Alberto Fuentes (e/acc)

59.1K posts

Alberto Fuentes (e/acc) banner
Alberto Fuentes (e/acc)

Alberto Fuentes (e/acc)

@AlberFuen

Founder of @daertml. Training LLaMAs as a hobby (and no profit yet).

Madrid, Comunidad de Madrid Katılım Nisan 2018
2.8K Takip Edilen759 Takipçiler
Sabitlenmiş Tweet
Alberto Fuentes (e/acc)
Alberto Fuentes (e/acc)@AlberFuen·
AGI achieved externally in the 4chan chat by miqudev anon, on 29th January 2024. Here goes a 🧵with Miqu rocking everything I ask (from datasets, random things I find from the internet and more). Feel the AGI!! Using the Q5 (biggest model) version, with this llama.cpp config:
Alberto Fuentes (e/acc) tweet media
English
1
1
22
6.9K
Alberto Fuentes (e/acc) retweetledi
@·
Why is your 405B model losing to an 8B model? 📉"Context Rot" is a logic problem, not a scale problem. Based on the amazing work of RLMs, we built $\lambda$-RLM: replacing messy AI-generated code with a typed $\lambda$-Calculus runtime. The results: ✅ +21.9 accuracy gain
 tweet media
English
0
2
14
570
Alberto Fuentes (e/acc) retweetledi
@·
Una china con acento gallego. Su padre no habla ni papa de español. Y juntos tienen un restaurante buffet en A Coruña que no para de llenarse. Mira esto 📽️komosushi_
Español
16
39
204
14.3K
Alberto Fuentes (e/acc) retweetledi
Alfonso C. Suárez
Alfonso C. Suárez@AlfonsoCSuarez·
🥩 Cinco consejos para comprar en la carnicería si te da vergüenza (o no sabes cómo pedir) El mostrador puede llegar a intimidar y la mayoría acaba yendo al supermercado a comprar la carne en bandejas. Error. El carnicero de tu barrio es un gran aliado para tu cocina. Quédate con estas claves: 👇 1️⃣ Pide por raciones, no por gramos. Calcular "300 gramos de lomo" es un lío. Pide "4 filetes finos" o "2 contramuslos". Tú sabes cuánto se come en tu casa y él sabe el grosor exacto para que quede bien. 2️⃣ Habla de recetas, no de anatomía. No hace falta que sepas lo que es la babilla o la tapilla de ternera. Dile al carnicero: "Quiero hacer un guiso para tres personas" o "algo a la plancha que quede jugoso". Déjale hacer su trabajo. 3️⃣ Aprovecha la mano de obra (es gratis). La carnicería no es el súper. ¿Quieres la carne en tiras para un wok? ¿Picada fina? ¿Sin un gramo de grasa? Pídelo. Te vas a casa con la mise en place hecha. 4️⃣ El 'truco' del pollo entero. Realmente no es un truco, pero es algo que puede que no sepas: comprar un pollo entero es mucho más barato que comprar las bandejas sueltas. Pídele que te lo despiece: pechugas fileteadas, muslos para asar, alitas... 5️⃣ No te dejes el oro líquido (los huesos). Cuando te despiecen ese pollo o te limpien una carne, pide siempre que te guarden los huesos y carcasas. Son la base gratuita para los mejores caldos y guisos que vas a hacer en tu vida. 💡 Solo hay que perderle el miedo al mostrador. ¿Cuál es vuestro truco más útil cuando compráis en la carnicería? 👇
Alfonso C. Suárez tweet mediaAlfonso C. Suárez tweet media
Español
7
25
323
35.4K
Alberto Fuentes (e/acc) retweetledi
@·
Here's a common misconception about RAG! When we talk about RAG, it's usually thought: index the doc → retrieve the same doc. But indexing ≠ retrieval So the data you index doesn't have to be the data you feed the LLM during generation. Here are 4 smart ways to index data: 1) Chunk Indexing - The most common approach. - Split the doc into chunks, embed, and store them in a vector DB. - At query time, the closest chunks are retrieved directly. This is simple and effective, but large or noisy chunks can reduce precision. 2) Sub-chunk Indexing - Take the original chunks and break them down further into sub-chunks. - Index using these finer-grained pieces. - Retrieval still gives you the larger chunk for context. This helps when documents contain multiple concepts in one section, increasing the chances of matching queries accurately. 3) Query Indexing - Instead of indexing the raw text, generate hypothetical questions that an LLM thinks the chunk can answer. - Embed those questions and store them. - During retrieval, real user queries naturally align better with these generated questions. - A similar idea is also used in HyDE, but there, we match a hypothetical answer to the actual chunks. This is great for QA-style systems, since it narrows the semantic gap between user queries and stored data. 4) Summary Indexing - Use an LLM to summarize each chunk into a concise semantic representation. - Index the summary instead of the raw text. - Retrieval still returns the full chunk for context. This is particularly effective for dense or structured data (like CSVs/tables) where embeddings of raw text aren’t meaningful. 👉 Over to you: What are some strategies that you commonly use for RAG indexing?
 tweet media
English
13
121
598
43K
Alberto Fuentes (e/acc) retweetledi
@·
hey folks if you are super duper worried about AI systems I highly recommend learning how they work internally like at a technical level will really ground you in the janky present and refrain you from posting a video asking interview questions to sonnet
 tweet media
English
3
3
86
3.3K
Alberto Fuentes (e/acc) retweetledi
Yizhou Liu
Yizhou Liu@YizhouLiu0·
Seems to be a systematic scaling study for diffusion language model 👍 Not surprised that the exponents are still similar to Chinchilla. But the origin of 21.8x speedup? So far I can imagine diffusion models enable better hyperparameter choices.
Chen-Hao (Lance) Chao@chenhao_chao

(1/7) We introduce MDM-Prime-v2 which scales 21.8× better than autoregressive models (ARMs) in compute-optimal comparisons. 📎 Paper: arxiv.org/abs/2603.16077 🌟 Blog: chen-hao-chao.github.io/mdm-prime-v2 ⌨️ Github: github.com/chen-hao-chao/… Here’s how we did it👇:

English
0
4
17
2.3K
Alberto Fuentes (e/acc)
RT @victormustar: Forgot about this qwen3.5-0.8B experiment, the results: - 0% -> 26.5% DOOM action prediction, 16 autonomous experiments…
English
0
1
0
0
Alberto Fuentes (e/acc) retweetledi
Wildminder
Wildminder@wildmindai·
3DreamBooth- high-fidelity 3D subject-driven video generation. AI just made running an e-commerce brand ridiculously easy. - view-consistent videos from multi-view references; - snap a few photos of your product - turns them into cinematic 3D videos - perfect 360-degree rotation in any scene - zero warped logos or lost textures it's HunyuanVideo-1.5 + LoRA. ko-lani.github.io/3DreamBooth/
English
0
5
21
1.6K
Alberto Fuentes (e/acc) retweetledi
Rosinality
Rosinality@rosinality·
Training set and benchmark for the problem require deriving mathematical objects. The training set itself is made using GPT-OSS-120B with self-consistency. They post-trained the model with an LLM verifier trained with RLVR.
Rosinality tweet media
English
1
3
23
483
Alberto Fuentes (e/acc) retweetledi
@·
🆕Hoy en @elDiarioes volvemos a actualizar nuestra calculadora de salarios con los datos de la EES 💵⚖️¿Cres que cobras poco? Te proponemos un baño de realidad para que conozcas tu posición en la escala salarial de España
 tweet media
Español
52
31
329
157.5K
Alberto Fuentes (e/acc) retweetledi
@·
These might have flown under the radar, but these open SWE models entered the top 20 of SWE-Bench Verified! A fully transparent framework for SWE agent training in Python, comprising 45,320 executable Docker environments spanning over 12.8k repositories!
 tweet media
English
0
3
9
1.6K
Alberto Fuentes (e/acc) retweetledi
@·
Yes! We're bullish on VLMs for tracking objects 💪 Handling occlusions, re-identification across shots, flexible text queries—these all need reasoning from LLMs🧠, not just a pure CV system. Just like VLMs have taken over OCR, VLMs should handle any grounding task like
English
1
5
46
3.4K
Alberto Fuentes (e/acc) retweetledi
Rosinality
Rosinality@rosinality·
If expert routing is completely independent the number of possible paths would be (# experts)^(# layers), which is larger than the number of training tokens. Practically it is not independent. What if we further reduce the paths by sharing router weights?
Rosinality tweet media
English
2
10
41
1.1K
Alberto Fuentes (e/acc) retweetledi
@·
FASTER achieves 10x faster action sampling for real-time VLAs By compressing multi-step denoising into a single step, it enables immediate reaction in highly dynamic tasks like table tennis—even on consumer GPUs such as the RTX 4060.
 tweet media
English
1
5
28
1.7K
Alberto Fuentes (e/acc) retweetledi
@·
the holy grail of robotics is a self improvement loop that can hill climb into unambiguously superhuman territory on real tasks by sharing internal VLA state with an adaptive policy, we got an off policy TD actor critic setup to do that figure 9 is the money shot
 tweet media
English
5
28
371
28.1K