jonah

1.6K posts

jonah banner
jonah

jonah

@drexalt

research @mixedbreadai

gangnam Katılım Aralık 2014
1.4K Takip Edilen554 Takipçiler
jonah
jonah@drexalt·
@matospiso another one has been hinted at in a linkedin post, hopefully will drop on monday? @zongyuxuan3 :D
English
2
0
1
140
tomaarsen
tomaarsen@tomaarsen·
⬆️ I've just released Sentence Transformers v5.3.0! This release upgrades training with MultipleNegativesRankingLoss with alternative InfoNCE formulations and hardness weighting, adds two new losses, and more. Details in 🧵
tomaarsen tweet media
English
4
9
96
4.5K
jonah
jonah@drexalt·
@din0s_ lobster too buttery, pants too big
English
1
0
1
45
dinos
dinos@din0s_·
the one thing nobody warned me about with cutting is that none of my pants fit anymore
English
1
0
5
195
jonah
jonah@drexalt·
@antoine_chaffin if we only ever eval colbertv2, splade-v3 can keep looking competitive with multi vector :)
English
1
0
2
197
Antoine Chaffin
Antoine Chaffin@antoine_chaffin·
Different day, same results Stop living in the past, have a taste of the future Small models outperforming billion parameters models, the same code as dense models, easy to use indexes… PyLate has it all, at this point, you are just missing out lightonai.github.io/pylate/
Omar Khattab@lateinteraction

Wow. It’s absolutely preposterous that ColBERTv2, a 100M parameter retriever, still fricking outperforms Qwen3-Embed-8B, an 80x bigger dense retriever. ColBERTv2 was trained by one dude in 2021 on 4 A100s for 4 days, on top of puny BERT-base. Single-vector models hold IR back.

English
3
9
57
7.8K
jonah
jonah@drexalt·
@candyflipline outting myself but 30 mins for the last few weeks on every all-in pod too
English
0
0
0
348
jonah
jonah@drexalt·
@airkatakana damn if she makes it through this one, she's a keeper
English
0
0
17
1.1K
xjdr
xjdr@_xjdr·
@cloud11665 i have no doubt its excellent. but the _best_ slice in sf?!
GIF
English
3
0
12
635
chris
chris@hingeloss·
lucidrains github gone? RIP sweet prince
chris tweet media
English
13
11
126
22K
Raphaël Sourty
Raphaël Sourty@raphaelsrty·
Releasing NextPlaid today at @LightOnIO. It’s a production-ready multi-vector database 🎉 NextPlaid lets you deploy an API in seconds using our pre-built containers. It embeds a multi-vector database and an inference engine for late-interaction models based on ONNX.
English
5
8
55
7.3K
jonah
jonah@drexalt·
@din0s_ gm. check the output dim on this guy
English
1
0
2
96
dinos
dinos@din0s_·
gm
dinos tweet media
2
1
29
2.5K
jonah retweetledi
Ben Clavié
Ben Clavié@bclavie·
I'm personally very bearish on MUVERA due to its many, many, failure cases, but I have a lot of respect for the @weaviate_io folks, so I gave this another deep read to see if there were things that could change my mind. However this has me a bit puzzled, if I'm reading the graph below right, it means that MUVERA itself produces a ~50+% incompressible performance degradation at commonly used indexing parameters, and still a ~20% degradation at near-bruteforce search tier parameters (ef=1024), meaning that the degradation would be purely due to MUVERA itself. For most retrieval uses, this would make the method completely unusable, as this degradation for many workflows is almost similar to the one we'd experience from replacing semantic search with pure bm25/keyword search. I feel like I'm missing something here so I'm very happy to be corrected if I'm misinterpreting the results!
Ben Clavié tweet media
Femke Plantinga@femke_plantinga

Multi-vector embeddings (ColBERT, ColPali) are budget killers. But MUVERA can cut your memory footprint by 70%. Multi-vector models offer incredible retrieval but suffer from massive memory overhead and slow indexing. MUVERA (Multi-Vector Retrieval via Fixed Dimensional Encodings) compresses these into single, fixed-dimensional vectors. How it works: MUVERA condenses a sequence of vectors (e.g., 100x96d) into one vector via: 1️⃣ Space Partitioning: Groups vectors into buckets using SimHash or k-means clustering. 2️⃣ Dimensionality Reduction: Applies random linear projection to compress each sub-vector while preserving dot products. 3️⃣ Repetitions: Repeats the process multiple times and concatenates results to improve accuracy. 4️⃣ Final Projection: Optional final compression (not used in Weaviate's implementation). The impact (LoTTE benchmark): - Memory: 12GB → <1GB. - Indexing: 20+ mins → 3-6 mins. - HNSW Graph: 99% smaller. There’s a trade-off: You trade a slight dip in raw recall for massive efficiency gains. However, by tuning the HNSW `ef` parameter (e.g., `ef=512`), you can recover 80-90%+ recall while keeping costs low. When should you use MUVERA? → Large-scale production RAG → Systems where memory/infrastructure costs are the direct bottleneck → Use cases requiring fast indexing MUVERA in @weaviate_io 1.31+ takes just a couple of lines of code. You can tune three parameters (k_sim, d_proj, r_reps) to balance memory usage and retrieval accuracy for your specific use case. Read the full technical deep-dive here: weaviate.io/blog/muvera?ut…

English
5
3
45
4.4K
jonah
jonah@drexalt·
@din0s_ 💀 get aerospace though
English
0
0
0
21
dinos
dinos@din0s_·
this is genuinely crazy
dinos tweet media
English
2
0
0
153
dinos
dinos@din0s_·
fresh install, what should i get?
dinos tweet media
English
1
0
2
196
jonah
jonah@drexalt·
@vikhyatk georgism? monarchism? many options left, don't despair
English
0
0
1
51
vik
vik@vikhyatk·
i like capitalism but i’m finding capitalism supporters tend to be low IQ and i don’t want to be associated with that socialism is too mainstream. i don’t want to be seen as a trend follower my only option now is to pivot to nihilism
English
17
0
92
7.1K
jonah
jonah@drexalt·
@staghado @Dorialexander thanks for the eval, was wondering about this one. nice to see lighton keeping the crazy lead 💪
English
0
0
3
212
staghado
staghado@staghado·
@Dorialexander a small improvement from v1 but overall still behind most recent models on OlmOCR-bench including PaddleOCR-VL, OlmOCR, Chandra, Mistral OCR 3, very close to dots.ocr, and way below LightOnOCR-2-1B 👑
staghado tweet media
English
2
0
13
5.6K