Omar Khattab

13.1K posts

Omar Khattab banner
Omar Khattab

Omar Khattab

@lateinteraction

Asst professor @MIT CSAIL @nlp_mit. Research includes https://t.co/VgyLxl0oa1, https://t.co/ZZaSzaRaZ7 (@DSPyOSS), RLMs, and GEPA. Prev: CS PhD @StanfordNLP. Research @Databricks.

Cambridge, MA Katılım Aralık 2022
3.4K Takip Edilen31.3K Takipçiler
Omar Khattab
Omar Khattab@lateinteraction·
@nileshgupta2797 @LightOnIO @mixedbreadai Yup I don’t believe late interaction should fundamentally revolve around MaxSim. It’s just a lot better than single vector. It is possible to design richer pruning-friendly set-level / multi-vector scoring functions I think!
English
0
0
1
22
Nilesh Gupta
Nilesh Gupta@nileshgupta2797·
@lateinteraction @LightOnIO @mixedbreadai I agree logical compositions is def where single vectors fail hard - eucledian space by design doesn't allow modeling logical relevance. Though I've been wondering what are the limits of maxsim, is it limited to AND like queries or goes beyond (e.g. (A or B) and (not C))?
English
2
0
1
23
Omar Khattab
Omar Khattab@lateinteraction·
late interaction model (150M) beats the 54x larger Qwen3-8B-Embedding by... hmm, looks like up to 34% relative increase :D also really funny that the entire top section of the BC+ leaderboard, sorted by Recall, is just late interaction models by @LightOnIO and @mixedbreadai
Omar Khattab tweet media
Antoine Chaffin@antoine_chaffin

BrowseComp-Plus, perhaps the hardest popular deep research task, is now solved at nearly 90%... ... and all it took was a 150M model ✨ Thrilled to announce that Reason-ModernColBERT did it again and outperform all models (including models 54× bigger) on all metrics

English
9
14
171
11.6K
Omar Khattab
Omar Khattab@lateinteraction·
@nileshgupta2797 @LightOnIO @mixedbreadai The difference is huge on every distribution not yet overfit by the single vector models. This was true back in the day on all of BEIR, all of LoTTE, and all the OpenQA datasets, etc. Single vector is uncompositional. It advanced only by moving distributions in domain.
English
1
0
1
54
Nilesh Gupta
Nilesh Gupta@nileshgupta2797·
@lateinteraction @LightOnIO @mixedbreadai Do you think such a big diff is primarily because of the logical AND like multi-constraint nature of queries in BrowseComp for e.g. find X that satisfies A and B and C and D ..
Nilesh Gupta tweet media
English
1
0
1
68
Omar Khattab retweetledi
Joel Dierkes
Joel Dierkes@joeldierkes·
Mixedbread just made 115h of videos accessible to my agent. With the new @mixedbreadai v3 release, you can upload any video to your Mixedbread store and make its content accessible to your agent.
English
6
4
21
1.9K
Omar Khattab retweetledi
Pamela Fox
Pamela Fox@pamelafox·
@SQLGene @dbreunig indeed, SLMs came up so much last night- lots of reports of great performance after a bit of DSpy to correct malformed JSON and tool calling.
English
0
1
2
563
Omar Khattab retweetledi
Aamir
Aamir@aaxsh18·
Mixedbread stores can handle videos now with 48h+ runtime
English
0
3
23
1.3K
Omar Khattab retweetledi
Omar Khattab
Omar Khattab@lateinteraction·
what do people mean by ex-founder? how did you edit history to un-founder yourself
English
4
0
44
6.2K
Omar Khattab
Omar Khattab@lateinteraction·
to do really well on OOLONG
English
2
1
5
750
Omar Khattab retweetledi
alex zhang
alex zhang@a1zhang·
someone should try having RLMs write REPL code primarily using DSPy
English
6
7
116
14.3K
Omar Khattab retweetledi
Igor Carron
Igor Carron@IgorCarron·
Omar is shy about this late interaction approach changing everything. As an outsider and seeing @LightOnIO 's crew delivering exceptional results day after day, I prefer to frame this type of result as being one of those rare AlexNet moments. There is simply no turning back.
Igor Carron tweet media
Omar Khattab@lateinteraction

imo, these kinds of regular amazing results don't quite mean that late interaction is extremely strong per se as much as they mean that dense single-vector retrievers are a permanent bottleneck on your quality and generalization they're so bad!

English
0
5
9
1.8K
Omar Khattab
Omar Khattab@lateinteraction·
@HansiZeng @raphaelsrty @LightOnIO @mixedbreadai I definitely agree it’s not trivial to implement it right. But why does one need to implement it? It’s already implemented and optimized. We don’t usually implement our own attention kernels or HNSW index or anything like that. Rerankers are also far more expensive. Need GPU.
English
0
0
1
118
Omar Khattab retweetledi
Anthony Ronning
Anthony Ronning@anthonyronning·
okay i like half understand RLMs now and it's sick
English
2
1
21
3.6K
Omar Khattab retweetledi
Pamela Fox
Pamela Fox@pamelafox·
@dbreunig The DSpy meetup tonight was fantastic - even though I've never used DSpy, the case studies were full of general insights. Will speakers be sharing their slides?
English
3
2
9
1K