Cesare Campagnano

52 posts

Cesare Campagnano

Cesare Campagnano

@caesar_one_

PostDoc @ Sapienza University of Rome | ex-Amazon

Rome, Italy Katılım Mart 2010
419 Takip Edilen226 Takipçiler
Sabitlenmiş Tweet
Cesare Campagnano
Cesare Campagnano@caesar_one_·
I’m thrilled to participate in such a prestigious conference with my first paper! See you in Dublin at #ACL2022 😎 #NLProc
SapienzaNLP@SapienzaNLP

#NLPaperAlert 📢 We bring together existing resources, revise them, and propose SRL4E, a unified evaluation on Semantic Role Labeling 4 Emotions! Read our #ACL2022 preprint: researchgate.net/publication/35… By @caesar_one_ @ConiaSimone @RNavigli + @ERC_Research @EuroLangTech #NLProc

English
0
0
6
0
Cesare Campagnano retweetledi
RSTLess group
RSTLess group@RSTLessGroup·
We are very excited to share that the work of @caesar_one_ , @antonio_mallia , @JackPertschuk and @fabreetseo has been accepted to #ECIR2025 as a #shortpaper. See you in #Lucca. @ecir2025 @pinecone #AI #Research #IR #industry
Pinecone@pinecone

Congratulations to our very own @antonio_mallia, @caesar_one_, and @JackPertschuk – as well as their co-authors – on their accepted #ECIR2025 research papers! 🎉 They continue to push the state-of-the-art forward on information retrieval, and we as an industry are better for it! 📚 📜 Sean MacAvaney, Antonio Mallia and Nicola Tonellotto: “Efficient Constant-Space Multi-Vector Retrieval", 2025 📜 Kaili Huang, Thejas Venkatesh, Uma Dingankar, Antonio Mallia, Daniel Campos, Jian Jiao, Christopher Potts, Matei Zaharia, Kwabena Boahen, Omar Khattab, Saarthak Sarup and Keshav Santhanam: “ColBERT-serve: Efficient Multi-Stage Memory-Mapped Scoring”, 2025 📜 Cesare Campagnano, Antonio Mallia, Jack Pertschuk and Fabrizio Silvestri: “E2Rank: Efficient and Effective Layer-wise Reranking”, 2025

English
0
3
10
596
Cesare Campagnano retweetledi
Pinecone
Pinecone@pinecone·
Congratulations to our very own @antonio_mallia, @caesar_one_, and @JackPertschuk – as well as their co-authors – on their accepted #ECIR2025 research papers! 🎉 They continue to push the state-of-the-art forward on information retrieval, and we as an industry are better for it! 📚 📜 Sean MacAvaney, Antonio Mallia and Nicola Tonellotto: “Efficient Constant-Space Multi-Vector Retrieval", 2025 📜 Kaili Huang, Thejas Venkatesh, Uma Dingankar, Antonio Mallia, Daniel Campos, Jian Jiao, Christopher Potts, Matei Zaharia, Kwabena Boahen, Omar Khattab, Saarthak Sarup and Keshav Santhanam: “ColBERT-serve: Efficient Multi-Stage Memory-Mapped Scoring”, 2025 📜 Cesare Campagnano, Antonio Mallia, Jack Pertschuk and Fabrizio Silvestri: “E2Rank: Efficient and Effective Layer-wise Reranking”, 2025
Pinecone tweet media
English
0
2
11
1.6K
Cesare Campagnano retweetledi
Min Choi
Min Choi@minchoi·
Llama 3 is insanely moving fast. People are really pushing Llama 3 to its limits in incredible ways. 10 wild examples (and use cases)
Min Choi tweet media
English
73
442
3.2K
1.2M
Cesare Campagnano retweetledi
Daniel Vila Suero
Daniel Vila Suero@dvilasuero·
This is actually huge: - No SFT stage (e.g., Zephyr used 200k examples) - Preference tuning with 7K examples only (other models trained with at least 60k samples) I've put a lot of care & love building the DPO version of the amazing Capybara dataset from @ldjconfirmed so I'm really pleased to see these results. Let's double down on useful open data for OSS AI developers and researchers
Daniel Vila Suero tweet media
Jiwoo Hong@jiwoohong98

📢New model, Mistral-ORPO-Capybara-7k in ORPO collection!🧵 With 💡ORPO💡 + 7k Capybara preference pair by @argilla_io🔥 + Mistral (7B), you can get the human-aligned chat model within 2.5 hours of fine-tuning👀 👉AlpacaEval 2.0 (LC): 15.9% 👉MT-Bench: 7.44 👉IFEval: 61.27%

English
2
17
72
6.5K
Cesare Campagnano retweetledi
Fabrizio Silvestri
Fabrizio Silvestri@fabreetseo·
🤯 Think adding nonsense to RAG systems is madness? Our new paper says otherwise! We found that including random documents boost accuracy by 30+%, challenging old paradigms and showing the complexity of integrating retrieval w/ language generation. #RAGSystems #surprisingresults
English
4
10
56
6.3K
Cesare Campagnano retweetledi
elvis
elvis@omarsar0·
Redefining Retrieval in RAG A nice comprehensive study that focuses on the components needed to improve the retrieval component of a RAG system. Confirms that the position of relevant information should be placed near the query. The model will struggle to attend to the information if this is not the case. Surprisingly, it finds that related documents don't necessarily lead to improved performance for the RAG system. Even more unexpectedly, irrelevant and noisy documents can actually help drive up accuracy if placed correctly. We need more systematic studies around RAG. The hard part of a RAG system is typically the retriever component. Just dumping relevant docs into the context is not an effective approach but it's what a lot of LLM devs do. I like that the Ragas library proposes the use of several metrics for assessing a RAG system at both the generation and retrieval stages, including an end-to-end evaluation. It's a good first step but we still need better ways to integrate external information that can be effectively leveraged by the generative component.
elvis tweet media
English
10
183
887
88.9K
Andrej Karpathy
Andrej Karpathy@karpathy·
With many 🧩 dropping recently, a more complete picture is emerging of LLMs not as a chatbot, but the kernel process of a new Operating System. E.g. today it orchestrates: - Input & Output across modalities (text, audio, vision) - Code interpreter, ability to write & run programs - Browser / internet access - Embeddings database for files and internal memory storage & retrieval A lot of computing concepts carry over. Currently we have single-threaded execution running at ~10Hz (tok/s) and enjoy looking at the assembly-level execution traces stream by. Concepts from computer security carry over, with attacks, defenses and emerging vulnerabilities. I also like the nearest neighbor analogy of "Operating System" because the industry is starting to shape up similar: Windows, OS X, and Linux <-> GPT, PaLM, Claude, and Llama/Mistral(?:)). An OS comes with default apps but has an app store. Most apps can be adapted to multiple platforms. TLDR looking at LLMs as chatbots is the same as looking at early computers as calculators. We're seeing an emergence of a whole new computing paradigm, and it is very early.
Andrej Karpathy tweet media
English
296
1.9K
9.2K
2.1M
Cesare Campagnano retweetledi
Giovanni Trappolini
Giovanni Trappolini@GioTrappolini·
Still can't handle the indecisiveness between Barbie and Oppenheimer? 😫💥 Don't fret! Come to the presentation of our new perspective paper, "Multimodal Neural Databases", where we lay out the vision for database-like queries on multimodal data. Tomorrow @SIGIR2023, 1.30pm GMT+8
Giovanni Trappolini tweet media
English
1
5
21
2.7K
Cesare Campagnano retweetledi
Suraj Srinivas
Suraj Srinivas@Suuraj·
Three papers accepted at NeurIPS'22 (!!) 1) Efficiently training low-curvature neural networks (arxiv.org/abs/2206.07144), w/ Kyle Matoba, @hima_lakkaraju, @francoisfleuret We propose to build NNs that are "as linear as possible", and thus eliminate excess model curvature.
Suraj Srinivas tweet media
English
5
34
219
0
Cesare Campagnano retweetledi
Ben Meer
Ben Meer@SystemSunday·
YouTube is free education. But 99% don’t know the best spots on its virtual campus. Here are the top channels to accelerate your learning:
English
1.6K
48.7K
234.3K
0
Cesare Campagnano retweetledi
Bojan Tunguz
Bojan Tunguz@tunguz·
This week @Google researchers announced Minerva, an internally developed project that can answer mathematical questions and tackle other complex topics such as physics. 1/5
Bojan Tunguz tweet media
English
18
362
1.9K
0