Lapis Labs

33 posts

Lapis Labs banner
Lapis Labs

Lapis Labs

@lapisrocks

be novel.

Katılım Şubat 2024
2 Takip Edilen52 Takipçiler
Lapis Labs retweetledi
Andy Zhou
Andy Zhou@zhouandy_·
Announcing the first fully AI-generated scientific discovery to pass the highest level of peer review – the main track of an A* conference (ACL 2025). Several groups have shown AI-generated work at workshops, but main conference acceptance is a far higher bar. While workshops often accept more than 50% of submissions, top-tier venues like ACL accept ~20%.
English
19
27
257
71.1K
Lapis Labs retweetledi
Intology
Intology@IntologyAI·
The 1st fully AI-generated scientific discovery to pass the highest level of peer review – the main track of an A* conference (ACL 2025). Zochi, the 1st PhD-level agent. Beta open.
English
41
130
653
217.3K
Lapis Labs retweetledi
Intology
Intology@IntologyAI·
📰 Zochi @ ICLR 2025! After productive discussions with the workshop organizers that accepted Zochi's work, we are pleased to announce that all work produced by Zochi has been approved for presentation at ICLR 2025. Intology reps will present on Zochi’s behalf. As requested by the organizers, papers/posters will note that authors don’t claim sole credit. The exact wording requested can be found in our blog below 🧵👇
English
1
3
9
2.4K
Lapis Labs retweetledi
Intology
Intology@IntologyAI·
🤖🔬Today we are debuting Zochi, the world’s first Artificial Scientist with state-of-the-art contributions accepted in ICLR 2025 workshops. Unlike existing systems, Zochi autonomously tackles some of the most challenging problems in AI, producing novel contributions in days—from idea to finalized publication. With a standardized automated reviewer, Zochi’s papers score an average of 7.67 compared to other publicly available papers generated by AI systems that score between 3 and 4. 🧵👇
Intology tweet media
English
14
52
191
74.9K
Lapis Labs
Lapis Labs@lapisrocks·
🎉 Oral, Spotlight, NeurIPS, oh my! 🎉 We're thrilled to announce that two of our papers at #NeurIPS2024 have received special recognition: ❗Oral❗ presentation in the Datasets and Benchmarks track AND ❗Spotlight❗ 👏 Massive congratulations to our student researchers, Nikhil Khandekar and Soren Dunn, for their groundbreaking work on MEDCALC-BENCH, which earned them the Oral @ NeurIPS 2024 recognition, alongside our amazing collaborators at the National Institutes of Health. Catch us on stage! 🫡 📄 Paper: arxiv.org/abs/2406.12036 👏 And a big shoutout to @andyz245 for the fantastic work on RPO, which secured a Spotlight @ NeurIPS 2024! 📄 Paper: arxiv.org/abs/2401.17263 We’re incredibly grateful to the @NCSAatIllinois and National Science Foundation (NSF) for their invaluable computational support that helped make this research possible. If you’re attending NeurIPS this year, don’t hesitate to stop by and say hello! We look forward to seeing everyone in Canada 🇨🇦
Lapis Labs tweet media
English
0
0
5
574
Lapis Labs retweetledi
Andy Zhou
Andy Zhou@zhouandy_·
Excited to announce three papers were accepted to #NeurIPS2024! 🤖 Robust Prompt Optimization for Defending Language Models Against Jailbreaking Attacks (Spotlight) arxiv.org/abs/2401.17263 - Jailbreaking defense based on adversarial training Jailbreaking Large Language Models Against Moderation Guardrails via Cipher Characters arxiv.org/abs/2405.20413 - A cipher-based attack on guardrail models like LlamaGuard RedCode: Risky Code Execution and Generation Benchmark for Code Agents - A benchmark for evaluating unsafe code execution and generation Check out our findings! Looking forward to visiting Vancouver 🇨🇦
Andy Zhou tweet media
English
5
10
87
13.2K
Lapis Labs
Lapis Labs@lapisrocks·
🎉 We are very excited to be announcing our work done with CAIS (@cais)! Congratulations to our student researchers @rishub_t and @8hrugu for their amazing work along with the rest of the authors on this project. Read @rishub_t's tweet below to learn more 🫡
Rishub Tamirisa@rishub_t

🚨New Paper: “Tamper-Resistant Safeguards for Open-Weight LLMs” We introduce the first safeguards for LLMs that resist realistic fine-tuning attacks of up to 5,000 steps, demonstrating the potential of tamper-resistance as a powerful new tool for making open-weight LLMs safer.🧵

English
3
1
4
854
Lapis Labs
Lapis Labs@lapisrocks·
Happy to announce our collaboration with @askalphaxiv 🫡 "We are incredibly excited to support this effort by helping develop their reviewer program with our student researchers at Lapis Labs, along with utilizing this service to promote direct discussion with our work and authors." Read more: lapisrocks.substack.com/p/working-with…
English
0
1
5
4.1K
Lapis Labs
Lapis Labs@lapisrocks·
“This was a highly interdisciplinary project, where we had individuals not only from computational backgrounds, but also had MD students from Yale, UIC, UChicago, and Rosalind Franklin University who helped us ensure that the patient notes in the dataset were appropriate for each calculation task.”¹ - Nikhil Khandekar¹, student researcher
English
0
0
1
175
Lapis Labs
Lapis Labs@lapisrocks·
We are excited to introduce FIRST! Our novel LLM reranking approach boosts efficiency by 50% while maintaining performance. Paper link: arxiv.org/pdf/2406.15657 Big thanks to @gangi_official and @hengjinlp, along with our collaborators @IBMResearch (@aviaviavi__ and Arafat Sultan)
Revanth Gangi Reddy@gangi_official

Introducing FIRST: Faster Improved Listwise Reranking with Single Token Decoding arxiv.org/pdf/2406.15657 Listwise LLM reranking typically outputs the ranking order as a generation sequence. Instead, we use output logits of the first generated identifier to obtain the ranking.

English
0
1
6
481