Ne Luo

36 posts

Ne Luo banner
Ne Luo

Ne Luo

@neluo19

PhD student @UniMelb | MSc @EdinburghNLP | interested in #NLP and #RL

Katılım Kasım 2016
516 Takip Edilen60 Takipçiler
Happy Researchers
Happy Researchers@hapyresearchers·
Undergrad --> PhD --> Postdoc --> PI --> bakery shop owner Who is thinking this career path?
English
90
266
2.8K
263.5K
Ne Luo retweetledi
Pasquale Minervini
Pasquale Minervini@PMinervini·
The amazing folks at @EdinburghNLP will be presenting a few papers at ACL 2025 (@aclmeeting); if you're in Vienna, touch base with them! Here are the papers in the main track 🧵
Pasquale Minervini tweet media
English
1
16
72
13.6K
Ne Luo
Ne Luo@neluo19·
@miniapeur Remind myself that happiness is a choice.
English
0
0
1
46
Mathieu
Mathieu@miniapeur·
What is your key to happiness?
English
24
2
30
59.5K
Ne Luo retweetledi
Aryo Pradipta Gema
Aryo Pradipta Gema@aryopg·
New Anthropic Research: “Inverse Scaling in Test-Time Compute” We found cases where longer reasoning leads to lower accuracy. Our findings suggest that naïve scaling of test-time compute may inadvertently reinforce problematic reasoning patterns. 🧵
Aryo Pradipta Gema tweet media
English
55
161
1.1K
338.6K
Ne Luo retweetledi
Alessio Devoto
Alessio Devoto@devoto_alessio·
🏆 Our @nvidia KV Cache Compression Leaderboard is now live! Compare state-of-the-art compression methods side-by-side with KVPress. See which techniques are leading in efficiency and performance. 🥇 huggingface.co/spaces/nvidia/…
Alessio Devoto tweet media
English
8
40
257
18.5K
Ne Luo
Ne Luo@neluo19·
Hi! I will be attending #NAACL2025 and presenting our paper on self-training for tool-use today, an extended work of my MSc dissertation at @EdinburghNLP, supervised by @PMinervini. Time: 14:00-15:30 Location: Hall 3 Let’s chat and connect!😊
Ne Luo tweet media
English
1
7
31
3.1K
Ne Luo retweetledi
Yu Zhao
Yu Zhao@yuzhaouoe·
NAACL 2025 Oral Presentation💥 Our work about using Sparse AutoEncoder to resolve knowledge conflict will present on 30 Apr 11:30–11:45 AM • Ballroom C Thank Hongru for presenting our work!!!
Hongru Wang@HongruWang007

🎉 Thrilled to share our TWO #NAACL2025 oral papers! 👇 Welcome to catch me and talk about anything! 1️⃣ Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering 📅 30 Apr • 11:30–11:45 AM • Ballroom C TLDR: A general representation learning framwork to detect knowledge conflicts and control the generation direction of LLMs 🔗 aclanthology.org/2025.naacl-lon… 2️⃣ Self-DC: When to Reason & When to Act? 📅 2 May • 11:30–11:45 AM • San Miguel First to balance 🧠 reasoning vs 🛠️ acting in language agents, enabling on-the-fly decomposition & execution of compositional unknown questions. 🔗 aclanthology.org/2025.naacl-lon… See you in Ballroom C & San Miguel! #LLM #Agent #NAACL2025

English
1
9
19
3.7K
🇺🇦 Dzmitry Bahdanau
🇺🇦 Dzmitry Bahdanau@DBahdanau·
Adam deserves the award, but in Singapore everyone still uses SGD
English
22
63
789
109.1K
Ne Luo
Ne Luo@neluo19·
Grateful to be part of this! I'll be presenting our paper "Self-Training Large Language Models for Tool-Use Without Demonstrations" at #NAACL2025! Also, I am currently seeking PhD opportunities. Please feel free to reach out if you're recruiting or know of any openings! :)
Pasquale Minervini@PMinervini

My amazing collaborators will present several works at ICLR and NAACL later this month -- please catch up with them if you're attending! I tried to summarise our recent work in a blog post: neuralnoise.com/2025/march-res…

English
0
2
14
2.4K
Ne Luo retweetledi
Richard Sutton
Richard Sutton@RichardSSutton·
David Silver really hits it out of the park in this podcast. The paper "Welcome to the Era of Experience" is here: goo.gle/3EiRKIH.
Google DeepMind@GoogleDeepMind

Human generated data has fueled incredible AI progress, but what comes next? 📈 On the latest episode of our podcast, @FryRsquared and David Silver, VP of Reinforcement Learning, talk about how we could move from the era of relying on human data to one where AI could learn for itself. Watch now → 00:00 Introduction 01:50 Era of experience 03:45 AlphaZero 10:19 Move 37 15:20 Reinforcement learning and human feedback 24:30 AlphaProof 29:50 Math Olympiads 35:00 Experience based methods 42:56 Hannah's reflections 44:00 Fan Hui joins

English
19
181
1K
181.9K
Ne Luo retweetledi
Peyman Milanfar
Peyman Milanfar@docmilanfar·
Researchers’ Credo
Peyman Milanfar tweet media
English
1
8
132
15.8K
Aryo Pradipta Gema
Aryo Pradipta Gema@aryopg·
Today, I'm starting as an AI Safety Fellow @AnthropicAI ! 🚀 Super excited to collaborate and learn from some of the brightest minds in AI! 🌟
English
3
0
14
700
Ne Luo retweetledi
elvis
elvis@omarsar0·
This is an excellent survey of post-training methods for LLMs
elvis tweet media
English
12
182
887
121.1K
Ne Luo retweetledi
Edward Hughes
Edward Hughes@edwardfhughes·
Agreed. Open-endedness is not just scaled-up objective optimisation. It is non-objective search. And it’s crucial for reaching ASI.
Thomas Wolf@Thom_Wolf

I shared a controversial take the other day at an event and I decided to write it down in a longer format: I’m afraid AI won't give us a "compressed 21st century". The "compressed 21st century" comes from Dario's "Machine of Loving Grace" and if you haven’t read it, you probably should, it’s a noteworthy essay. In a nutshell the paper claims that, over a year or two, we’ll have a "country of Einsteins sitting in a data center”, and it will result in a compressed 21st century during which all the scientific discoveries of the 21st century will happen in the span of only 5-10 years. I read this essay twice. The first time I was totally amazed: AI will change everything in science in 5 years, I thought! A few days later I came back to it and, re-reading it, I realized that much of it seemed like wishful thinking at best. What we'll actually get, in my opinion, is “a country of yes-men on servers” (if we just continue on current trends). Let me explain the difference with a small part of my personal story. I’ve always been a straight-A student. Coming from a small village, I joined the top French engineering school before getting accepted to MIT for PhD. School was always quite easy for me. I could just get where the professor was going, where the exam's creators were taking us and could predict the test questions beforehand. That’s why, when I eventually became a researcher (more specifically a PhD student), I was completely shocked to discover that I was a pretty average, underwhelming, mediocre researcher. While many colleagues around me had interesting ideas, I was constantly hitting a wall. If something was not written in a book I could not invent it unless it was a rather useless variation of a known theory. More annoyingly, I found it very hard to challenge the status-quo, to question what I had learned. I was no Einstein, I was just very good at school. Or maybe even: I was no Einstein in part *because* I was good at school. History is filled with geniuses struggling during their studies. Edison was called "addled" by his teacher. Barbara McClintock got criticized for "weird thinking" before winning a Nobel Prize. Einstein failed his first attempt at the ETH Zurich entrance exam. And the list goes on. The main mistake people usually make is thinking Newton or Einstein were just scaled-up good students, that a genius comes to life when you linearly extrapolate a top-10% student. This perspective misses the most crucial aspect of science: the skill to ask the right questions and to challenge even what one has learned. A real science breakthrough is Copernicus proposing, against all the knowledge of his days -in ML terms we would say “despite all his training dataset”-, that the earth may orbit the sun rather than the other way around. To create an Einstein in a data center, we don't just need a system that knows all the answers, but rather one that can ask questions nobody else has thought of or dared to ask. One that writes 'What if everyone is wrong about this?' when all textbooks, experts, and common knowledge suggest otherwise. Just consider the crazy paradigm shift of special relativity and the guts it took to formulate a first axiom like “let’s assume the speed of light is constant in all frames of reference” defying the common sense of these days (and even of today…) Or take CRISPR, generally considered to be an adaptive bacterial immune system since the 80s until, 25 years after its discovery, Jennifer Doudna and Emmanuelle Charpentier proposed to use it for something much broader and general: gene editing, leading to a Nobel prize. This type of realization –"we've known XX does YY for years, but what if we've been wrong about it all along? Or what if we could apply it to the entirely different concept of ZZ instead?” is an example of out-side-of-knowledge thinking –or paradigm shift– which is essentially making the progress of science. Such paradigm shifts happen rarely, maybe 1-2 times a year and are usually awarded Nobel prizes once everybody has taken stock of the impact. However rare they are, I agree with Dario in saying that they take the lion’s share in defining scientific progress over a given century while the rest is mostly noise. Now let’s consider what we’re currently using to benchmark recent AI model intelligence improvement. Some of the most recent AI tests are for instance the grandiosely named "Humanity's Last Exam" or "Frontier Math". They consist of very difficult questions –usually written by PhDs– but with clear, closed-end, answers. These are exactly the kinds of exams where I excelled in my field. These benchmarks test if AI models can find the right answers to a set of questions we already know the answer to. However, real scientific breakthroughs will come not from answering known questions, but from asking challenging new questions and questioning common conceptions and previous ideas. Remember Douglas Adams' Hitchhiker's Guide? The answer is apparently 42, but nobody knows the right question. That's research in a nutshell. In my opinion this is one of the reasons LLMs, while they already have all of humanity's knowledge in memory, haven't generated any new knowledge by connecting previously unrelated facts. They're mostly doing "manifold filling" at the moment - filling in the interpolation gaps between what humans already know, somehow treating knowledge as an intangible fabric of reality. We're currently building very obedient students, not revolutionaries. This is perfect for today’s main goal in the field of creating great assistants and overly compliant helpers. But until we find a way to incentivize them to question their knowledge and propose ideas that potentially go against past training data, they won't give us scientific revolutions yet. If we want scientific breakthroughs, we should probably explore how we’re currently measuring the performance of AI models and move to a measure of knowledge and reasoning able to test if scientific AI models can for instance: - Challenge their own training data knowledge - Take bold counterfactual approaches - Make general proposals based on tiny hints - Ask non-obvious questions that lead to new research paths We don't need an A+ student who can answer every question with general knowledge. We need a B student who sees and questions what everyone else missed. --- PS: You might be wondering what such a benchmark could look like. Evaluating it could involve testing a model on some recent discovery it should not know yet (a modern equivalent of special relativity) and explore how the model might start asking the right questions on a topic it has no exposure to the answers or conceptual framework of. This is challenging because most models are trained on virtually all human knowledge available today but it seems essential if we want to benchmark these behaviors. Overall this is really an open question and I’ll be happy to hear your insightful thoughts.

English
3
3
37
5.1K
Ne Luo retweetledi
Yu Zhao
Yu Zhao@yuzhaouoe·
We find a single biased direction encodes a KV Cache selection mechanism in Self-Attention -- Key vector with a strong component in this direction results in this Key-Value pair being ignored by Query🚀🚀🚀
Nathan Godey@nthngdy

🚀 New Paper Alert! 🚀 We introduce Q-Filters, a training-free method for efficient KV Cache compression! It is compatible with FlashAttention and can compress along generation which is particularly useful for reasoning models ⚡ ⬇️R1-Distill-Llama-8B with 128 KV pairs ⬇️ 🧵

English
0
11
25
2.3K