Joe Davison

2.1K posts

Joe Davison banner
Joe Davison

Joe Davison

@joeddav

AI Research & Eng. @huggingface alumn.

Lehi, UT 🏔 Sumali Ocak 2009
1.1K Sinusundan3K Mga Tagasunod
Joe Davison nag-retweet
Adam Zweiger
Adam Zweiger@AdamZweiger·
We introduce a new approach for fast and high-quality context compaction in latent space. Attention Matching (AM) achieves 50× compaction in seconds with little performance loss, substantially outperforming summarization and other baselines.
Adam Zweiger tweet media
English
23
147
942
128.9K
Joe Davison nag-retweet
Nathan Lambert
Nathan Lambert@natolambert·
Really excited for the things @rosstaylor90 has been building to go out into the world. He's been one of the people I can always rely on to have non-cope takes on what we need to do to make the open ecosystem great. What a great time in RL.
General Reasoning@GenReasoning

Introducing OpenReward. 🌍 330+ RL environments through one API ⚡ Autoscaled sandbox compute 🍒 4.5M+ unique RL tasks 🚂 Works like magic with Tinker, Miles, Slime Link and thread below.

English
7
18
197
27.6K
Joe Davison nag-retweet
Benjamin Todd
Benjamin Todd@ben_j_todd·
Opus 4.6 is hugely better at Pokemon: • Opus 4.0 took 1,000 hours to get half way through • Opus 4.5 could almost finish in 1,000 hours • Opus 4.6 was another 10x faster!
Benjamin Todd tweet media
English
28
51
766
123.6K
Joe Davison nag-retweet
Perturb.ai
Perturb.ai@perturbai_tx·
Introducing PerturbAI. Today we announced our emergence from stealth with the release of the world’s largest in vivo CRISPR data engine, interrogating the effects of thousands of genetic perturbations across 8 million cells throughout the whole brain. This dataset represents a new category of biological data: organism-level, circuit-resolved causal genomics leading to novel targets and therapeutics. By combining scalable in vivo CRISPR perturbation with AI, we model biological systems at unprecedented resolution and simulate therapeutic interventions before committing to expensive downstream development. We’re grateful to our collaborators at @NVIDIAHealth and @10xGenomics for helping make this landmark dataset possible. Read More: perturb.ai/news #CRISPR #AI #DrugDiscovery #FunctionalGenomics #Biotech
English
7
77
364
60.8K
Joe Davison nag-retweet
Kimi.ai
Kimi.ai@Kimi_Moonshot·
Introducing 𝑨𝒕𝒕𝒆𝒏𝒕𝒊𝒐𝒏 𝑹𝒆𝒔𝒊𝒅𝒖𝒂𝒍𝒔: Rethinking depth-wise aggregation. Residual connections have long relied on fixed, uniform accumulation. Inspired by the duality of time and depth, we introduce Attention Residuals, replacing standard depth-wise recurrence with learned, input-dependent attention over preceding layers. 🔹 Enables networks to selectively retrieve past representations, naturally mitigating dilution and hidden-state growth. 🔹 Introduces Block AttnRes, partitioning layers into compressed blocks to make cross-layer attention practical at scale. 🔹 Serves as an efficient drop-in replacement, demonstrating a 1.25x compute advantage with negligible (<2%) inference latency overhead. 🔹 Validated on the Kimi Linear architecture (48B total, 3B activated parameters), delivering consistent downstream performance gains. 🔗Full report: github.com/MoonshotAI/Att…
Kimi.ai tweet media
English
334
2K
13.5K
5M
Joe Davison nag-retweet
Luca Soldaini 🎀
Luca Soldaini 🎀@soldni·
wait why Opus got a blog but not Golden Gate Claude
English
1
2
13
2K
Joe Davison
Joe Davison@joeddav·
@natolambert What’s your read on that Sonnet 4.5->4.6 leap? Genuine long context capability leap, narrow benchmark-specific phase change, or something else?
English
0
0
0
134
Joe Davison nag-retweet
Nathan Lambert
Nathan Lambert@natolambert·
World will converge on 3 types of models 1. Closed frontier (Ant, OAI, Gemini) 2. Open frontier (2-3 labs, much consolidation coming) 3. Open small / tool (fairly empty now) The open frontier will be far from the closed frontier, but way cheaper. Other statements are cope.
English
43
31
483
36.5K
Joe Davison nag-retweet
Omar Khattab
Omar Khattab@lateinteraction·
The Gemini Embedding 2 baseline here is.. 2 days old. Was just being celebrated and is now outperformed by a median of 14% and up to 91 points. If I didn't kind of know how powerful scaling ColBERTs and ColPalis can be compared to a single-vector model, I'd be in disbelief!
Omar Khattab tweet media
Ben Clavié@bclavie

I'm so excited to introduce this! We've worked on a million different moving parts to produce this. I'm fairly confident it's the best multimodal model that exists, period -- and it's not too shabby at pushing back the LIMITs of retrieval either...

English
13
46
692
91.2K
clem 🤗
clem 🤗@ClementDelangue·
I’m back. The girls and their superhero of a mom are doing great 😍😍😍 What did I miss?
clem 🤗@ClementDelangue

After almost 10 years of near nonstop grind, I’m taking 2 months of paternity leave to support my hero of a wife and welcome our twin daughters. @huggingface is in great hands with the team and @julien_c acting as interim CEO. Hope to return a changed man, to an even stronger HF, and to an AI field that’s more open and collaborative than ever!

English
99
8
1.1K
103.8K
Joe Davison nag-retweet
AT
AT@AliesTaha·
was skeptical but gave it a shot because @karpathy anyways 2x kernel perf (fp4 matmul) 3 minutes of work (1 prompt) triton beat cutlass (?!)
AT tweet media
Jaber@Akashi203

i open-sourced autokernel -- autoresearch for GPU kernels you give it any pytorch model. it profiles the model, finds the bottleneck kernels, writes triton replacements, and runs experiments overnight. edit one file, benchmark, keep or revert, repeat forever. same loop as @karpathy autoresearch, applied to kernel optimization 95 experiments. 18 TFLOPS → 187 TFLOPS. 1.31x vs cuBLAS. all autonomous 9 kernel types (matmul, flash attention, fused mlp, layernorm, rmsnorm, softmax, rope, cross entropy, reduce). amdahl's law decides what to optimize next. 5-stage correctness checks before any speedup counts the agent reads program.md (the "research org code"), edits kernel.py, runs bench.py, and either keeps or reverts. ~40 experiments/hour. ~320 overnight ships with self-contained GPT-2, LLaMA, and BERT definitions so you don't need the transformers library to get started github.com/RightNow-AI/au…

English
13
12
244
39.6K
Joe Davison nag-retweet
Will Knight
Will Knight@willknight·
Scoop from me: Nvidia will spend a total of $26 billion over the next five years building the world's best open source models. America is back in the open source AI race! wired.com/story/nvidia-i…
English
82
224
1.9K
689.4K
Joe Davison nag-retweet
Florian Brand
Florian Brand@xeophon·
time to look at your data, anon
Florian Brand tweet media
English
21
10
224
21.8K
Joe Davison nag-retweet
Dan Turner-Evans
Dan Turner-Evans@DanTurnerEvans·
I worked on the fly connectome for over 6 years, and let me just say that y’all have to slow this hype train way down. Connectomes are amazing. Biomechanical models are amazing. Linking the two is awesome. But scientists at the HHMI Janelia Research Campus, Princeton, and other institutes have been working on this for years now, and it’s not clear to me what’s new in the below. And connectomes are still missing a LOT of information. We’ve had the connectome of the worm for over 30 years now, and we still can’t reliably simulate a virtual worm. For example, connectomes don’t capture information about neuromodulator or neuropeptide release sites or receptors. These molecules are constantly changing the properties of neurons in the brain in ways that we have yet to really understand. And we don’t yet understand animal behavior well enough to refine and/or evaluate whole-brain simulations effectively. @AdamMarblestone and @doristsao already made many of these points, as well as many other good ones, but I just wanted to also add my two cents.
Dr. Alex Wissner-Gross@alexwg

x.com/i/article/2029…

English
71
298
2.3K
152.9K
Joe Davison nag-retweet
Andrej Karpathy
Andrej Karpathy@karpathy·
I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)
Andrej Karpathy tweet media
English
1.1K
3.6K
28.3K
11M