Bread

11 posts

Bread banner
Bread

Bread

@ai_bread

We bake things into LLMs; founded by @abhargava2000 & @witkowski_cam

San Francisco, CA Joined Aralık 2024
21 Following1K Followers
Bread
Bread@ai_bread·
Bread is hiring! We're looking for two exceptional individuals to join our team in SF / Bay Area: Developer Relations: → Build a vibrant developer ecosystem → Create killer tutorials/docs → Python, ML, and OSS familiarity → Attend conferences, speak publicly, and be an evangelist for our mission. → 3+ years experience in a similar role. Machine Learning Engineer → Run experiments + bake models → Develop good recipes & share best practices → Python/PyTorch/huggingface → Applied experience Prompt Engineering → 2+ years experience training LLMs If you want to give people greater control over their AI by making them learn more like humans, click the link below to learn more!
English
1
1
11
1.2K
Bread retweeted
Cameron Witkowski
Cameron Witkowski@witkowski_cam·
"We really care about being seen positively by society. We care about being in good standing. All these social intuitions that we have, I feel strongly that they're baked in."
English
3
2
26
4.2K
Bread
Bread@ai_bread·
Cameron Witkowski@witkowski_cam

Had not heard of context distillation when we wrote the paper back in 2024 but this is great stuff & way ahead of its time! Our initial paper showed us that prompts could in principle be converted into weight updates — and surprisingly fast with new advances like LoRA, and increasingly capable open source LLMs. We started Bread to take the next step: making LLMs learn at human speed, autonomously, from the real world, in the real world. That involves - Taking the same data from inference time - Using it to structure coherent weight updates repeatedly (like humans learning new concepts). - Doing that usably and reliably at scale (distillation to + from many model versions) Technically, that involves solving 1. Ingestion problem: how does incremental inference time data update the configuration of the bake? 2. Digestion problem: how do you structure updates to retain model coherence? 3. Learning infra problem: How do you build a fast training platform which depends upon a rapidly expanding repertoire of new models for on-policy updates? 4. Serving problem: how do you take advantage of GPU + bandwidth resources to serve inference & hot-swap new models created throughout inference time? The first research paper only put us at the starting line, and our work at Bread is running the marathon. Please check out the papers in this thread to see some early precursors to Baking! (wish we had known to cite them in the original paper). It’s satisfying validation to see others recognizing the value in converting prompts to weight updates, and I celebrate these prior works from Askell and Snell which bolster our commitment to and confidence in our mission.

QME
0
0
4
843
Bread
Bread@ai_bread·
Why Prompt Baking is the only known method for sample-efficient, on-the-job learning. Our new blog post explores why no other method can achieve the same sample efficiency and composability as baking in prompts, and why we’re placing our long-term bets on this approach for solving human-like on-the-job learning.
English
4
3
19
6.6K
Bread
Bread@ai_bread·
Our round also included participation from Golden Ventures, Northside Ventures, and a number of angels. None of this would have been possible without our early employees’ extremely hard work and monumental technical lifts. Special thanks to Alexander Detkov & Prof. Matt Thomson at Caltech for their support in the early research that led to Bread!
English
2
0
30
27.6K
Bread
Bread@ai_bread·
Announcing Bread Technologies. We’re building machines that learn like humans. We raised a $5 million seed round led by Menlo Ventures and have been building in stealth for 10 months. Today, we rise 🍞
English
128
129
931
761.4K
Bread
Bread@ai_bread·
Happy SF tech week
English
2
0
10
2.8K