Amund Tveit

240 posts

Amund Tveit banner
Amund Tveit

Amund Tveit

@atveit

Principal AI Engineer | Personal Opinions Only

Trondheim, Norway Katılım Mayıs 2007
4.6K Takip Edilen2.2K Takipçiler
Rohit Durvasula
Rohit Durvasula@0xrodu·
Have often seen Jeff mention Sanjay. Didn't remember reading anything in particular from him or ever coming across an interview from him. Fell into a rabbit hole of searching for content on him and 15 minutes later I still can't find anything. This guy is invisible!
Jeff Dean@JeffDean

Performance Hints Over the years, my colleague Sanjay Ghemawat and I have done a fair bit of diving into performance tuning of various pieces of code. We wrote an internal Performance Hints document a couple of years ago as a way of identifying some general principles and we've recently published a version of it externally. We'd love any feedback you might have! Read the full doc at: abseil.io/fast/hints.html

Seven Trees, CA 🇺🇸 English
17
0
94
40.7K
Amund Tveit
Amund Tveit@atveit·
4 bits is all you need (and 3.6 bit you have?) for resource-efficient LLMs? A few months ago OpenAI published their open weights model(s) GPT-OSS (20B and 120B), and one of the eye-catching characteristics was that it was heavily quantized - in other words “shrunk” or compressed - to 4 bits per parameter (MXFP4) instead of the commonly used 16 bit (BF16). 4 bit (“half byte”) means that the model uses much less memory, energy and also can run efficiently on hardware natively supporting 4 bit MXFP inference (e.g. Nvidia Blackwell B200/300 and AMD MI355X). With 4 bit compared to 16 bit there is some quality loss, but typically marginally. But can we go even lower and represent the model in fewer bits per parameter, e.g. 3 bit, 2 bit or even 1 bit? There is highly interesting and seemingly promising research on more efficient quantization (few bit representation) of deep neural networks, however perhaps 4 bit is very close to the lower plateau - based on: a) findings in an article with large empirical studies by Meta, Google DeepMind, Cornell University and Nvidia that language models have a capacity of 3.6 bits per parameter - Language Models Don’t Store Facts: They Compress Patterns - so perhaps 3.6 bit is a lower bound per parameter for efficient representation of deep neural networks? b) and to a lesser degree personal testing on doing further quantizations on GPT-OSS 20B, it seems like when going to 2 and 3 bit even with (LoRA-based) finetuning it is hard to lift it back to close to 4 bit. Amund atsentia.com
English
0
0
0
483
Amund Tveit
Amund Tveit@atveit·
AI on Apple Watch
English
0
0
2
247
Simon Willison
Simon Willison@simonw·
What's the latest research on how much baked-in knowledge an LLM needs in order to be useful? If I want a specialist coding model can I trim the size of that model down by stripping out detailed knowledge of human history, geography etc? Can we even do that?
English
35
9
193
32K
Amund Tveit
Amund Tveit@atveit·
@xamat Hope to get complementary byte-based grokking results scrutinized and published soonish :)
English
0
0
0
47
Amund Tveit
Amund Tveit@atveit·
@jobergum Congratulations! Curious to learn more :) perhaps 2025 is the year of entrepreneurship..? 👀
English
0
0
5
139
Amund Tveit
Amund Tveit@atveit·
@jobergum yes, quite the jump with reasoning models from O1 and onwards to GPT-5. However, believe many are a bit behind in learning prompting/context engineering skills.
English
0
0
2
62
Amund Tveit
Amund Tveit@atveit·
Exciting times in AI (Agent) devtools and vibe coding — and for the historically inclined, you might enjoy a 2008 throwback: my blog post Increase the automation of Test-Driven Development?amundblog.blogspot.com/2008/01/increa…
English
2
0
7
1.9K