Alignment Lab AI

7.2K posts

Alignment Lab AI banner
Alignment Lab AI

Alignment Lab AI

@alignment_lab

Devoted to addressing alignment. We develop state of the art open sourced AI. https://t.co/oANsMnut7V https://t.co/6aJDLUvuU5

Your Digital Ecosystem Inscrit le Nisan 2023
4K Abonnements12.5K Abonnés
Alignment Lab AI retweeté
dr. jack morris
dr. jack morris@jxmnop·
Learning to write kernels might be the highest-ROI activity for displaced SWEs: → prereq: reasonable engineering ablity → six to twelve months of study → millions of dollars, mark zuckerberg showing up at your house to hire you, etc. i wish this were an exaggeration
English
43
62
1.9K
122.5K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
Particularly in terms of quantization of features to effective regimes, there's a very large amount of that explicitly operating on actual measurements of entropy unsupervised for the purpose of allowing a maximally efficient representation to emerge, because the computational substrate is itself still dominated by the entropy costs as a primary consideration
English
1
0
2
161
quetzal_rainbow
quetzal_rainbow@quetzal_rainbow·
The other problem with this paper is that discretization is treated here as black box which only mysterious "mapmaker" can do. But discretization happens constantly in nature. Sedimentation creates separate rock layers, cells are discretized by membranes, \
Séb Krier@sebkrier

An excellent paper for anyone interested in rigorous physicalist argument against computational functionalism. Alex is a fantastic, careful thinker and influenced my views a lot; we're working on a broader blog post breaking these concepts down, stay tuned! 🐙

English
7
2
40
3.9K
Vuk Rosić 武克
Vuk Rosić 武克@VukRosic99·
do you think most of the research is useless?
English
12
0
8
2.3K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
absolutely disagree, even if we stopped with just what we have now it would take years forr society and for the delpoyment of it to really be appreciable with the scale, the current stopgap is just how long it takes people to understand, not whats avaliable as currently known/extant implementation
English
0
0
0
94
Benjamin Todd
Benjamin Todd@ben_j_todd·
If AI progress stopped now, it would be a normal technology. One-off 5-10% productivity growth. Some routine white collar tasks automated. We chat to AI tools a lot. But no big economic or scientific acceleration. Ergo we don't have AGI.
English
55
6
182
22.9K
Alignment Lab AI retweeté
Mariusz Kurman
Mariusz Kurman@mkurman88·
Need more Claude, need more Codex, need more OpenCode or Pi? Gemini, Kimi? You got this
English
3
1
15
2.7K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
until i read this paper i was losing my mind not able to figure out wh this architecture i had was outperforming everything else so hard (fully constructing mostly reasonable sentences out of bytes in a few minutes at 5m parameters) after reading the paper and doing some analysis and ablations, its because i was using 768d model and 256 vocab (plus some other stuff to do with num params to dim) that avoided the bottleneck they mention almost entirely by acident
English
2
0
14
1.7K
bycloud
bycloud@bycloudai·
how big of a problem is this? > When backproping through the LM head, about 95-99% of the logit-gradient norm lies in directions that get projected away seems like the current workaround is just to use scaling to brute force it
bycloud tweet media
English
22
36
351
40.3K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
@pmddomingos if only performance were the only thing that kept it in the frontier labs
English
0
0
0
168
Alignment Lab AI
Alignment Lab AI@alignment_lab·
so it turns out the fast inv sqrt trick from Quake III Arena, (according to the internet from either or both of Greg Walsh and @ID_AA_Carmack ) entirely critical for some work im doing building linear models out of pretrained nonlinear ones. rmsnorm and softmax both would have gone unsolved if not for it. the unlock here is extremely op, im stoked
English
0
1
3
451
Alignment Lab AI
Alignment Lab AI@alignment_lab·
@nthngdy this is getting me so hard in the confirmation bias right now, this explains a ton!
English
0
0
0
163
Nathan Godey
Nathan Godey@nthngdy·
🧵New paper: "Lost in Backpropagation: The LM Head is a Gradient Bottleneck" The output layer of LLMs destroys 95-99% of your training signal during backpropagation, and this significantly slows down pretraining 👇
Nathan Godey tweet media
English
24
91
853
73.6K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
@sebkrier is this paper operating on the premise that what happens inside of a computer is *not* happening in reality/subject to thermodynamic constraints?
English
0
0
3
114
Alignment Lab AI
Alignment Lab AI@alignment_lab·
@sebkrier ive read this twice now, i dont get where it identifies which party is which and why, and what the delta is between a compression algorithm producing a codebook of class labels like a rANS, or me definitely learning language from my parents?
English
1
0
2
268
Séb Krier
Séb Krier@sebkrier·
An excellent paper for anyone interested in rigorous physicalist argument against computational functionalism. Alex is a fantastic, careful thinker and influenced my views a lot; we're working on a broader blog post breaking these concepts down, stay tuned! 🐙
Séb Krier tweet media
Alexander Lerchner@AlexLerchner

🧵1/4 The debate over AI sentience is caught in an "AI welfare trap." My new preprint argues computational functionalism rests on a category error: the Abstraction Fallacy. AI can simulate consciousness, but cannot instantiate it. philpapers.org/rec/LERTAF

English
47
44
519
56.3K
Alignment Lab AI
Alignment Lab AI@alignment_lab·
@sebkrier It's genuinely crazy, people have no idea how efficient the tech actually is, no one ever really considers what something like mohrs law running for so long actually means You can only double something so many times before it gets entirely out of hand
English
0
0
1
142
Séb Krier
Séb Krier@sebkrier·
Every day I notice inefficient processes that could be automated, yet won't be for a while bc of bureaucracy, legacy infra, misaligned incentives, inertia & status quo bias. Eventually competition forces it but it's so slow! "What could be, completely burdened by what has been."
English
14
8
107
9.8K
great auto
great auto@GreatA21473·
@alignment_lab @servamind Great lineup—interested in the computational cognition + encoding angle. Will catch the recording if available.
English
1
0
1
6
Alignment Lab AI retweeté
Servamind
Servamind@servamind·
New lecture drop 🎓 In our latest "Learning from Bio to AI" session, Andrew Coward explores Procedural Memory—how the brain learns skills and sequences. Join us tomorrow 7pm EST for a live Q&A on X Spaces to dig deeper. 🎥 Watch now
English
0
2
4
1.9K
attentionmech
attentionmech@attentionmech·
text and hilbert curve
English
3
1
42
2K