David Andrzejewski

1.1K posts

David Andrzejewski banner
David Andrzejewski

David Andrzejewski

@davidandrzej

Software! Systems! Data! @SFMachineLearn Elsewhere: @[email protected] @davidandrzej.bsky.social

San Francisco Katılım Eylül 2008
1.4K Takip Edilen2.3K Takipçiler
Sabitlenmiş Tweet
David Andrzejewski
David Andrzejewski@davidandrzej·
Proposal to re-brand "deep learning" as "stochastic regularized estimation of compositionally structured nonlinear functions".
English
5
25
76
0
David Andrzejewski retweetledi
Jason Alan Fries
Jason Alan Fries@jasonafries·
🚀 Checkout our ICLR 2024 Spotlight Poster #165 today (Session 3) 🩺 "MOTOR: A Time-to-Event Foundation Model For Structured Medical Records" ✨ Highlights: - The first TTE foundation model for structured EHRs - Improves SOTA TTE by 4.6% & boosts label efficiency up to 95% - TTE pretraining scales to 16k tasks & reduces GPU memory usage by ~35% - Model weights available for research use! 🔍 Tutorial: github.com/som-shahlab/mo… 🤖 Model: huggingface.co/StanfordShahLa… #ICLR2024 #AI #Healthcare
Jason Alan Fries tweet media
English
1
11
49
4.4K
David Andrzejewski retweetledi
Korl
Korl@Korl_co·
Korl is available for public beta! We’ve built a platform that auto-generates consumable product presentations in seconds, each optimized for a common use case and audience. Start with a 2-week free trial from our site: korl.co
English
1
2
14
1.9K
David Andrzejewski
David Andrzejewski@davidandrzej·
Fun stuff - "In-Context Learning Creates Task Vectors" (via @_akhaliq) isolates something (more or less) resembling the "program key" described below as an internal attention state, demonstrating with state substitution/patching experiments huggingface.co/papers/2310.15…
François Chollet@fchollet

My interpretation of prompt engineering is this: 1. A LLM is a repository of many (millions) of vector programs mined from human-generated data, learned implicitly as a by-product of language compression. A "vector program" is just a very non-linear function that maps part of the latent space unto itself. 2. When you're prompting, you're fetching one of these programs and running it on an input -- part of your prompt serves as a kind of "program key" (as in database key) and part serves as program argument(s). Like, in "write this paragraph in the style of Shakespeare: {my paragraph}", the part "write this paragraph in the stye of X: Y" is a program key, with arguments X=Shakespeare and Y={my paragraph}. 3. The program fetched by your key may or may not work well for the task at hand. There's no reason why it should be optimal. There are lots of related programs to choose from. 4. Prompt engineering represents a search over many keys in order a find a program that is empirically more accurate for what you're trying to do. It's no different than trying different keywords when searching for a Python library. 5. Everything else is unnecessary anthropomorphism on the part of the prompter. You're not talking to a human who understands language the way you do. Stop pretending you are.

English
0
1
1
640
David Andrzejewski
David Andrzejewski@davidandrzej·
Frequent SF (Nob Hill) sighting recently: astonished & delighted tourists taking photos/videos of the self-driving cars.
English
0
0
0
230
David Andrzejewski
David Andrzejewski@davidandrzej·
Paper itself is an absolute treat: great diagrams, abundant code samples, and as a nice bonus some historical context around the origins and development of the research ideas. Credit: saw it in @deliprao's "AI Research & Strategy" newsletter deliprao.substack.com
David Andrzejewski tweet media
English
0
0
6
0
David Andrzejewski
David Andrzejewski@davidandrzej·
...furthermore: outperforms XGBoost, does Lasso in one-pass, seems not to rely on nearest-neighbor. Future work: "...an intriguing possibility where we might be able to reverse engineer the Transformer to obtain better learning algorithms." (!)
English
2
1
25
0
David Andrzejewski
David Andrzejewski@davidandrzej·
Easy to get "breakthrough fatigue" in ML recently, but in this work the NN learns *how to learn* linear regression, decision trees, 2-layer ReLU nets 😲 "...we train Transformer models to discover algorithms for different learning problems." arxiv.org/abs/2208.01066
English
8
68
448
0
David Andrzejewski
David Andrzejewski@davidandrzej·
That said, it worked! Big thanks to all.
English
0
0
1
0
David Andrzejewski
David Andrzejewski@davidandrzej·
ML on the M1 has magically transported me back to my youth: a hellscape of Python toolchain chaos with insane fixes like “find a random .py file in your /lib and swap the order of two import statements” #issuecomment-975178763" target="_blank" rel="nofollow noopener">github.com/tensorflow/ten…
English
1
2
3
0
David Andrzejewski
David Andrzejewski@davidandrzej·
@hkarthik @Carnage4Life is there an innocuous technical / data reason why (anecdotally) users overwhelmingly experience asymmetric “slippage” (ie, longer waits) in the predictions?
English
1
0
0
0
Karthik Hariharan
Karthik Hariharan@hkarthik·
@Carnage4Life We have 7 data scientists and over 10 engineers working on predictions and dynamic pricing, the second of which is absolutely necessary to load balance the supply issues of delivery drivers.
English
1
0
2
0
David Carlton
David Carlton@davidcarlton·
Or why, in pop music, do we grudgingly accept remaking songs but make sure to label it as a cover, whereas it’s just the norm in classical music and play the same pieces over and over again?
English
2
0
2
0
David Carlton
David Carlton@davidcarlton·
It’s kind of weird which categories of art are ones where we embrace frequent remakes and which are ones where we don’t. Like, why are people suspicious of the concept of remaking a TV series but we perform the same plays over and over?
English
1
0
2
0
David Andrzejewski
David Andrzejewski@davidandrzej·
Pondering the TBs of data crunched in an ultramodern cloud ML platform in order to power cutting-edge causal models targeting tightly business-aligned KPIs in one of the world’s most esteemed tech firms, all leading up to this email.
David Andrzejewski tweet media
English
2
0
2
0