Samson

2.6K posts

Samson banner
Samson

Samson

@SamsonDota

life enjoyer pos 1 https://t.co/865MFeWWM9

Denver Beigetreten Şubat 2013
324 Folgt3.8K Follower
Samson retweetet
Andrej Karpathy
Andrej Karpathy@karpathy·
There was a nice time where researchers talked about various ideas quite openly on twitter. (before they disappeared into the gold mines :)). My guess is that you can get quite far even in the current paradigm by introducing a number of memory ops as "tools" and throwing them into the mix in RL. E.g. current compaction and memory implementations are crappy, first, early examples that were somewhat bolted on, but both can be fairly easily generalized and made part of the optimization as just another tool during RL. That said neither of these is fully satisfying because clearly people are capable of some weight-based updates (my personal suspicion - mostly during sleep). So there should be even more room for more exotic approaches for long-term memory that do change the weights, but exactly - the details are not obvious. This is a lot more exciting, but also more into the realm of research outside of the established prod stack.
Awni Hannun@awnihannun

I've been thinking a bit about continual learning recently, especially as it relates to long-running agents (and running a few toy experiments with MLX). The status quo of prompt compaction coupled with recursive sub-agents is actually remarkably effective. Seems like we can go pretty far with this. (Prompt compaction = when the context window gets close to full, model generates a shorter summary, then start from scratch using the summary. Recursive sub-agents = decompose tasks into smaller tasks to deal with finite context windows) Recursive sub-agents will probably always be useful. But prompt compaction seems like a bit of an inefficient (though highly effective) hack. The are two other alternatives I know of 1. online fine-tuning and 2. memory based techniques. Online fine-tuning: train some LoRA adapters on data the model encounters during deployment. I'm less bullish on this in general. Aside from the engineering challenges of deploying custom models / adapters for each use case / user there are a some fundamental issues: - Online fine-tuning is inherently unstable. If you train on data in the target domain you can catastrophically destroy capabilities that you don't target. One way around this is to keep a mixed dataset with the new and the old. But this gets pretty complicated pretty quickly. - What does the data even look like for online fine tuning? Do you generate Q/A pairs based on the target domain to train the model? You also have the problem prioritizing information in the data mixture given finite capacity. Memory based techniques: basically a policy for keeping useful memory around and discarding what is not needed. This feels much more like how humans retain information: "use it or lose it". You only need a few things for this to work: - An eviction/retention policy. Something like "keep a memory if it has been accessed at least once in the last 10k tokens". - The policy needs to be efficiently computable - A place for the model to store and access long-term memory. Maybe a sparsely accessed KV cache would be sufficient. But for efficient access to a large memory a hierarchical data structure might be beter.

English
275
299
4.6K
586.4K
Samson
Samson@SamsonDota·
@astinidota glad to see hes doing so well, thanks for remembering I posted this astini
English
0
0
1
247
Samson
Samson@SamsonDota·
@ninajirachi you wouldn't happen to have 2 extra tickets to Sundays show would you? am a big fan but the resale a little out of my price range. thought it's worth asking since you been replying.
English
1
0
1
407
Samson
Samson@SamsonDota·
Recently bought a Tesla and use the self driving occasionally. It is super impressive in some ways and has a long way to go for others but if used in combination with manual driving and with focus it's unequivocally the safest way to drive. Maybe LiDAR is the future of this tech but Teslas camera only is very impressive.
English
0
0
0
246
Andrej Karpathy
Andrej Karpathy@karpathy·
I am unreasonably excited about self-driving. It will be the first technology in many decades to visibly terraform outdoor physical spaces and way of life. Less parked cars. Less parking lots. Much greater safety for people in and out of cars. Less noise pollution. More space reclaimed for humans. Human brain cycles and attention capital freed up from “lane following” to other pursuits. Cheaper, faster, programmable delivery of physical items and goods. It won’t happen overnight but there will be the era before and the era after.
English
793
2K
21.9K
1.5M
Samson
Samson@SamsonDota·
@LyricalDota Channel 5 has some of the best coverage, period.
English
0
0
1
157
Samson retweetet
signüll
signüll@signulll·
karpathy speaks like someone who’s running a mental compiler in real time with minimal interpretive latency & almost zero runtime garbage. he’s not verbose. he just threads complexity into compressed lossless statements. most smart people can be dense, but they lose clarity. karpathy keeps clarity while cranking the bitrate like an llm tuned with perfect temperature control. it’s so beautiful to listen to that i had to hear this three times already. i might add a fourth.
Dwarkesh Patel@dwarkesh_sp

The @karpathy interview 0:00:00 – AGI is still a decade away 0:30:33 – LLM cognitive deficits 0:40:53 – RL is terrible 0:50:26 – How do humans learn? 1:07:13 – AGI will blend into 2% GDP growth 1:18:24 – ASI 1:33:38 – Evolution of intelligence & culture 1:43:43 - Why self driving took so long 1:57:08 - Future of education Look up Dwarkesh Podcast on YouTube, Apple Podcasts, Spotify, etc. Enjoy!

English
145
330
6K
824.3K
Samson retweetet
Dwarkesh Patel
Dwarkesh Patel@dwarkesh_sp·
The @karpathy interview 0:00:00 – AGI is still a decade away 0:30:33 – LLM cognitive deficits 0:40:53 – RL is terrible 0:50:26 – How do humans learn? 1:07:13 – AGI will blend into 2% GDP growth 1:18:24 – ASI 1:33:38 – Evolution of intelligence & culture 1:43:43 - Why self driving took so long 1:57:08 - Future of education Look up Dwarkesh Podcast on YouTube, Apple Podcasts, Spotify, etc. Enjoy!
English
540
2.9K
18.7K
10.7M
SabeRLighT-
SabeRLighT-@jonas_volek·
1/3 In the past Valve made all the content for Battlepass + the prize pools were already super high so it made complete sense that only 25% of BP purchases went towards the prize pool. Now however, except for stickers its the teams that create all the content.
English
13
12
353
33.4K
rouri404
rouri404@rouri404·
MISS IT OUT EVERYWHERE LINK BELOWWW
rouri404 tweet media
English
2
9
44
1.3K
Samson retweetet
Andrej Karpathy
Andrej Karpathy@karpathy·
@lexfridman 100%, it's very interesting to me when a lot value is created but the awareness/esteem of it, the story of it, or the people involved is next to zero. It's not fully true but it's sufficiently true.
English
58
40
2.5K
106.3K
Samson
Samson@SamsonDota·
@Griz just sat next to you at Cart Driver Rino but didnt realize fucking cheers mate
English
0
0
0
43
Jesshrac 🐐
Jesshrac 🐐@jessjenius·
my local shops was SOLD OUT OF @redbullau wtf i had to buy a v blue instead ☠️
Jesshrac 🐐 tweet media
English
2
0
4
2.3K
Samson retweetet
Davai
Davai@Davai_Dota·
Super happy for the oppurtunity to proof myself, and especially to represent NA once again! Shopify's been amazing to work with so far. Time to make good things happen.
Shopify Rebellion@ShopifyRebels

Getting our reps in across the pond with @fissure_en Universe Ep. 5 Play-in! @DavaiLama will be joining us as our Offlaner for ongoing and upcoming events. Welcome! DAY 2 - Group Stage 10:00 CEST - SR 🆚 One Move 16:00 CEST - SR 🆚 @Wildcard_GG #ArmTheRebels

English
4
2
85
4.8K
Samson
Samson@SamsonDota·
@Skrillex thanks for giving these 500 tickets to the people, much better to wait in an irl line compared to rng online sweepstakes
English
1
0
0
82
Nahaz
Nahaz@NahazDota·
@Noxville Can you please elaborate on what you mean?
English
3
0
6
1.1K
Ben Steenhuisen
Ben Steenhuisen@Noxville·
At least for the short term, I would suggest no pro teams reveal/experiment with anything very interesting in their scrims, unless they're okay with that being leaked.
Berlin, Germany 🇩🇪 English
10
3
80
15K
ixmike88
ixmike88@ixmike88·
pistons are such chokers jfc
English
2
0
8
2.6K
Quinn Callahan
Quinn Callahan@ccncdota2·
Highly recommend the Pitt to anyone looking for a new show to watch. My gosh it's good
English
6
3
118
13K