Joscha Bach

67.6K posts

Joscha Bach banner
Joscha Bach

Joscha Bach

@Plinz

San Francisco, CA เข้าร่วม Nisan 2009
789 กำลังติดตาม158.3K ผู้ติดตาม
Joscha Bach
Joscha Bach@Plinz·
RT @chris_percy: I'm looking forward to joining @Plinz and many others at this conference - May in the Bay! I'll be talking about differen…
English
0
1
0
68
Joscha Bach รีทวีตแล้ว
T Wolf 🌁
T Wolf 🌁@Twolfrecovery·
We're living in the upside down when a U.S. Senator is working with the CCP to regulate AI in the U.S.
T Wolf 🌁 tweet media
English
49
228
1K
150K
Joscha Bach รีทวีตแล้ว
Daniel Jeffries
Daniel Jeffries@Dan_Jeffries1·
AI has had one safest technology roll-outs in history. Read that again, because it's a fact. It's used by billions with a tiny fraction of a percent of actual problems. And yet it's seen as dangerous or unsafe by many. There's a constant chorus of people shouting about its
English
33
14
140
26.7K
Joscha Bach รีทวีตแล้ว
Rebecca Wang
Rebecca Wang@rbccawang·
agi-timeline gap relationship
English
9
20
361
25.6K
Joscha Bach
Joscha Bach@Plinz·
@realantonmaier It makes sense, but there is no obvious reason why models could not be prompted into repairing their taste
English
1
0
1
201
anton
anton@realantonmaier·
@Plinz If it didn't think their style was better, they wouldn't use it.
English
1
0
1
195
Daniel Jeffries
Daniel Jeffries@Dan_Jeffries1·
Never let critical thinking get in the way of extremist fantasy stories. If you're on either side of this slider, congratulations, you may be a child in an adult costume body! Black and white thinking is for children. When you become an adult it's time to put away childish things.
Ethan Mollick@emollick

English
34
24
232
49.1K
Joscha Bach
Joscha Bach@Plinz·
our timeline has only llms because the resistance sent someone back in time to prevent the birth of everyone who was going to build actual ai
English
44
14
320
17.7K
Joscha Bach
Joscha Bach@Plinz·
@Dan_Jeffries1 I just learned that "AI cannot understand things and cannot feel pain, because it does not have mirror neurons"
English
12
0
28
2.6K
Daniel Jeffries
Daniel Jeffries@Dan_Jeffries1·
Never let actual understanding of AI get in the way of good populist one liner for idiots.
GDP@bookwormengr

Let us settle this distillation debate with science. ============= Can anyone share any papers that show distillation based on text tokens alone (which is glorified fine tuning) - in absence of log probabilities, because frontier labs don't share them - can produce near frontier performance? Let me know if you found any. I haven't seen any. Why you are not likely to find such papers: ------------------- If you try to distill/fine tune based on text outputs from another model, it is "off policy" update - i.e. fine tuning based on tokens produced by another models. It is a well known fact that such approaches degrade model performance in unexpected ways. Distillation in such cases can provide style, but not 'wisdom/capabiltiies' to describe in simple terms. The claim here is that the malicious labs are acquiring 'wisdom/capabilities' through distillation. Where distillation actually works: ------------------- When frontiers labs distill their large models into smaller models (the way DeepSeek and ZAI also do, in addition to OpenAI, Ant, GDM), they use token probability distribution and not text output alone. The student model's objective is to match teacher's probability distribution at each token position. Further more, now 'on policy distillation' is in vogue. You can read more here about one such approach thinkingmachines.ai/blog/on-policy… where in text is generated by the student model, and token probabilities are extracted from the teacher model for the same set of tokens. However, such token probabilities are not available from APIs of closed source labs. So, for distillation by malicious labs - with intent to acquire "wisdom/capabilities" - to be successful there must have been some scientific breakthrough that I have missed that enables them to distill without log probabilities. Can anyone please enlighten me and share relevant papers? I haven't found any. If such literature does not exist, then this debate is meaningless. Caveats: --------- 1. I myself have distilled from open source reasoning models to open source non-reasoning model with text only, i.e. without log probabilities. There was a slight boost in performance in one area and degradation in performance on many other areas. 2. Text only distillation can still help to get a few good agentic trajectories to seed that behaviour in the model before conducting further re-inforcement learning (RL). However, here model still acquire style only. The "wisdom/capabilities" in this approach still comes reinforcement learning training done by the labs. 3. Capable closed source models could be used during RL training as a judge (Generative Reward). However, they can be very expensive and labs have figured they can use smaller more effective models as a judge. @teortaxesTex

English
4
7
36
8.6K
Joscha Bach
Joscha Bach@Plinz·
@_chaosmos_ Peter Watts is an absolute favorite of mine (the Rifters are brilliant, too!). Ted Chiang is great. Greg Egan! Neal Stephenson is becoming more hard science (eg. Seveneves), but I love his early, more playful works better.
English
0
0
1
113
сhaosmos
сhaosmos@_chaosmos_·
@Plinz That sounds really intriguing and interesting! Thanks for the recommendation! I’m really curious to know which hard science fiction books and authors are your favourites? Have you read Blindsight and Echopraxia by Ritter Watts?
English
1
0
1
102
Joscha Bach
Joscha Bach@Plinz·
Alastair Reynold's "Revelation Space" universe is certainly not the hardest SF out there, but it's hard to beat his scope. He takes the largest ideas in SF and zooms so far out that they become tiny details in his stories.
English
10
5
94
9.7K
Joscha Bach
Joscha Bach@Plinz·
Academia is crucial to the survival of our civilization, but without effective external oversight and pruning, it evidently succumbs to the corrupting forces of entropy, declining peer quality and self serving ideology.
English
8
7
53
2.9K
Joscha Bach
Joscha Bach@Plinz·
Mentioning this is considered impolite. Sometimes, entire disciplines become infertile but are kept alive like endangered species, or they get cordycepted by political movements that wear the shell of the dead host to legitimize academic status and funding.
English
6
2
61
3.3K
Joscha Bach
Joscha Bach@Plinz·
An interesting aspect of academic paradigms that is that many of them are dead ends, made obsolete by intellectual or practical breakthroughs in different areas of intellectual inquiry, or due to internal exhaustion or ideological corruption.
English
11
10
97
6.4K
Joscha Bach
Joscha Bach@Plinz·
@bokuHaruyaHaru I reject the notion that factual statistical statements about humanity constitute normative claims and are furthermore impermissible on moral grounds, but I freely offer my contempt to you if helps you
English
1
0
2
78
Haru Haruya (春夜 ハル)
Haru Haruya (春夜 ハル)@bokuHaruyaHaru·
“Recursive joke” is doing a lot of cleanup work here. You phrased it like an actual cynical argument about human unreliability, got a response to the argument-shaped thing you posted, and then added “cognitively useless” on top. That’s not just clarification. That’s contempt with an escape hatch. 😉
English
1
0
3
64
Joscha Bach
Joscha Bach@Plinz·
If you want to get a taste, I recommend the Revelation Space short stories collection. You can also read Chasm City if you dig the noir B movie style of Altered Carbon, or the first book of the Inhibitor trilogy (also called Revelation Space) for the full-on experience.
English
2
1
25
2.5K
Joscha Bach
Joscha Bach@Plinz·
Human hive minds? Sure, Conjoiners have scaled brain implants into transenlightenment since the early Mars settlements. Hyperreligion? Of course, lots of people get infected by the various strains of engineered indoctrinal viruses these days.
English
1
1
10
2.6K