Robert Ritz

891 posts

Robert Ritz banner
Robert Ritz

Robert Ritz

@RobertERitz

American building in Mongolia. Education, manufacturing, and data.

Mongolia 가입일 Temmuz 2008
316 팔로잉2.4K 팔로워
@jason
@jason@Jason·
We started an AI founder twitter group... reply with "I'm in" if you're a founder and want to be added
English
3.8K
37
1.5K
190K
Matteo Franceschetti
Matteo Franceschetti@m_franceschetti·
BREAKING: @eightsleep is now available in mainland China. Our biggest expansion ever. A major company milestone. And a sign of where health technology is going. For more than a decade, we’ve engineered sleep technology that measures the body, adjusts the environment in real time, and improves sleep, not just tracks it. The future of health and longevity will not be built on data alone. It will be built on outcomes. That shift is happening around the world, and Eight Sleep is establishing itself as the leading global sleep brand. 300 million people in China suffer from sleep disorders. Average nightly sleep is 6.97 hours. The unmet need here is bigger than in any market we’ve entered. We’re excited for what’s ahead. Learn more at eightsleep.com/china
English
47
26
726
2.8M
Chris Albon
Chris Albon@chrisalbon·
Every month like clockwork
Chris Albon tweet media
English
1
0
6
1.3K
Robert Ritz
Robert Ritz@RobertERitz·
They will get there eventually. My guess is they want to be careful. Hilariously drive is a great place to work on files because even with accidental deletion they can be recovered. Having worked with the Google Drive api their read/write speeds are trash though. Like several seconds for a 10MB PDF.
English
0
0
0
67
Kevin Rose
Kevin Rose@kevinrose·
co-work feature request, let me select google drive/dropbox/etc. as a place to "work" on files. -- that way I can create on the fly vs relying dispatch (my laptop to be open/on)
English
37
3
160
25.1K
Robert Ritz
Robert Ritz@RobertERitz·
@bscholl Do you think members even believe it’s a problem? I’d bet about 40% think it’s not actually a problem. And with that 40% they make another 20-30% hold enough to screw everything up. And old people will vote out anyone who cuts their gravy train.
English
0
0
0
15
Robert Ritz
Robert Ritz@RobertERitz·
I think we humans are very “sparse” thinkers. Base models exhibit very creative and often random outputs. The RLHF models are much more coherent but also more boring and less creative. I think our cognition is somewhere in between, and in general we are much more sparse in our thinking than current neural networks which are comparatively very dense. This sparse vs dense is in reference to a vanilla neural network which is extremely dense. All information passes through every neuron. Current large state of the art models are mixtures do experts where portions of the model will be activated depending on the prompt. Humans are sparser even still, and our RLHF (our upbringing) is filled with a large variety of things that models are carefully shown only examples and rewarded to achieve desired behavior. Or the ultimate home schooled kid.
English
0
0
0
29
Zachary Schlosser
Zachary Schlosser@Zach_Schloss·
Yeah, I’m sympathetic to both those points, but I think it’s underestimating both the fact that the training data is vastly larger than the information any human studies and there’s likely to be a lot of patterns to mine in that much data we wouldn’t anticipate, plus the fact that a lot (not all) of science moves forward iteratively and by translating known insights from one domain into other domains, usually not new paradigmatic interpretations, and that we’re already seeing meaningful scientific discovery across multiple deeply technical fields. I work with folks that are all coming from different directions to argue that human cognition is demonstrably uniquely capable, which might turn out to be true, but I think there’s a lot of wishful and god of the gaps thinking going on.
English
2
0
0
28
Andy Hall
Andy Hall@ahall_research·
My dream for automated, replicable research: "papers" become open repos that anyone---agent or human---can fork and build on. But I have a funny story that suggests we've got some way to go to get there. In their fantastic paper, @xuyiqing and @YangYang_Leo show that my dream is approaching feasibility. For papers in journals with sufficiently strong data/code requirements, AI can now automatically replicate almost all of them. This is super exciting! But here's my story... In January, I released my Claude Code vote-by-mail paper, which got a lot of attention. As part of that release, I shared a public repo with the data and code (github.com/andybhall/vbm-…). Recently, I discovered to my great joy that the repo had 70 forks. Maybe my dream of open research was coming true?? So today I fired up Claude Code and asked it to analyze the 70 forks, so I could see what new insights we're learning. Here's Claude's summary: “Based on what I just investigated, the answer is simple: virtually none of them do anything.” It turns out, none of the 70 forks do any new research whatsoever. LOL. It seems like the tech is getting close---now we need to solve the incentive problem. We need to make it exciting and valuable from a career perspective for people to fork and build on projects. Then we can get real knowledge aggregation moving and really leverage AI's ability to let us do continuous, automatically replicable empirical work. We need to find a way to reward people for productive repo forks as a means to encouraging new and better research. Would love to figure out how best to do this!
Andy Hall tweet media
Yiqing Xu@xuyiqing

1/🧵 A major update to our paper: "Scaling Reproducibility" w/ @YangYang_Leo. We move beyond reanalyzing a single design to (almost) full-paper replication! Paper: bit.ly/repro-ai

English
12
28
165
31.6K
Robert Ritz
Robert Ritz@RobertERitz·
So there are two processes to training the models. The pretraining does have a lot more data than humans will ever see. This is where the model learns from pretty much all text data available. But the RLHF steers the model to a particular set of responses with the reward function. This is a curated set of data with a reward function that trains the model using conversation turns with their desired outputs. So the model we use is much more narrow than the base one. The data is all there, but it’s been trained to give responses in a certain way after the base pretraining.
English
1
0
0
22
Robert Ritz
Robert Ritz@RobertERitz·
On the ideation side it’s all about training data in the RLHF pipeline not having enough examples and rewards for that type of ideation. It’s also possible that the safety process neuters out good ideation for safety. In my own experience it’s pretty woke, and just shallow thinking. Not a political judgement but it’s just not useful to have ideological ideation in an academic context.
English
1
0
0
16
TravelGov
TravelGov@TravelGov·
Hong Kong: On March 23, 2026, the Hong Kong government changed the implementing rules relating to the National Security Law. It is now a criminal offense to refuse to give the Hong Kong police the passwords or decryption assistance to access all personal electronic devices including cellphones and laptops. This legal change applies to everyone, including U.S. citizens, in Hong Kong, arriving or just transiting Hong Kong International Airport. In addition, the Hong Kong government also has more authority to take and keep any personal devices, as evidence, that they claim are linked to national security offenses. Read more: hk.usconsulate.gov/security-alert…
TravelGov tweet media
English
487
2.7K
7.9K
3.5M
Robert Ritz
Robert Ritz@RobertERitz·
@Zach_Schloss @ahall_research @xuyiqing @YangYang_Leo I’m finding that Opus 4.6 can pretty easily do a full paper. Yes you have to go over it, but so do humans. What it’s NOT good at is ideation on impactful ideas. You get weird often biased ideas that are not generally useful.
English
1
0
0
36
Zachary Schlosser
Zachary Schlosser@Zach_Schloss·
When you say “exciting and valuable from a career perspective” do you mean paying human researchers to think, design and implment experiments and data analysis, and to rigorously and clearly communicate results? Concretely, what would it take for that to be economically compatible with automated, replicable research? Or are you biting the bullet that research will also become hyper unequal? Or do you anticipate some other career incentivization mechanism other than economic compensation?
English
1
0
0
120
Robert Ritz
Robert Ritz@RobertERitz·
This is just like open source software though. Lots and lots of forks with very little new insight or work. The biggest move is when an actor finds missing utility, a business case, or some sort of grant to build on the open work. Hilariously I’d also say the poli sci community places less value on this paper 1) because it was written by AI, and 2) it’s just out there flapping in the wind
English
0
0
0
89
Ian Reader
Ian Reader@IanReader4·
@RobertERitz @lydiahallie Hi, just asking, is that a good thing? as to protect you from unwanted changes when you forgot you had enabled the permission bypass before you closed your laptop?
English
1
0
0
13
Lydia Hallie ✨
Lydia Hallie ✨@lydiahallie·
Claude Code on desktop lets you select DOM elements directly, much easier than describing which component you want updated! Claude gets the tag, classes, key styles, surrounding HTML, and a cropped screenshot. React apps also get the source file, component name and props
English
194
297
4.7K
602.5K
Noah Zweben
Noah Zweben@noahzweben·
You can now schedule recurring cloud-based tasks on Claude Code. Set a repo (or repos), a schedule, and a prompt. Claude runs it via cloud infra on your schedule, so you don’t need to keep Claude Code running on your local machine.
English
298
563
7.5K
2M
Robert Ritz
Robert Ritz@RobertERitz·
@levelsio You can hit C on the keyboard and it copied it.
English
1
0
1
138
@levelsio
@levelsio@levelsio·
I do know copy pasting the big login URL via SSH might have some extra spaces and line breaks that MIGHT break it but maybe they should just be aware of that and remove those
English
4
0
27
8.3K
@levelsio
@levelsio@levelsio·
Every week at least one day I am stuck in some weird Claude Code logging me out somewhere and then their OAuth is completely broken and I get stuck and can't code
@levelsio tweet media@levelsio tweet media@levelsio tweet media
English
120
9
370
86.3K
Robert Ritz
Robert Ritz@RobertERitz·
@JaimeOrtega @trq212 Losing money faster could mean they have a funding crisis, which not only hurts that company but the entire AI wave right now.
English
0
0
0
38
JaimeOrtega
JaimeOrtega@JaimeOrtega·
@RobertERitz @trq212 making money -> losing money = bad losing money -> losing money = not bad, just neutral
English
1
0
0
423
Thariq
Thariq@trq212·
i think we might have undersold 1M context tbh, the performance is so so good, I really just don't clear the context window much these days
Thariq tweet media
English
260
68
2.1K
95.6K
Robert Ritz
Robert Ritz@RobertERitz·
@JaimeOrtega @trq212 So they are already likely losing money on the subscriptions? How could they make it cheaper?
English
1
0
0
441