Johnny Ni

12.1K posts

Johnny Ni banner
Johnny Ni

Johnny Ni

@JohnnyNi13

Vanguard Defense @Harvard - Prev. @northropgrumman @MITLL

LA Katılım Ocak 2021
4.6K Takip Edilen4.3K Takipçiler
Johnny Ni retweetledi
Sean Cai
Sean Cai@SeanZCai·
The infrastructure to make robotics data at scale useful is extremely neglected. Many different ego data vendors in the market right now selling piecewise ego datasets, without any sense of how to collaborate with end researchers on making it useful, or developing infra to collect more of it en masse once learning signal is discovered in any given format. What is diversity as a proxy of quality? Why is IMU data at all important? Is teleop or ego the way forward for robotics scaling (likely a combination of both)? What/when will the RL environments (GRPO Jan 2025) moment for robotics happen, such that a post-training format emerges that formalizes the robotics data market as the text-based reasoning data markets. have? That most data vendors in robotics data now have no particular sophistication on the research side, nor ideas on developing scalable infrastructure besides creative operational practices for overindexing on one fleeting modality, means that working on the above questions have extremely high EV.
Natasha Malpani 👁@natashamalpani

there is no hugging face for robotics data. no standardized pipeline for collecting, labeling, versioning, training on real-world robot data at scale. no tooling that handles contact dynamics and material deformation well enough for industrial manipulation. no teleoperation infrastructure where human supervisor intervention automatically becomes training data. no vertical-specific manipulation datasets for any specific industrial task. the actual bottleneck in physical AI is the data and the infrastructure to generate it. and this is a structural problem. for language AI, training data was the internet. abundant, cheap, already labeled by human intent. for robotics, the gap between where foundation models are and where they need to be cannot be closed by deploying more robots. three bets are being made right now: simulation-first works brilliantly for locomotion. domain randomization has essentially solved quadruped walking in unstructured terrain. but it breaks down completely for manipulation. simulated cameras have no noise, blur, or friction error. real cameras and grippers have all of it. cable insertion, fabric folding, dexterous assembly are exactly where simulation fails. teleoperation as data collection is the second move. deploy semi-autonomous robots, capture human-guided trajectories, iterate. theoretically sound. but the capital math is brutal and the execution evidence isn't there yet. human video as proxy is the third. if robots could learn from watching humans, you tap unlimited data. the problem: human hand geometry and force feedback don't map onto robot actuators. you're learning the shape of motion without the physics that make it work. what's actually working today is locomotion. narrow manipulation in structured environments. inspection and sensing. quadrupeds doing thermal inspection. no general-purpose manipulation required. the hardware race is loud, capital-intensive, winner-take-few. but the data infrastructure race is quiet, undercapitalized, wide open.

English
8
8
115
14.8K
Ian Zelbo
Ian Zelbo@ianzelbo·
If you can reply you’re cool and i wanna know you
English
380
2
456
43.8K
Johnny Ni retweetledi
Tommy
Tommy@thomasehendrix·
Defense-tech VC: “I didn’t see you at Hill and Valley this week…” Me: “I didn’t see you in Baghdad ever.”
English
65
41
1.1K
63.3K
Johnny Ni
Johnny Ni@JohnnyNi13·
Tokyo is now on the AI infra lab map
Sakana AI@SakanaAILabs

The AI Scientist: Towards Fully Automated AI Research, Now Published in Nature Nature: nature.com/articles/s4158… Blog: sakana.ai/ai-scientist-n… When we first introduced The AI Scientist, we shared an ambitious vision of an agent powered by foundation models capable of executing the entire machine learning research lifecycle. From inventing ideas and writing code to executing experiments and drafting the manuscript, the system demonstrated that end-to-end automation of the scientific process is possible. Soon after, we shared a historic update: the improved AI Scientist-v2 produced the first fully AI-generated paper to pass a rigorous human peer-review process. Today, we are happy to announce that “The AI Scientist: Towards Fully Automated AI Research,” our paper describing all of this work, along with fresh new insights, has been published in @Nature! This Nature publication consolidates these milestones and details the underlying foundation model orchestration. It also introduces our Automated Reviewer, which matches human review judgments and actually exceeds standard inter-human agreement. Crucially, by using this reviewer to grade papers generated by different foundation models, we discovered a clear scaling law of science. As the underlying foundation models improve, the quality of the generated scientific papers increases correspondingly. This implies that as compute costs decrease and model capabilities continue to exponentially increase, future versions of The AI Scientist will be substantially more capable. Building upon our previous open-source releases (github.com/SakanaAI/AI-Sc…), this open-access Nature publication comprehensively details our system's architecture, outlines several new scaling results, and discusses the promise and challenges of AI-generated science. This substantial milestone is the result of a close and fruitful collaboration between researchers at Sakana AI, the University of British Columbia (UBC) and the Vector Institute, and the University of Oxford. Congrats to the team! @_chris_lu_ @cong_ml @RobertTLange @_yutaroyamada @shengranhu @j_foerst @hardmaru @jeffclune

English
0
0
1
499
Christian Keil
Christian Keil@pronounced_kyle·
my body is a machine that turns Sweetgreens into deal flow
English
14
4
119
6.7K
Johnny Ni retweetledi
Max the VC 👨‍🚀
AI Game Theory Anthropic takes share fast. Then hits two walls: compute and pricing. Both cap how deeply it embeds into real workflows. For now. OpenAI plays its hand. What seemed like excessive compute commitments and irrational DRAM stockpiling in 2025 suddenly looks prescient. Enterprise usage snaps back on the back of cheaper, faster models. Not better in isolation per se, but on a per dollar, millisecond, and workflow basis. Their fate isn’t sealed, but they aren’t disappearing without a fight. Consulting and PE partnerships prove to be the real wedge into the enterprise at scale. They sell “AI transformation,” push OpenAI into the enterprise, and pipe real workflows back into “thinking” model training. Short term, consulting gets one last boom. Long term, they train the systems that replace them. Google keeps shipping quietly and relentlessly. Gemini gets bundled and subsidized. Distribution is prioritized over margins. It remains the lingering question mark and investor boogeyman that can break frontier economics at any moment. Their right to win remains. Meanwhile, routers quietly prove their value and become a core component of the stack. @OpenRouter and others become the meta layer. Best model per task fully abstracted away. Users stop caring (or knowing) who built what. Open source keeps closing the gap. Not frontier parity per se, but doesn’t matter. “Good enough” + free + customizable wins huge surface area. The other boogeyman. The frontier doesn’t stop, but it gets more expensive. Progress shifts from brute force scaling to systems design and algorithmic breakthroughs: memory, reasoning, and architectures we haven’t yet invented. Essentially, intelligence (i) becomes a function where i = p x c x a • p = power (energy, infra) • c = compute (chips, memory, throughput) • a = algorithms (efficiency, architecture, breakthroughs) As those inputs scale and diffuse, intelligence itself commoditizes and moats moves up the stack. Value concentrates in: • distribution • workflow ownership • proprietary data loops • embeddedness in decision making The goal isn’t the best model. It’s owning the outcome. At the same time, smaller, specialized models move on prem. Intelligence decentralizes. Competition comes from every direction, ultimately benefiting consumers. By 2030, 2026 looks primitive. The constraint is no longer compute. It’s imagination.
English
1
5
35
5.1K
Johnny Ni
Johnny Ni@JohnnyNi13·
Saw folks on the DC plane with Hill and valley merch already
English
0
0
4
284
Johnny Ni retweetledi
Catalyst
Catalyst@CatalystLabsX·
Introducing Catalyst, the agent layer for all of finance. Turn any natural language idea into a live strategy: research, backtesting & execution. Don’t get left behind. Waitlist open, join now.
English
121
157
1.2K
398.4K
Johnny Ni retweetledi
Nathan Lands
Nathan Lands@NathanLands·
The level of people supporting Lore is mindblowing Can't share everything yet, but we're building something that America genuinely needs right now
English
3
4
13
2.5K
Johnny Ni
Johnny Ni@JohnnyNi13·
The amount of secondary allocation offers I get from hardtech companies is bewildering
English
0
0
8
698
Johnny Ni retweetledi
Santiago Pliego
Santiago Pliego@SantiagoPliego·
Great to see friends win and win big.
English
0
6
34
2.3K
Johnny Ni retweetledi
delian
delian@zebulgar·
Excited to announce the next round of speakers for @HillValleyForum 2026 Obv my favorite is that I'll be doing a fireside chat w/ @NASAAdmin... Literally 15 year old me's dream come true.... & also... Jamie Dimon,@zoink,@friedberg, @jgebbia, Klaus Hommels, Thomas Laffont & more
delian tweet media
English
25
24
465
67.1K
Cyan Banister
Cyan Banister@cyantist·
The @LongJourneyVC team is heading to the Hill and Valley conference. We are hosting something fun - who will be there??
English
26
4
142
12.4K
Michelle Volz 🇺🇸🚀
Michelle Volz 🇺🇸🚀@MichelleVolz·
Excited to officially announce the launch of Pax Fund I, a $50M early stage vehicle dedicated to founders transforming the foundational categories of society. 🇺🇸🚀 Wrote a bit about my journey and the thinking behind Pax below:
English
156
42
871
99.7K