M

19.6K posts

M banner
M

M

@init_malachi

continually learning in a state of delight | ex sr member of technical staff | interested in ai epistemology

synthesis Katılım Mayıs 2022
2.9K Takip Edilen939 Takipçiler
M retweetledi
mattparlmer 🪐 🌷
mattparlmer 🪐 🌷@mattparlmer·
Robotics of this type is a solved problem to the degree that undergrads will be putting stuff that beats anything they’ve demo’d here up on HuggingFace by the end of the year
Skild AI@SkildAI

Robotics is a data problem. Today, we’re partnering with @ABBRobotics, @Universal_Robot, and @NVIDIARobotics to deploy the Skild Brain across real-world industries from manufacturing to factory lines. This will help us build the world’s biggest data flywheel for physical AI.

English
9
17
287
23.2K
M
M@init_malachi·
get its activations in context without spewing a single rambly token. no more observing by talking
English
0
0
0
9
M
M@init_malachi·
they should invent a cantrip but for shutting the f up
English
1
0
0
17
M retweetledi
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞)
> MDM-Prime-v2 is 21.8× more compute-efficient than autoregressive models I may be humiliated extremely hard with my diffusionLM skepticism.
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) tweet mediaTeortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) tweet media
You Jiacheng@YouJiacheng

HUGE if true. If true, this is probably a larger efficiency gain than ALL publicly available techniques since DeepSeekMoE(Jan 2024) COMBINED. And it can just win modded-nanogpt speedrun. (1e18 is 250s@50%MFU, but the loss is significantly lower than 3.28) cc @classiclarryd

English
8
5
83
15.2K
M retweetledi
anand iyer
anand iyer@ai·
This feels like physical product design's ChatGPT moment. This team just ran an autonomous agent against the entire chip design process: 219-word spec in, tape-out-ready silicon layout out, 12 hours later. The agent ran continuously against a simulator, found its own bugs, rewrote its own pipeline, and iterated to a working CPU! Chip design costs well over $400M and takes up to 9 years. Not because writing hardware code is hard (it is actually brutally hard) but because a respin costs 10 of millions. So teams spend more than half their total budget just verifying the design is correct before a single transistor is placed. That cost structure is why most chip designs never get built. Entire product categories that were previously too low-volume to justify a tape-out are now buildable.
Towaki Takikawa / 瀧川永遠希@yongyuanxi

Design Conductor: an AI agent that can build a RISC-V CPU core from design specs. The agent is given access to a RISC-V ISA simulator and manuals... to enable an end-to-end verification-driven generation. The most important thing for design intelligence is a verifier 😎

English
10
36
253
32.7K
M retweetledi
Felix Rieseberg
Felix Rieseberg@felixrieseberg·
A small ship I love: We made Claude.ai and our desktop apps meaningful faster this week. We moved our architecture from SSR to a static @vite_js & @tan_stack router setup that we can serve straight from workers at the edge. Time to first byte is down 65% at p75, prompts show up 50% sooner, navigation is snappier. We're not done (not even close!) but we care and we'll keep chipping away. Aiming to make Claude a little better every day.
English
71
68
1.7K
213.8K
M retweetledi
Dylan Garcia
Dylan Garcia@_dylanga·
The first thing I did at @tryramp was set up distributed tracing, structured logging, and metrics for Inspect, our background coding agent. We now have full visibility in to everything the system is doing: the browser, CF workers/DOs, @modal sandboxes, database calls, etc. Most importantly, Inspect now has visibility in to itself. It can self-triage runtime errors it encounters and create PRs to fix them. Every morning, it reviews the past 24 hours of its own @datadoghq dashboard, identifies systemic issues, new errors, and long tail latencies, and has a summary + PR waiting for me at 9am.
Dylan Garcia tweet media
English
22
21
446
47.1K
M retweetledi
elie
elie@eliebakouch·
(continual) pre-training is not dead! some thoughts about cost per task (on cursorbench) being 2x lower, imo it can be due to: - new base model: seems straightforward but it's not imo, you need to optimize the inference/training stack (both for rl and consumer inference), GLM-5 > V3.2 doesn't mean glm-5-base > V3, they may not be equally malleable to post/mid training - more optimized kernel and inference stack/serving (most likely) - rl/mid-training with objective/data to make smaller CoT (most likely) - mid-training with more efficient arch: i would love for this to be true, and i can see how it's necessary if they use the previous base model generation and need efficient memory for long context, but since they also released some tricks with self-summary, i'd say unlikely? (they can and imo should be combined together for very long tasks)
elie tweet media
Cursor@cursor_ai

Composer 2 is now available in Cursor.

English
3
4
119
10.6K
m_11
m_11@instance_11·
i envision a world without duplicates
English
57
544
5K
336.7K
M
M@init_malachi·
@instance_11 yeah, it used to be my main thing
English
0
0
0
6
m_11
m_11@instance_11·
@init_malachi thank you! the audio here is a snippet from a long session of improvising, imagining this video as i played; i then made the video based on the audio, it was a joy to draw from both sides
English
1
1
12
1K
M
M@init_malachi·
i’m going thru my old schizophrenic notebooks and it reads like 2026 twitter takes
English
0
0
0
44
M retweetledi
m_11
m_11@instance_11·
@r0b0t_sp1der i’m still very much learning the ropes of audio, i wanted a very metallic/vibrational sound so i recorded the piano with an iPhone laying directly on the frame; the voice was generated from a similarly tinny source, i then applied some light filters and added a hard limiter
English
2
1
64
3K
M retweetledi
Jesse Abraham Lucas 🌃
Jesse Abraham Lucas 🌃@JesseLucasSaga·
You must read God-Emperor of Dune understanding "precognition" as the computability and predictability of human individual/group actions, which will take off in our lifetimes; the computers/Bene Gesserit are the Fates, who Leto II frees us from. It is in dialogue with Foundation.
English
29
76
1.5K
47.9K
M retweetledi
Simone Foti
Simone Foti@simo_foti·
🧠AGC. Most surface CNNs use fixed patch sizes. AGC uses our diff. framework to dynamically learn the optimal receptive field for each channel & layer. It adapts to local geometry, beating fixed-patch methods.
Simone Foti tweet media
English
1
1
11
818
M retweetledi
Simone Foti
Simone Foti@simo_foti·
🌊MeshFlow🐤: our alternative to Riemannian flow matching (RFM) on meshes. We skip the ODE solver by back-propagating through our Exp map to learn a static vector field. We are 16,000x faster in inference, use 97% less GPU memory, and deliver superior results than RFM.
Simone Foti tweet media
English
1
2
15
1K