

Trevor McCourt
1.3K posts

@trevormccrt1
"Life is a beautiful, magnificent thing, even to a jellyfish"



If you are attending @APSphysics March meeting, come learn more about thermodynamic computing! Our work on taming non-equilibrium thermal electron fluctuations in silicon is now accepted in Physical Review Applied. Read more here: journals.aps.org/prapplied/abst…



Okay here's the first thing I did with THRML by @extropic It's just a basic sudoku solver. Thermodynamic computing is a bit overkill for this task but I think since humans can actually do sudoku, it's a good intuition for what's going on under the hood. With sudoku, there are many overlapping constraints. You start with a partially filled puzzle, which are the initial conditions, but then other rules are: no duplicates on any row, column, or square. Now, with a sudoku problem, you know there is ONE singular solution, or a "low energy state" i.e. where there are no rule violations or collisions. So then what you do is you program those "clamped" initial values into the TSU, and you bake in the rules (no duplicates) and then, due to the laws of thermodynamics and electricity... it just sort of settles into the correct solution (this is "annealing") The reason I think this is such a good example of what TSUs do is because for humans (and classical computers) it's more or less a "guess and check" process. No matter what method you use with classical computation or human computation, it's an iterative refinement process of sequential steps. But, with sudoku, as you can see in the output below, it's a single step. That's because the TSU looks at the whole problem globally. Here's how I did this: ChatGPT PRO 🤣 No joke, ChatGPT pro one-shotted this entire problem. There were several refinements we made, though it was mostly around UI and validation (not the core logic). However, we did do an optimization step to make sure we were using the correct block batching from the THRML library.






Hello Thermo World.






Carl, thanks for the thoughtful engagement with our material! You are absolutely right that random sampling makes up a vanishing part of the computational workload in something like a transformer or diffusion model. These algorithms evolved alongside the GPU, and therefore naturally mostly use the things GPUs are good at (matrix-vector multiplication). However, we arent trying to run these models. In fact, we are trying to take things in a completely different direction! At Extropic, we are co-designing hardware that is really good at sampling and machine learning algorithms that make heavy use of sampling. Our paper provides early evidence that this approach could be really really energy efficient. In this paper, we show that sampling hardware can run a denoising model (parent concept of diffusion) that generates images that are similar to a simple benchmark dataset. This, of course, does not mean we are going to replace ChatGPT any time soon. There is a ton of work needed to scale our approach to something comparable with today's LLMs. However, the way we are doing things now is insanely costly from an energy perspective, which IMO makes pursuing less incremental (but riskier) technology a no-brainer. arxiv.org/pdf/2510.23972






The foundations have been laid. Now it's time to scale. Excited for the Thermodynamic Intelligence takeoff ahead.

Hello Thermo World.

@trevormccrt1 the big flaw in both papers is that they only run weird EBMs that nobody wants to use! imo new hardware needs to be able to run conventional SOTA models to get meaningful adoption (though i realize that's not your view)