Hooman Shayani

896 posts

Hooman Shayani banner
Hooman Shayani

Hooman Shayani

@hsh95

Research Scientist @ Autodesk AI Lab

London, England Katılım Haziran 2009
1.2K Takip Edilen347 Takipçiler
Sabitlenmiş Tweet
Hooman Shayani
Hooman Shayani@hsh95·
Excited to share WaLa: our open-source, 1B-param 3D generative model. It supports diverse inputs: point clouds, voxels, images, text, depth, and multi-view. Offering rapid 3D shape generation in 2-4 seconds. Explore the code and pre-trained models:
English
2
0
9
683
Hooman Shayani
Hooman Shayani@hsh95·
@lucasmaes_ Very nice demonstration of the point. Isn’t explainability the price?
English
0
0
1
375
Hooman Shayani
Hooman Shayani@hsh95·
@fchollet What’s your definition of symbolic? Or rather, when do you say a rule is not symbolic? I think compression always ends up in symbols that can present more than one thing. But some symbols are understandable and communicatable by humans, some aren’t. Don’t you think?
English
0
0
1
78
François Chollet
François Chollet@fchollet·
All the great breakthroughs in science are, at their core, compression. They take a complex mess of observations and say, "it's all just this simple rule". Symbolic compression, specifically. Because the rule is always symbolic -- usually expressed as mathematical equations. If it isn't symbolic, you haven't really explained the thing. You can observe it but you can't understand it.
English
486
730
6K
13.2M
Yuandong Tian
Yuandong Tian@tydsh·
Several of my team members + myself are impacted by this layoff today. Welcome to connect :)
English
469
267
6.4K
4.4M
Muthu / முத்து
Muthu / முத்து@muthutalks·
I was not impacted by the Meta PAR/FAIR layoffs today. I’m a research scientist passionate about computer vision, generative models, multimodal understanding, and AI safety. Over the past decade, I’ve contributed to advancing a broad range of applied machine learning problems —
Susan Zhang@suchenzang

👀

English
2
0
0
677
Mimansa Jaiswal
Mimansa Jaiswal@MimansaJ·
I was impacted by Meta layoffs today. As a Research Scientist working on LLM posttraining (reward models, DPO/GRPO) & automated evaluation pipelines, I’ve focused on understanding why/wehere models fail & how to make them better. I’m looking for opportunities; please reach out!
Susan Zhang@suchenzang

👀

English
121
231
3.1K
841K
Mariya I. Vasileva
Mariya I. Vasileva@mariyaivasileva·
I was impacted by the Meta PAR/FAIR layoffs today. I’m a research scientist passionate about computer vision, generative models, multimodal understanding, and AI safety. Over the past decade, I’ve contributed to advancing a broad range of applied machine learning problems — from multimodal generative models and visual recommender systems, to 2D-to-3D human modeling, large-scale data generation for foundation model training, mechanistic interpretability, and rigorous evaluations for trust & alignment of large-scale AI systems. I’m actively seeking new opportunities — please reach out if you have any openings!
Susan Zhang@suchenzang

👀

English
50
67
1K
336.7K
Hooman Shayani retweetledi
evan
evan@evan__casey·
New #ICCV2025 paper: ✨ Aligning Constraint Generation with Design Intent in Parametric CAD ⚙️ We apply post-training techniques to the task of generating engineering sketch constraints found in parametric CAD, using a constraint solver for verifiable rewards.
GIF
English
1
4
6
1K
Hooman Shayani
Hooman Shayani@hsh95·
@GaryMarcus @francoisfleuret Biological systems have a Markov Blanket. They model the world according to their Markov Blanket Boundary. LLMs model their input output distribution which is their MBB and what they can see of the world. Free Energy Principle. Right?
English
1
0
0
150
Gary Marcus
Gary Marcus@GaryMarcus·
@francoisfleuret not the same. the LLM models an output distribution (most/many) biological systems model the world.
English
2
1
10
1.5K
François Fleuret
François Fleuret@francoisfleuret·
What a very strange angle to demonstrate limitations of LLMs. Exactly like a bird mimicking another bird's song, an LLM (or an MLP!) matches the visible signal, but devises its internal states ("motor actions") without direct supervision.
Richard Sutton@RichardSSutton

@eigenrobot Even in birdsong learning in zebra finches the motor actions are not learned by imitation. The auditory result is reproduced, not the actions; in this crucial way it differs from LLM training.

English
46
13
323
56.2K
Hooman Shayani retweetledi
Sam 𝕏u
Sam 𝕏u@SamXu03799145·
We are hosting live demo at the AI tomorrow booth! Autodesk unleashes Neural CAD 3D generative AI foundation models develop3d.com/cad/autodesk-u…
English
0
1
9
2.6K
Jon Richens
Jon Richens@jonathanrichens·
@hsh95 @KarlFristonNews Similar statement, but Friston assumes the agent operates based on a world model. Whereas we prove that any agent conforming to our assumptions has learned a world model. But also, we say nothing about if / how the world model is used by the agent.
English
1
0
1
75
Jon Richens
Jon Richens@jonathanrichens·
Are world models necessary to achieve human-level agents, or is there a model-free short-cut? Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world models… 🧵
Jon Richens tweet media
English
35
173
1.1K
185.6K
Jon Richens
Jon Richens@jonathanrichens·
Turns out there’s a neat answer to this question. We prove that any agent capable of generalizing to a broad range of simple goal-directed tasks must have learned a predictive model capable of simulating its environment. And this model can always be recovered from the agent.
Jon Richens tweet media
English
3
9
63
8.2K
Hooman Shayani
Hooman Shayani@hsh95·
@docmilanfar Well eq2 is very intuitive while I can’t even understand what the assumptions of eq1 are! 1/2PNr is simply the area of the triangle representing the integral of the interest over time assuming the principal is paid back uniformly.
English
0
0
3
782
Peyman Milanfar
Peyman Milanfar@docmilanfar·
I published this in a 1-pager: P. Milanfar, “A Persian Folk Method of Figuring Interest”, Mathematics Magazine, vol. 69, no. 5, Dec. 1996 My late dad refused to be a co-author. But when it appeared, he printed it out, framed it, and hung it on the wall of the house 🙂 4/4
Peyman Milanfar tweet media
English
26
98
1.5K
52.9K
Peyman Milanfar
Peyman Milanfar@docmilanfar·
Years ago when my wife and I we were planning to buy a home, my dad stunned me with a quick mental calculation of loan payments. I asked him how - he said he'd learned the strange formula for compound interest from his father, who was a merchant in 19th century Iran. 1/4
Peyman Milanfar tweet media
English
50
359
3.9K
483.3K
Hooman Shayani
Hooman Shayani@hsh95·
@keenanisalive This is beautiful but rule 5 needs more explanation. What’s the meaning of the notation S(pq)? Why do you say *or*? pq is *and*. Right?
English
0
0
4
85
Keenan Crane
Keenan Crane@keenanisalive·
Rule 5: If two improbable events are unconnected, then I'm surprised if either of them occurs. E.g., if I see a shooting star *or* my engine dies, I'm surprised. More generally, my overall surprise at two events is the sum of individual surprises: S(pq) = S(p) + S(q).
Keenan Crane tweet media
English
6
2
99
10.9K
Keenan Crane
Keenan Crane@keenanisalive·
Entropy is one of those formulas that many of us learn, swallow whole, and even use regularly without really understanding. (E.g., where does that “log” come from? Are there other possible formulas?) Yet there's an intuitive & almost inevitable way to arrive at this expression.
English
72
691
4.4K
493.7K
Hooman Shayani retweetledi
ぬこぬこ / NUKO 🇯🇵
arxiv.org/abs/2411.08017 Autodesk の 3D モデル生成技術の論文 Wavelet Latent Diffusion (WaLa) という手法を提案しているのですが、残念ながら Latent Diffusion の僕の理解が浅く適当なことを言いかねないので、手法のすごさは論文をご参照ください笑 huggingface.co/ADSKAILab モデルは HF に公開、Google Colab のデモも用意されています。 colab.research.google.com/drive/1W5zPXw9… 飛行機のサンプルを見るとこのくらいであればローカルで動くようになったのかと。 Colab 課金中で L4 / A100 を使える方はぜひ(VRAM 20GB 以上必要)
日本語
0
3
10
1.1K
Hooman Shayani
Hooman Shayani@hsh95·
Excited to share WaLa: our open-source, 1B-param 3D generative model. It supports diverse inputs: point clouds, voxels, images, text, depth, and multi-view. Offering rapid 3D shape generation in 2-4 seconds. Explore the code and pre-trained models:
English
2
0
9
683