Peter Hizalev

684 posts

Peter Hizalev banner
Peter Hizalev

Peter Hizalev

@petrohi

TT-Lang at Tenstorrent. Also retro computing and homebrew electronics.

San Jose, CA Katılım Temmuz 2009
125 Takip Edilen438 Takipçiler
Peter Hizalev retweetledi
Artem Y
Artem Y@artem_aero·
Tenstorrent can run AI magically fast. Try it yourself. console.tenstorrent.com - Super fast large LLMs. DS R1 685B model flying at 350t/s/u on 16 Blackhole galaxies - Super fast video generation with Wan 2.2 Lighting (Prodia) finishing 5 seconds clip within 3 seconds on 4 galaxies. I know you want more. You’ll get it fast. The team is cooking more.
Davor Capalija@davorVDR

We’re ready, and we’re very committed to this. 😎

English
8
20
102
22.2K
Peter Hizalev retweetledi
Tenstorrent
Tenstorrent@tenstorrent·
Tune in tomorrow! Run fast video, speech, code all on Tenstorrent Galaxy Blackhole. Powered by our Networked AI architecture with native scale-out. Hear from our partners and customers deploying at scale. Watch the livestream on May 1st @ 1:30 PM PDT: tenstorrent.com/deploy
English
1
4
35
2K
Peter Hizalev retweetledi
Colman Glagovich
Colman Glagovich@ColmanGlag·
FlashAttention on Tenstorrent, a technical report
Colman Glagovich tweet media
English
1
4
45
6.1K
Peter Hizalev
Peter Hizalev@petrohi·
@jimkxa In hardware they already moved to tiled and async programming models, which is why Triton existed and now is extended by OpenAI and Meta to support explicit async threads. Once there is critical mass of these kernels in GenAI training datasets they become AI-portable.
English
0
0
0
127
JLarky
JLarky@JLarky·
Dad, we have been through this, stop telling people you take Ketamine like Elon Musk, you take Creatine like Joe Rogan!
English
1
0
1
682
Peter Hizalev retweetledi
Modular
Modular@Modular·
Part 3 of "Matrix Multiplication on Blackwell" is here! It continues our epic journey of describing how Modular implemented the fastest B200 matmul in the industry, revealing the techniques to go from 16% to 85% of SOTA modular.com/blog/matrix-mu…
English
0
11
48
9.4K
JLarky
JLarky@JLarky·
that's like the most messages I get in DMs nowadays :(
JLarky tweet media
English
6
0
4
1.1K
JLarky
JLarky@JLarky·
Do you think people who are saying that AI will replace jobs ever heard about pumping gas in Oregon?
English
1
0
3
425
Peter Hizalev
Peter Hizalev@petrohi·
@clattner_llvm I would be super interested in your take on performance. Is there a true need of automatic exploration of tiling space given today’s hardware and workloads? What’s the place for runtime auto tuning (TVM, Triton)? And for compile-time exploration (Part-IR, XLA Shardy)?
English
0
0
0
51
Chris Lattner
Chris Lattner@clattner_llvm·
Let me know if you have any specific suggestions or request on these topics.
English
4
0
9
3.6K
Chris Lattner
Chris Lattner@clattner_llvm·
I got busy with other things, so the next Democratizing AI Compute post is aiming for next week. Stay tuned to dive into C++’y OpenCL/SyCL/OneAPI and compiler’y XLA/MLIR tech. We will look at what worked and what struggled, with a goal of understanding the past.
English
4
5
146
7.9K
antirez
antirez@antirez·
Not the right moment, give me some time, but soon or later I'll ask for the help of fellow hackers (something like 20 folks), create a first "open technology news group", and create a competitor of Hacker News that has transparent rules and lacks any personal goal.
English
5
2
37
3K
Kurt Schrader
Kurt Schrader@kurt·
Does anyone have an example of a company that's built a piece of SaaS that works great for startups and ALSO works well for enterprise companies?
English
12
1
108
4.8K
Peter Hizalev
Peter Hizalev@petrohi·
@burkov is there a place to submit errors in your LM book?
English
1
0
0
33
Peter Hizalev
Peter Hizalev@petrohi·
@JLarky I am using UML and mouse to visually design my software. It automatically generates all necessary code. (Circa 2002)
English
1
0
1
31
JLarky
JLarky@JLarky·
- oh, so you are a developer? What language do you use? - mostly English and a lot of pressing of the [tab] key
English
3
0
9
667
Peter Hizalev
Peter Hizalev@petrohi·
@antirez Let’s keep open mind: Stochastic parrots are the new frontier for AI reasoning
English
0
0
4
28
antirez
antirez@antirez·
You may think contortionists are able, but then give a look at AI researches that need to move from "LLMs are just stochastic parrots" to "LLMs are the new frontier for AI reasoning".
English
2
0
6
2.1K
Peter Hizalev
Peter Hizalev@petrohi·
@eevblog What’s on the logo? Is it “staring into the abyss” or “going down the drain”?
English
0
0
1
49
Dave Jones
Dave Jones@eevblog·
The economy is going to crap, so I'm branching out into a smart business to capitalise.
Dave Jones tweet media
English
3
1
18
2.1K
Peter Hizalev
Peter Hizalev@petrohi·
@JLarky Who knew that communism only needed few more yottaflops to work!
English
0
0
1
35
JLarky
JLarky@JLarky·
@petrohi Or go straight to communism
English
1
0
1
139
JLarky
JLarky@JLarky·
Why does it have to be coding? Like if AI is so smart, just ask it to make you money. Skip the middle man. Just ask it to create you a business, why is your first thought to ask it to build an app or a website?
English
9
4
39
3.2K
Peter Hizalev
Peter Hizalev@petrohi·
@burkov Same way we can argue that a typical RGB input image for Conv2d has naturally a NHWC layout. It beats me why PyTorch insists on putting channels after batch, but it does for some reason.
English
0
0
0
42
BURKOV
BURKOV@burkov·
@petrohi 1D kernels are used in text processing where we have: batch dimension, sequence length, and embedding. So, to use Conv1d we have to x.permute(0, 2, 1).
English
1
0
0
88
BURKOV
BURKOV@burkov·
I love PyTorch but what the hell: why nn.Conv1d expects the input of the shape (batch_size, channels, length) and not (batch_size, length, channels)?
English
2
0
8
1.6K