Post

Grummz
Grummz@Grummz·
Every AI company should be optimizing for speed of inference. It’s the difference between at 10fps game and a 60fps game. The same game feels horrible at 10 and amazing at 60. It’s not just a quality be degrees situation, it’s transformative. The same model feels 100x smarter and can get more done with realtime inference.
English
16
6
121
8.7K
Sunny Golovine
Sunny Golovine@sunnygg·
It’s a real balancing act. For example I use Codex Spark a lot and while it’s a great model, it has its limitations around reasoning and most importantly context window size. I see it more like tools in a toolbox. There are times you want to reach for the ultra fast model cause you don’t need reasoning or a deep context window, you just need to do this small thing FAST. But there are other times you want to build something big and need that larger context window size and the more advanced reasoning.
English
0
0
0
174
Wallscreet
Wallscreet@Wallscreet·
@Grummz I honestly think this is putting more expectation than necessary on models and providers. As long as they move at an equal or greater pace than human capability we should be satisfied overall. Context and model adaptation/learning ability is still the frontier imho.
English
0
0
0
68
テストさん
テストさん@testo_san·
@Grummz I didn't find out until later that DLSS 5.0 currently requires two RTX5090s & they didn't even specify which cards can use it efficiently b/c they prob don't even know yet. What an absolute sh*tshow. Also I get the impression many ppl still can't tell diff between 30fps & 60fps.
English
0
0
0
2
Dochex
Dochex@doc_hex1337·
@Grummz There are already dedicated ASIC inference chip that would work. Soon it will be CPU, GPU and IPU
English
0
0
0
47
Petri Kuittinen
Petri Kuittinen@KuittinenPetri·
@Grummz Faster response is of course nice, but if it speed comes at a noticeable cost to accuracy, this is a big NO for research tasks, fixing hard bugs, translating difficult texts (e.g. old poetry). In those tasks I would be happy with a slower, yet high quality response.
English
0
0
0
69
Mark Kretschmann
Mark Kretschmann@mark_k·
@Grummz It depends on the task. If I ask a complex question that requires research, I'm irritated if the model gives an instant answer. That's because I know that the task requires reasoning time.
English
2
0
8
390
BetMGM 🦁
BetMGM 🦁@BetMGM·
Pick which twin will win! You could score a share of $2 million in Bonus Bets.
English
266
164
3.4K
43.1M
BlaiseBits
BlaiseBits@BlaiseBits·
@Grummz Inference is nice, but being correct is better. In this case, 30fps of awesome is better than 60fps of slop.
English
0
0
1
144
🎯🔫👌
🎯🔫👌@gurgle_io·
@Grummz 100 %. I can't have 30 seconds while it's thinking. Also it never gives a better answer anyway. Maybe for some tailored workflows, but not for general stuff. Response should be instant and it should fit on the screen. Any game dev would have told you that. @xai you're welcome.
English
0
0
1
28
aidenpryde
aidenpryde@aidenpryde·
@Grummz My question is why are they focused so much on "enhancing the graphics." NPCs are still hard coded and running off of the CPU. The promised land for me is a local LLMs running the NPCs. They respond to what you are doing, the world events, etc in real time.
English
1
0
0
38
nør 🏴‍☠️
nør 🏴‍☠️@meta_acc·
@Grummz I would say first fps is more akin to output quality. Inference speed is more like ping.
English
0
0
0
97
DHYohko
DHYohko@DHYohko·
@Grummz Depends on what I ask. Most of the time im willing to wait minutes for more accurate data, expecially regarding tcg rulings and other such things. If its too fast im suspicious about innacurate info.
English
0
0
0
16
Amsanir
Amsanir@Amsanir·
@Grummz This tweet was written by ai
English
0
0
1
51
Teilen