grasgor

992 posts

grasgor banner
grasgor

grasgor

@grasgor

world Katılım Eylül 2023
667 Takip Edilen73 Takipçiler
grasgor
grasgor@grasgor·
I realized that I find myself using google's ai mode more often than chatgpt now for general queries
English
0
0
0
8
grasgor
grasgor@grasgor·
@classiclarryd llm vessels is true, im already burning tokens without having put in much thought except while creating the prompt, perhaps I will be more involved when claude cant get me far enough
English
0
0
0
844
Larry Dial
Larry Dial@classiclarryd·
Very cool. My 2 cents for participants: most compute will be spent on undifferentiated hill climbing from people functioning as LLM vessels. Agents can climb hills, but humans are still superior at finding them. What paradigm can you introduce? Sparse circuit discovery and compression during training? Variable embedding sizing? Manifold-ultra-connections? Paired head attn on steroids? Decision tree distillation? The list is endless.
OpenAI@OpenAI

Are you up for a challenge? openai.com/parameter-golf

English
12
14
472
58.2K
grasgor
grasgor@grasgor·
i'm yet to come across a way more effective than pen and paper while learning
English
0
0
1
15
grasgor
grasgor@grasgor·
@stevibe thoughts on temporal memory rather than re-captioning each frame?
English
0
0
0
162
stevibe
stevibe@stevibe·
I'm obsessed with pushing local small models to their limits. Qwen3.5:0.8b doing real-time video captioning on a Mac Studio M2 Ultra, streaming descriptions as the video plays. Under 1s per frame — 269 frames captured & described from a 3m49s video. Pause anywhere and read the captions, it describes every frame surprisingly well. This model is barely 1GB. Local AI is moving absurdly fast.
English
62
130
1.9K
103.2K
grasgor
grasgor@grasgor·
tell me mirofish (if it works as expected) won't become minority report (it will)
English
0
0
0
33
grasgor retweetledi
Discerner
Discerner@Discerner4u·
@bnjmn_marie Yes 4b is amazingly slow.. and borderline unusable for agents
English
0
1
0
71
grasgor
grasgor@grasgor·
@ashvanth_s1 now that you ask, it might be because the model is a reasoning one so the actual output takes timei plus the agent loop (idk what hermes does under the hood; if it does something specific), although I get ~76 tok/sec for plain llm inference
English
0
0
1
26
Ashvanth.S
Ashvanth.S@ashvanth_s1·
@grasgor Your laptop has 5060 and it is somewhat slow still ??
English
1
0
0
39
grasgor
grasgor@grasgor·
maybe it's coincidental or maybe because it just knew, nonetheless quoting since this popped on my TL. I've got a qwen3-4b-q4 quant with 128k context running on my laptop's 5060. It's somewhat slow, and I don't know what I'll do with it but definitely a start
grasgor tweet media
Sudo su@sudoingX

cancel your chatgpt subscription and delete your openclaw slop. i'm serious. go on ebay and buy a used RTX 3060 for the price of two months of pro. or check your drawer because half of you already own one and forgot about it. install hermes agent from @NousResearch. one framework, 31 tools, file operations, terminal, browser, code execution. connect it to your local llama.cpp server running qwen 3.5 9B Q4. total download is 5.3 gigs. that's it. that's the whole setup. every experiment you hesitated to run on API. every project you shelved because you didn't want your data on someone else's server. every late night idea you didn't test because you hit your rate limit. all of that is gone. runs 24/7 on your electricity. your machine. your data never leaves your house. connect it to telegram if you want it on your phone. hook up whatever tools you need. the model thinks at 29 tok/s with 128K context and it never bills you. qwen 3.5 9B and one RTX 3060 is the setup most people will never try because they've been trained to believe intelligence has to come from a datacenter. it doesn't. it runs on 12 gigs of VRAM under your desk right now. stop giving your thinking away for free.

English
2
0
2
181
Shreyas Rao
Shreyas Rao@pareto_pakodas·
Ok done. This took a bit longer than expected because my ISP provider had some "fair-use" data quota which I busted downloading all this. Had to bump up my subscription to get more data. Glad that this (and the AI models ) have been downloaded. Homage to @DejaRu22 !
Shreyas Rao tweet media
DR22 Ω 🪬🎭@DejaRu22

Monthly reminder: Back up everything you want to access later Download all of it External hard drive Backup external hard drive Cloud storage if you are so inclined

English
3
2
27
16.1K
grasgor
grasgor@grasgor·
It's probably going to be a tweak it as you go, add things you want, I can see this becoming what neovim was for terminal editors. I'm currently running the model via llama.cpp but I'll see if I can manage to speed up the inference
English
0
0
0
35
Kaito | 海斗
Kaito | 海斗@_kaitodev·
5 minutes ago, @karpathy just dropped karpathy/jobs! he scraped every job in the US economy (342 occupations from BLS), scored each one's AI exposure 0-10 using an LLM, and visualized it as a treemap. if your whole job happens on a screen you're cooked. average score across all jobs is 5.3/10. software devs: 8-9. roofers: 0-1. medical transcriptionists: 10/10 💀 karpathy.ai/jobs
Kaito | 海斗 tweet media
English
967
1.8K
12.1K
3.5M
grasgor
grasgor@grasgor·
someone please tell my friend
grasgor tweet media
English
1
0
2
32
divya venn
divya venn@divya_venn·
software is no longer a technical field, it's a linguistic one
divya venn tweet media
English
4
1
64
3.3K
grasgor
grasgor@grasgor·
@willccbb @seconds_0 good at kernels? curious, where does a human's moat lie when an agent can iterate much faster - adapt to newer architecture given documentation better
English
0
0
0
446
will brown
will brown@willccbb·
@seconds_0 either getting good at evals or getting good at kernels
English
12
9
297
11K
0.005 Seconds (3/694)
0.005 Seconds (3/694)@seconds_0·
Hypothetically, ai research will still exist as a field in 2027 Hypothetically , if someone wanted to spend the next 9mo upskilling to be an AI researcher, what does that look like? What are the type of outputs labs would expect? Any existing study tracks that are valuable?
English
28
13
501
51.5K
grasgor
grasgor@grasgor·
@ashvanth_s1 always english, it repeated my prompt thrice in the name of reasoning trace
grasgor tweet media
English
0
0
1
11
Ashvanth.S
Ashvanth.S@ashvanth_s1·
@grasgor How does the thinking outputs look like ? Does it do in the same language you gave as input or in english ?
English
1
0
0
23
grasgor
grasgor@grasgor·
so much for saying regional and multilingual ai
grasgor tweet media
English
2
0
1
56
grasgor
grasgor@grasgor·
nah man this shouldn't be happening, disappointed
grasgor tweet media
English
0
0
1
21
Lazarz
Lazarz@Laz4rz·
Ahh the classic > I just had a perfect final interview I wonder when I get the offer into > we decided not to move forward pipeline
English
10
3
242
11.8K