Onik Dev

95 posts

Onik Dev

Onik Dev

@OnikDeveloper

Software Engineer, with a focus on innovating, not debating

Earth شامل ہوئے Kasım 2025
74 فالونگ3 فالوورز
پن کیا گیا ٹویٹ
Onik Dev
Onik Dev@OnikDeveloper·
AI Companies are kinda screwed. Eventually there will be a LLM that’s small but powerful enough that no one will bother using AI companies API Prediction: 1-2 years, people will run their own models on a single GPU. Won’t be surprised if gov gets involved
English
0
0
0
28
Onik Dev
Onik Dev@OnikDeveloper·
@LottoLabs Can’t tell if you’re saying this is a good thing or a bad thing
English
1
0
1
593
Lotto
Lotto@LottoLabs·
Qwen 3.5 27b could probably 5 shot anything lol
English
17
4
216
17.7K
Onik Dev
Onik Dev@OnikDeveloper·
Yup. Cannot wait for that to happen x.com/onikdeveloper/…
𝗭𝗲𝗻 𝗠𝗮𝗴𝗻𝗲𝘁𝘀@ZenMagnets

I change my mind. I would have bet against @TheAhmadOsman 's Jan'06 prediction that an Opus 4.5 tier model would run on a single RTX 6000 Pro by end of 2026. "Because ultimately the number of parameters matters, and there's no way to escape the physics of vram," I thought. But then Qwen3.5 27B dropped late Feb, which entirely broke the intelligence density barrier. 42 on AA, can run on 3090 with full context, albeit slowly, which now fights Minimax m2.5 which would have required $20k 2x RTX 6K Pros on Feb 12, and Deepseek v3.2 $80k 8x RTX6K in Dec 2025. Now I have little doubt that Minimax 3.3 or Qwen4 or Deepseek v4.5 or whatever will be at or very near Opus 4.5 by Dec 2026. But, also, it's hard to imagine how far ahead Opus 5.6/Gpt-6/etc will be by then. x.com/i/status/20122…

English
0
0
0
12
Onik Dev
Onik Dev@OnikDeveloper·
Just learnt Go and did a small program for a client. My god. It’s so easy to compile. I’m used to .net. Just need to get used to not using OOP and use struct instead of classes
English
0
0
0
8
Onik Dev
Onik Dev@OnikDeveloper·
@BLUECOW009 But once these models catch up with closed source (which I think they pretty much are) the models will eventually be smaller where a normal GPU will suffice. Plus once that happens, I estimate that pc parts prices will crash. That’s when you should definitely get one ;)
English
0
0
1
310
@bluecow 🐮
@bluecow 🐮@BLUECOW009·
Running models locally is pretty useful but reality is that most people, even developers dont have much more than 1 gpu and ~32gb of ram. The best local models in the open need >90gb vram to run, that is not a realistic expectation for general usecase
English
55
10
347
15.9K
Onik Dev
Onik Dev@OnikDeveloper·
Ah man. These Open Source Models. Seriously cannot wait for the PC parts crash price to occur
English
0
0
0
5
Onik Dev
Onik Dev@OnikDeveloper·
@TheAhmadOsman Won’t be surprised that top AI companies will be made redundant. Can’t wait for the PC part price crash 😂
English
0
0
0
100
Ahmad
Ahmad@TheAhmadOsman·
You guys have no idea how happy I am seeing > Qwen3.5-27B going head-to-head with > DeepSeek-V3.2 & MiniMax-M2.5 on the ArtificialAnalysis leaderboard Deep down I believe in 2 things - Small & specialized models - Dense models, not MoEs Plays a major role in why I prefer GPUs
Ahmad tweet media
Ahmad@TheAhmadOsman

i still believe in dense models

English
42
24
460
44.6K
Onik Dev
Onik Dev@OnikDeveloper·
@wesbos Do you blame him? lol Other models are catching up and LLMs are plateauing, so they don’t have the leverage as much anymore. Just marketing advantages
English
0
0
0
99
Onik Dev
Onik Dev@OnikDeveloper·
Called it. Specialised small models for specialised purposes. In the future it won’t be one model that runs all. But small fast models with specialised. I know it’s similar to MoE but still
Hugging Models@HuggingModels

Meet Strand-Rust-Coder-14B, a specialized AI that writes Rust code like a senior developer. It's not just another coding assistant, it's specifically fine-tuned for Rust, making it a game-changer for systems programming and performance-critical applications. This is exactly what the Rust community has been waiting for.

English
0
0
0
12
Onik Dev
Onik Dev@OnikDeveloper·
@HaihaoShen @MiniMax_AI 32B params? What’s the performance hit compared to the original? Am I missing something?
English
0
0
0
43