Stellan Haglund

800 posts

Stellan Haglund banner
Stellan Haglund

Stellan Haglund

@Stellanhaglund

Entrepreneur and code ninja. https://t.co/GyK8fQSFZe

Katılım Mart 2009
213 Takip Edilen251 Takipçiler
Ivan Fioravanti ᯅ
Ivan Fioravanti ᯅ@ivanfioravanti·
Playing with my italian wine classification and Qwen3.5-0.8B fine-tuning with MLX on Apple Silicon! From 10% to 72% after first 100 iters with batch size 12. Let's see what I can reach with it!
Ivan Fioravanti ᯅ tweet media
English
7
4
105
7K
Mariusz Kurman
Mariusz Kurman@mkurman88·
Qwen 3.5 4B !!!! 🌶️🌶️🌶️
Mariusz Kurman tweet media
Magyar
29
9
356
31K
Elon Musk
Elon Musk@elonmusk·
@tetsuoai Banger 🤣🤣 How dare they steal the stuff Anthropic stole from human coders??
English
762
748
13.9K
715.8K
tetsuo
tetsuo@tetsuoai·
I can't believe someone would just steal from Anthropic like this. The millions of man-hours Anthropic spent hand-writing code, text, art, books, etc. to generate enough data for training must be taken into consideration here. Where is the respect for IP?
Anthropic@AnthropicAI

We’ve identified industrial-scale distillation attacks on our models by DeepSeek, Moonshot AI, and MiniMax. These labs created over 24,000 fraudulent accounts and generated over 16 million exchanges with Claude, extracting its capabilities to train and improve their own models.

English
366
1K
13.6K
1.2M
Logan Kilpatrick
Logan Kilpatrick@OfficialLoganK·
We just shipped "Batch mode" in the Gemini API with 50% discounts on our 2.5 models and the ability to enqueue billions of tokens at a time!
Logan Kilpatrick tweet media
English
112
163
1.9K
203.5K
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@_philschmid @multimodalart A bit out of context, but I’ve been trying to find an answer to if it’s possible to use a fine tuned model with the batch prediction api?
English
0
0
0
19
Philipp Schmid
Philipp Schmid@_philschmid·
Gemini 2.5 is production ready! We just launched 3 new Gemini models with 2.5 Pro and Flash being now generally available and a new Gemini 2.5 Flash Lite preview! 🧠⚡️🔦 Here is all you need to know: 🔦 New Gemini 2.5 Flash Lite (Preview) with Thinking, 1M context, only $0.1/$0.4, better as 2.0 flash and tool-use. 🧠 ⚡ Gemini 2.5 Flash and 2.5 Pro are now Generally Available (GA) and production-ready. 💰 Updated pricing for 2.5 Flash to $0.30/1M input and $2.50/1M output tokens. 🤗 Start building all 3 Models today in AI Studio and via Gemini API! New Model IDs: `gemini-2.5-flash-lite-preview-06-17`, `gemini-2.5-flash`, `gemini-2.5-pro`
Philipp Schmid tweet media
English
14
38
330
26.6K
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@LingYang_PU That looks pretty sequential, isn’t the upside with diffusion to predict all the tokens every step?
English
1
0
2
242
Victor M
Victor M@victormustar·
You can really feel the A3B here 🚀 (hardware is mbp m4 max 128GB btw)
English
4
0
25
5.5K
Victor M
Victor M@victormustar·
Qwen/Qwen3-30B-A3B is absolute game changer 🤯 ⬇️ Are you on a Mac? then use MLX weights - getting 100 tokens/sec here. It changes everything for real world use cases.
English
47
101
992
142.6K
Rody Davis
Rody Davis@rodydavis·
I was able to create a CRDT in SQLite with pure C (no dependencies) custom loadable extensions! 🚨 uuid.c (From official SQLite misc extensions) hlc.c (Dart HLC ported to C) crdt.c (Dart logic with triggers ported to C) gist.github.com/rodydavis/d197…
English
3
2
57
4.4K
Steren
Steren@steren·
With @ollama 0.6.2, Gemma 3 27B now runs on 1 Cloud Run GPU. That's 27 billion parameters. The biggest Gemma 3 variant. The most capable open model you can run on a single GPU.
Steren tweet media
English
22
64
512
56.1K
Logan Kilpatrick
Logan Kilpatrick@OfficialLoganK·
We just updated the Gemini API docs to make native image generation 🖼️ even more simple to access and get started with across Python, JS, and cURL, happy building : ) ai.google.dev/gemini-api/doc…
English
126
196
2.3K
212.4K
Piers Morgan
Piers Morgan@piersmorgan·
Blocking so many loons tonight. Is it just me or is X getting madder by the day?
English
5.7K
496
17.4K
1.2M
anton
anton@abacaj·
GRPO llama3.2-1B inject "Wait a second..." when model is wrong and let it regenerate then compute loss
anton tweet media
English
21
19
337
36.2K
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@cerebras Like is it possible to make sure the data doesn’t leave Europe when using the api?
English
0
0
0
22
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@cerebras what are the options for European customers with GDPR requirements?
English
1
0
0
59
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@cerebras I mean for inference, the data centers are located in the us right?
English
0
0
0
30
Stellan Haglund
Stellan Haglund@Stellanhaglund·
@elonmusk when something is community noted, why aren’t other posts containing the same argument also marked with the same note? I see posts all day saying stuff I already saw was community noted as false several days ago.
English
0
0
0
21