Tweet ghim
bifkn
11.5K posts

bifkn
@therealbifkn
AI | gaming | fitness. Testing local models, memory, dashboards, and evals on my own messy performance data.
Tham gia Mayıs 2022
1.6K Đang theo dõi1.4K Người theo dõi

@aaron_vdp @jcthebadcat @cgarciae88 Artificial life is actual life kept alive by artificial means.
If you go down that road, anything can be alive. Is my keyboard alive if I say it is?
English

Just change the power settings ffs.
Can Vardar@icanvardar
if you don’t hold your macbook like this you’re not agentmaxxing enough
English

@cwolferesearch In what domain? So far 27B handles all of my coding tasks significantly better than Gemma 4
English

@tawer1O Qwen’s game looks like much better game play.
That’s been my experience too. Gemma has faster results, but the quality isn’t there.
bifkn.vercel.app/lab/octopus-ar…
English

Comparison of Gemma 4 and Qwen 3.6 in the same coding task
Same hardware, same prompt, comparable model size
> gemma 4 31b: 27 tok/s, 3m 51s, 6,209 tokens, stronger game logic
> qwen 3.6 27b: 32 tok/s, 18m 04s, 33,946 tokens, better visuals
Gemma 4 with lower tok/s finished 14 minutes faster because it used 5.5x fewer tokens to reach a complete answer
English

@AlexFromAtomic @kimmonismus @atomic_chat_hq Now you're talkin!
Also, here are all the models/variants I've ran
bifkn.vercel.app/lab/octopus-ar…
English

@therealbifkn @kimmonismus @atomic_chat_hq I like Qwen version ;) Let's host it and make our own B2B AI SaaS with 10k MRR))
English

/1 Gemma 4 31B just crushed Qwen 3.6 27B in a local LLM gamedev contest inside @atomic_chat_hq (prompt is below)
Device: MacBook Pro M5 Max, 64GB RAM
Results:
Qwen 3.6 27B: 32 tokens/sec · 18m 04s · 33,946 tokens
Gemma 4 31B: 27 tokens/sec · 3m 51s · 6,209 tokens
So what is more important: tokens per second, or the quality of the final answer?
Qwen made a very long response and showed more creativity and visual style. But Gemma gave a shorter, clearer, and more logical answer in much less time. In this one-shot Pac-Man gamedev contest, Gemma 4 31B was the clear winner. Its game logic was stronger: click reactions were smoother, and it handled interactions with elements like walls, ghosts, and particle effects better.
But this was only one test. Maybe Qwen 3.6 27B can show better results with better settings. Open the comments, try our prompt, and share your result below.
English

Yes - the site is not great - but minibif is working on it 😂
I made a thing.
bifkn.vercel.app/lab/octopus-ar…
English

@leftcurvedev_ @OpenRouter Can't deny how insanely good 27B is, but also props to 397B, very pretty!
We need 3.6 397B and 122B!!!
English

🥊 Time for a new fight
Qwen3.5 397B A17B vs Qwen3.6 27B
🌸 "Cherry Blossom" (↓ prompt below)
Using @OpenRouter for 397B
Running 27B locally on a single RTX 5080
Wow! 🤯
I'm sure this one is going to divide opinions. 397B feels cinematic with the flare and shadows it added, while 27B is just beautiful; it has a completely different color palette and the leaves are falling from the branches as asked. It's really tough to decide which model is the best. I was about to call it a tie, but then I remembered that 397B needs 180GB VRAM to run locally 😅
Don't get me wrong the model is amazing, but right now the VRAM it requires doesn't feel justified. It's simply not 10× better than 27B. We'll have to dig deeper with more prompts to be sure, let's see what each one has in store for us.
What do you think?
English

@loktar00 From wife's point of view:
"Please don't spend the rest of the night working on small demos with Qwen 3.6....
Please don't spend the rest of the night working on small demos with Qwen 3.6....
Please don't spend the rest of the night working on small demos with Qwen 3.6...."
English
bifkn đã retweet

I'm 22 years old and Claude Code is deteriorating my brain.
Every single day for the last 6 months I've had 6 to 8 Claude Code terminals open, waiting for a response just so I can hit 'enter' 75% of the time. And it's doing something to me.
In convos with a couple of friends, it's been a point that's been brought up pretty frequently.
None of us feel as sharp as we used to.
I don't know if it's just us, or others in their 20s are feeling the same thing, but it's something I've been thinking about a lot.
P.S. I know this is a problem with my reliability/usage of it, not Claude Code itself, but the effects are real nonetheless
English














