Alfc Oleol

41 posts

Alfc Oleol

Alfc Oleol

@Hcfthbfdgy

Katılım Ocak 2025
52 Takip Edilen2 Takipçiler
GMI Cloud
GMI Cloud@gmi_cloud·
shipped a new console. to celebrate: Gemini 3.1 Flash is free on GMI for the next 24 hours, GLM 5.1 is free for 12 just go use it 👇
English
8
6
49
135.1K
송준 Jun Song
송준 Jun Song@jun_song·
Working on only loading MoE expert on GPU, 25tok/s on very short context right now
English
3
0
38
4.3K
송준 Jun Song
송준 Jun Song@jun_song·
Running Kimi-k2.6 1T 8bit with only 21GB RAM on my Macbook at speed of 25tok/s. Some of my theory worked, but architecture is not perfect. Need to fix a lot of stuff, but there is hope. Working hard on this future method of Local LLM.
송준 Jun Song tweet media
English
50
43
751
40.6K
nick
nick@thecsguy·
@Hesamation rebranding as thinking machines just to ship a gpt-4o clone is the ultimate 2026 tech move. we’re officially in the 're-skinning 2024' era and nobody’s even trying to hide the wrappers anymore
English
2
0
32
5.1K
Flowers ☾
Flowers ☾@flowersslop·
images v2 definitely has a knowledge cutoff of 2024 because it often assumes Schulman, Karpathy, Murati and so on still work at OpenAI. 5.5 absolutely knows they dont. 4o thinks they still work there. So I think its reasonable to assume images v2 is 4o
Flowers ☾ tweet media
English
16
0
157
10.6K
Alfc Oleol
Alfc Oleol@Hcfthbfdgy·
@SIGKITTEN Parallel commands for terminal used by agent pls
English
1
0
0
38
SIGKITTEN
SIGKITTEN@SIGKITTEN·
ive got myself in a bit of a predicament. i have an insane line up of new features basically all ready - opencode, pi, claude - new QUIC qr code type of connection type thats a lot better for mobile - file picker to mount dirs from Files - a bunch of updates for the local linux but the kitty app is pretty good right now and i dont wanna mess up the experience while these are being tested the issue is that ive got an open TestFlight and android beta and there are already like 3000 people in it, a lot of them using it regularly. i need to somehow get the people who dont understand that this is for testing off the test channels
English
24
4
95
5.7K
Alfc Oleol
Alfc Oleol@Hcfthbfdgy·
@SIGKITTEN Why not use fork of ish like ish-arm64 for more performance?
English
1
0
0
585
BridgeMind
BridgeMind@bridgemindai·
GPT 5.5 just debuted on BridgeBench. It ranks below GPT 5.4. Read that again. The "most intelligent model ever built" scores worse than its predecessor on real world vibe coding. #8 overall. 84.5 quality. GPT 5.4 sits at #6 with 85.1. Behind Claude Opus 4.6. Behind Claude Opus 4.7. Behind Claude Sonnet 4.6. Behind Grok 4.20. Behind Qwen 3.6 Plus. The smartest model in the world is not the best coding model in the world. BridgeBench just proved it. bridgebench.ai
BridgeMind tweet media
English
133
11
194
48.9K
Alfc Oleol
Alfc Oleol@Hcfthbfdgy·
@neural_avb Why with this logic you dont use Minimax M2.7 (230B-A10B) Only 10B active parameters man! Better then deepseek v4 flash and grok
English
0
0
1
181
Boyuan Chen
Boyuan Chen@BoyuanChen0·
We are committed to continually improving the GPT Image 2 model! I am actively fixing various issues from the community feedback. Just reply or DM me your GPT conversation! Features like 2K or 4K images are already available via the experimental API. Hope you enjoy the model!
Boyuan Chen tweet media
English
255
55
993
230.3K
Mark Kretschmann
Mark Kretschmann@mark_k·
GPT-Image-2 is fundamentally broken and should not have been released in this state. The noise artifact issue is no joke. It affects all images, some more than others, and the problem is amplified by iterative editing.
English
108
26
641
77K
Alfc Oleol
Alfc Oleol@Hcfthbfdgy·
@mark_k Nano banana using upscale for 2k/4k
English
0
0
0
63
Mark Kretschmann
Mark Kretschmann@mark_k·
GPT-Image-2 has image quality issues, and nobody is talking about them... Except me. So the image is slightly grainy, fuzzy, and also low resolution (1K). Nano Banana 2 is better in that regard. After the initial hype you will see that other people start noticing it too.
English
98
24
492
46.6K
Anemll
Anemll@anemll·
WIP: Merged: prefill dedup for Flash-MoE. MiniMax only for now while I validate the other models Layer-major prefill runs one layer at a time, groups repeated expert picks across the prefill chunk, loads each unique expert from SSD once, and reuses it for every token that picked it , instead of reloading per token. Bigger prefill chunk → more reuse → cheaper prefill. Prefill is now compute-bound, not I/O-bound. That should help both GPU and ANE prefill at large batch sizes. Can we speedup regular MoE MLX prefill 🤔 github.com/Anemll/anemll-…
Anemll tweet media
English
3
3
39
8.8K
Chris
Chris@chatgpt21·
Honestly this chart makes me more bullish on GPT 5.4 Pro than anything else. People are focusing on Mythos looking strong, but what stands out to me is how well 5.4 Pro already stacks up on the overlap we actually have. GPQA is basically a tie at 94.4 vs 94.5. BrowseComp is a win for GPT 5.4 Pro at 89.3 vs 86.9. Yes, Mythos is ahead on Humanity’s Last Exam, 56.8 vs 42.7 without tools and 64.7 vs 58.7 with tools, but the bigger point is that 5.4 Pro is already this competitive right now. So if GPT 5.4 Pro is already THIS COMPETITIVE here, then Spud Pro, the next OpenAI flagship, is guaranteed to beat Mythos. This chart makes OpenAI look extremely close before its next real jump, and once that next jump lands I do not think Mythos stays ahead.
Chris tweet media
English
129
125
1.8K
323K
DogeDesigner
DogeDesigner@cb_doge·
Grok 4.20 Reasoning just took the #1 spot on the BridgeBench reasoning benchmark. 🔥 Beating GPT-5.4, Claude Opus 4.6, Google Gemini and others. Week after week, Grok keeps climbing across benchmarks. 🚀
English
350
306
1.5K
9.8M
Alfc Oleol
Alfc Oleol@Hcfthbfdgy·
@svpino Its bug, find fixed llama cpp fork with usable tool calling with gemma 4
English
0
0
0
285
Santiago
Santiago@svpino·
I'm running Gemma 4 on my computer with Ollama. Unusable with Claude Code. It can't even load and execute skills, so I had to stop. But the model is pretty decent as a chatbot using the Ollama UI. I've been cross-posting questions across Claude and Gemma 4, and I can use Gemma's answers without any problems. I wish we had a better UI harness for the model (with projects, memory, etc.)
English
218
16
619
159K
Lotto
Lotto@LottoLabs·
Is everyone just not using Gemma 31B locally because tool calls are broken
English
68
1
191
27K