Fisherman

7.4K posts

Fisherman

Fisherman

@Rybens92

Here just to support good open source, Linux Gaming and AI projects.

Polska Katılım Ocak 2013
291 Takip Edilen59 Takipçiler
Fisherman retweetledi
FFmpeg
FFmpeg@FFmpeg·
FFmpeg is moving to Rust 🦀 Our use of C and Assembly in FFmpeg has been an unacceptable violation of safety. FFmpeg will be running 10x slower - but we're doing it for your safety. All your videos will appear green - safety first, working software later.
English
41
54
693
9.9K
Fisherman retweetledi
clem 🤗
clem 🤗@ClementDelangue·
Today we’re releasing TRL v1. 75+ methods. SFT, DPO, GRPO, async RL to take advantage of the latest and greatest open-source. 6 years from first commit to the library that post-trains most open models in the world. Built to be future proof. pip install trl
clem 🤗 tweet media
clem 🤗@ClementDelangue

After @Pinterest @Airbnb @NotionHQ @cursor_ai, today it’s @eoghan @intercom publicly sharing that they’re finding it better, cheaper, faster to use and train open models themselves rather than use APIs for many tasks. And hundreds of other companies are doing the same without sharing. Ultimately, I believe the majority of AI workflows will be in-house based on open-source (vs API). It took much more time than we anticipated but it’s happening now!

English
4
7
18
870
Fisherman retweetledi
Liquid AI
Liquid AI@liquidai·
Today, we release LFM2.5-350M. Agentic loops at 350M parameters. A 350M model trained for reliable data extraction and tool use, where models at this scale typically struggle. <500MB when quantized, built for environments where compute, memory, and latency are constrained. 🧵
Liquid AI tweet media
English
39
107
909
54.7K
Fisherman retweetledi
OpenHands
OpenHands@OpenHandsDev·
We’re live today: Anti-Slop: High Quality Agentic Coding with Ray Myers. If you want practical ways to use coding agents to create well-tested, maintainable software, join us.
OpenHands tweet media
English
1
1
6
269
Fisherman retweetledi
0xSero
0xSero@0xSero·
The first company to make AI boxes, with specialised AI models trained to fit on that hardware will be the next Apple. Would you buy? Should I start a company doing this?
0xSero tweet media
English
213
15
561
33.3K
Fisherman retweetledi
metamike
metamike@itsmetamike·
TurboQuant. Deployed it on an old 3070 running @nousresearch Hermes Agent. OmniCoder 9B (finetuned Qwen 3.5 built from Claude Opus 4.6 agentic and code reasoning traces) soaring at 60toks/s with decent tool calling. Tagging those who inadvertently pressured me to make it work.
metamike tweet media
English
14
25
442
22K
Fisherman retweetledi
Sudo su
Sudo su@sudoingX·
the past week i've been publishing head to head tests between nvidia vs alibaba AI models on consumer hardware like RTX 3090. same GPU, same tests, same inference engine. letting the architectures fight. nvidia failed twice on my 3090. cascade 2 at 3B active hit 187 tok/s but couldn't build a working game in five attempts. openreasoning 32B dense couldn't stop solving math problems long enough to write a single line of code. two architectures, two failures on same card. alibaba's qwen has been winning on every card i've tested. 27B dense on a single 3090 is still undisputed for agent coding. 9B also built a full game on a 12GB card from 2020. so i'm done with consumer hardware for this fight. loading both flagships on 2x H200 NVL. 287GB of VRAM. full BF16 precision. no quants. nvidia nemotron super 120B-A12B. 120 billion parameters, 12 billion active per token. latent mixture of experts with mamba-2. their GTC flagship. against alibaba's qwen 3.5 122B-A10B. 122 billion parameters, 10 billion active per token. the lineup that has been winning at every tier. both unquantized. both on the same hardware. same octopus invader game test, same agent, same prompt. the only variable is the model. if nvidia can't code at full precision on datacenter hardware it can't code anywhere. downloading now.
Sudo su tweet mediaSudo su tweet mediaSudo su tweet mediaSudo su tweet media
Sudo su@sudoingX

so far i test 2 nvidia models tested on my 3090. two different failures. cascade 2: 187 tok/s, blank screens on every coding test. speed king, can't code. openreasoning 32B dense: 36 tok/s, overthinks everything. prompted hello and it solved math problems for 2 minutes straight. gave it a real build task and it wandered in its own reasoning until it was unusable. 5 million deepseek reasoning traces turned this model into a thinker that forgot how to ship. 3B active MoE couldn't hold coherence. 32B dense couldn't stop reasoning long enough to write code. nvidia's 3090 tier is not it for agent coding. but i am not done. the question is whether nvidia needs scale to compete with qwen or if the architecture itself is the problem. loading the big fight next.

English
17
9
221
10.4K
Fisherman retweetledi
Lotto
Lotto@LottoLabs·
Qwopus 27b in Claude Code is coming soon
English
5
3
70
5.1K
Fisherman retweetledi
Weekend Warrior ☝️🐉🏆🐏
AND it wasn''t just some guy he helped. it was eric wolpaw who was one of the writers in portal 1 and 2, one of the best selling duologies ever. so when gabe does nothing, he wins. and when gabe does something, HE WINS EVEN HARDER
Weekend Warrior ☝️🐉🏆🐏 tweet mediaWeekend Warrior ☝️🐉🏆🐏 tweet mediaWeekend Warrior ☝️🐉🏆🐏 tweet media
DAKKADAKKA@DAKKADAKKA1

Lmao Gabe Newell when faced with an employee with cancer changed their employed status to “get better” and kept them at full salary the entire ordeal and paid out the insurance.

English
4
115
1.1K
17.3K
Fisherman retweetledi
Lotto
Lotto@LottoLabs·
Life moments before installing hermes agent and qwen 27b
Lotto tweet media
English
11
6
259
7.9K
Fisherman retweetledi
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞)
Actually very impressive. Non-reasoner with relatively brief outputs, 110 t/s, $1.2 per 1M output. Overall scores just like Sonnet 4.6 non-reasoning, but completes the suite almost 3x faster and 20x cheaper. I get that AA is noisy, but take a look at new KAT-coder.
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) tweet mediaTeortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) tweet media
Artificial Analysis@ArtificialAnlys

KwaiKAT has released KAT-Coder-Pro V2, a non-reasoning model that scores 44 on the Artificial Analysis Intelligence Index, an 8 point improvement from KAT-Coder-Pro V1 @KwaiAICoder has updated their flagship proprietary coding model with the release of KAT-Coder-Pro V2. KAT-Coder-Pro V2 achieves 44 on the Artificial Analysis Intelligence Index, matching Claude Sonnet 4.6 (non-reasoning) and trailing only Claude Opus 4.6 (non-reasoning, 46) among non-reasoning models. At ~9M output tokens, it is also more token efficient than Claude Opus 4.6 (~11M), Claude Sonnet 4.6 (~14M), and reasoning models with similar intelligence such as DeepSeek V3.2 (reasoning, ~61M) and Qwen3.5 397B A17B (reasoning, ~86M). KAT-Coder-Pro V2 is a non-reasoning model, unlike all of the current frontier language models which ‘think’ before answering. Typically, reasoning variants score higher on the Intelligence Index than their non-reasoning counterparts, but consume more output tokens and are less suited to latency-sensitive workloads. Key Highlights: ➤ 🧠 Higher overall intelligence, but regression in long context reasoning and knowledge recall: KAT-Coder-Pro V2 scores 44 on the Artificial Analysis Intelligence Index, an 8 point improvement from KAT-Coder-Pro V1 and matching Claude Sonnet 4.6 (non-reasoning, max effort). It performs well on tool use (90% on Tau2-Telecom), but regresses compared to KAT-Coder-Pro V1 on long-context reasoning and knowledge, falling 8 p.p. on AA-LCR (66%) and 17 p.p. on HLE (16%). ➤ 🤖 Agentic capability improvements: KAT-Coder-Pro V2 shows major improvements on our agentic evaluations. On Terminal-Bench Hard, it scores 49%, up 40 p.p. from KAT-Coder-Pro V1, making it the highest-scoring non-reasoning model, matching Claude Opus 4.6 (non-reasoning, 49%) and ahead of Claude Sonnet 4.6 (non-reasoning, 46%). KAT-Coder-Pro V2 also shows improvement in GDPval-AA, scoring 1123 (+304 Elo from V1), but still sits behind models such as DeepSeek V3.2 (1198) and Qwen3.5 397B A17B (1202). ➤ ⚙️ High token efficiency: KAT-Coder-Pro V2 is a non-reasoning model and uses fewer tokens than peers with similar intelligence. It uses 8.7M output tokens to run the Artificial Analysis Intelligence Index, below Claude Opus 4.6 (non-reasoning, ~11M) and Claude Sonnet 4.6 (non-reasoning, ~14M), though this is ~2x higher than its predecessor, KAT-Coder-Pro V1 (~4.5M). It also uses significantly fewer tokens than similarly intelligent reasoning models such as DeepSeek V3.2 (reasoning, ~61M) and Qwen3.5 397B A17B (reasoning, ~86M). ➤ $ Improved cost efficiency: KAT-Coder-Pro V2 costs $73 to run the Artificial Analysis Intelligence Index, down from $76 for V1, as it uses fewer input tokens by requiring fewer turns in agentic evaluations. This makes it one of the most cost-efficient models at its intelligence level, costing less than Qwen3.5 397B A17B (reasoning, $418) and Claude Sonnet 4.6 (non-reasoning, $1397). KAT-Coder-Pro V2 is currently priced at $0.30/$1.20 per 1M input/output tokens on StreamLake and AtlasCloud API endpoints. ➤ ⚡ Low end-to-end response time: KAT-Coder-Pro V2 runs at ~109 output tokens per second, far ahead of Claude Opus 4.6 (non-reasoning, 39 OTPS) and Claude Sonnet 4.6 (non-reasoning, 43 OTPS). Because it also has a low time to first token without any reasoning delay, it delivers one of the fastest end-to-end response times, which measures the time taken from request sent to final output returned. Model details: ➤ Availability: KAT-Coder-Pro V2 is available via StreamLake and AtlasCloud API endpoints ➤ Context Window: 256K tokens (equivalent to KAT-Coder-Pro V1) ➤ Multi-modal capabilities: Text input and output only

English
4
4
103
13.8K