Marvin Gabler

1.2K posts

Marvin Gabler banner
Marvin Gabler

Marvin Gabler

@bigmarvin

liquidity provider, founder @ https://t.co/3ldLWkThF1

Zürich, Schweiz Katılım Temmuz 2022
1.3K Takip Edilen258 Takipçiler
Sabitlenmiş Tweet
Marvin Gabler
Marvin Gabler@bigmarvin·
Tired: Building the 5000st language model Wired: Building a digital twin of the earth, learning physics and simulating our nature techcrunch.com/2024/02/05/jua…
English
0
2
13
4.1K
Marvin Gabler
Marvin Gabler@bigmarvin·
@DeItaone The average price and starts at are 2 fundamentally different things Sherlock, a basic Bentley starts at 180k
English
0
0
0
485
*Walter Bloomberg
*Walter Bloomberg@DeItaone·
BENTLEY KYIV RANKS 3RD IN THE WORLD Bentley Kyiv has achieved 3rd place out of 61 Bentley dealerships worldwide in the prestigious “Best of the Best” nomination, according to the Business Performance Matrix. The average price of a Bentley starts at $400,000.
*Walter Bloomberg tweet media
English
197
249
2K
440.8K
Marvin Gabler
Marvin Gabler@bigmarvin·
If inference was near free and instant (meaning you could run 1000s of agents in parallel in milliseconds), do you have any doubts we could brutforce agi?
English
1
0
1
50
Marvin Gabler
Marvin Gabler@bigmarvin·
Our new combat drone prototype, wdyt?
Marvin Gabler tweet media
English
0
0
0
24
Marvin Gabler
Marvin Gabler@bigmarvin·
@MarvinTBaumann Of course dehumanizing ad hominem, German heritage. The cognitive dissonance caused by your world view clashing with reality must be painful
English
0
0
1
23
Marvin Gabler
Marvin Gabler@bigmarvin·
@mtxpost @paularambles Nah, only the pseudo intellectuals were unhappy, real life losers with good ideas. Since Ancient Greece we know talking is cheaper than walking
English
0
0
0
62
Sam Clemens
Sam Clemens@mtxpost·
@paularambles Historically, in literature, intelligent people are never happy. Hamlet, Dedalus, Raskolnikov, Frankenstein, Holmes, Solomon, Faust, Finch, etc... Intelligence always leads to misery, grief, or something similar. Ignorance is bliss.
English
15
29
775
31.4K
Eneasz Brodski
Eneasz Brodski@EneaszWrites·
Half the local rats are at the Stop The AI Race March
Eneasz Brodski tweet mediaEneasz Brodski tweet mediaEneasz Brodski tweet mediaEneasz Brodski tweet media
English
141
33
840
1.3M
kache
kache@yacineMTB·
slavic accent living in Switzerland, 1000 views, talking about technical concept, high rise apartment. many such cases
English
12
3
351
33.8K
Marvin Gabler
Marvin Gabler@bigmarvin·
So that I may perceive whatever holds / The world together in its inmost folds.
Demis Hassabis@demishassabis

@elonmusk I feel like it might be possible with the help of AI tools to find some very elegant and compact descriptions that explain some of the deepest mysteries of the universe, but it might take a lot of pattern processing and matching to get there...

English
0
0
0
96
Marvin Gabler
Marvin Gabler@bigmarvin·
pretty fun seeing your agents visually jumping around your code base
English
0
1
1
87
Theo - t3.gg
Theo - t3.gg@theo·
Since OpenAI dropped gpt-oss-120b, Mistral has released 4 models that are worse than gpt-pss-120b
Artificial Analysis@ArtificialAnlys

Mistral has released Mistral Small 4, an open weights model with hybrid reasoning and image input, scoring 27 on the Artificial Analysis Intelligence Index @MistralAI's Small 4 is a 119B mixture-of-experts model with 6.5B active parameters per token, supporting both reasoning and non-reasoning modes. In reasoning mode, Mistral Small 4 scores 27 on the Artificial Analysis Intelligence Index, a 12-point improvement from Small 3.2 (15) and now among the most intelligent models Mistral has released, surpassing Mistral Large 3 (23) and matching the proprietary Magistral Medium 1.2 (27). However, it lags open weights peers with similar total parameter counts such as gpt-oss-120B (high, 33), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, 36), and Qwen3.5 122B A10B (Reasoning, 42). Key takeaways: ➤ Reasoning and non-reasoning modes in a single model: Mistral Small 4 supports configurable hybrid reasoning with reasoning and non-reasoning modes, rather than the separate reasoning variants Mistral has released previously with their Magistral models. In reasoning mode, the model scores 27 on the Artificial Analysis Intelligence Index. In non-reasoning mode, the model scores 19, a 4-point improvement from its predecessor Mistral Small 3.2 (15) ➤ More token efficient than peers of similar size: At ~52M output tokens, Mistral Small 4 (Reasoning) uses fewer tokens to run the Artificial Analysis Intelligence Index compared to reasoning models such as gpt-oss-120B (high, ~78M), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, ~110M), and Qwen3.5 122B A10B (Reasoning, ~91M). In non-reasoning mode, the model uses ~4M output tokens ➤ Native support for image input: Mistral Small 4 is a multimodal model, accepting image input as well as text. On our multimodal evaluation, MMMU-Pro, Mistral Small 4 (Reasoning) scores 57%, ahead of Mistral Large 3 (56%) but behind Qwen3.5 122B A10B (Reasoning, 75%). Neither gpt-oss-120B nor NVIDIA Nemotron 3 Super 120B A12B support image input. All models support text output only ➤ Improvement in real-world agentic tasks: Mistral Small 4 scores an Elo of 871 on GDPval-AA, our evaluation based on OpenAI's GDPval dataset that tests models on real-world tasks across 44 occupations and 9 major industries, with models producing deliverables such as documents, spreadsheets, and diagrams in an agentic loop. This is more than double the Elo of Small 3.2 (339) and close to Mistral Large 3 (880), but behind gpt-oss-120B (high, 962), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, 1021), and Qwen3.5 122B A10B (Reasoning, 1130) ➤ Lower hallucination rate than peer models of similar size: Mistral Small 4 scores -30 on AA-Omniscience, our evaluation of knowledge reliability and hallucination, where scores range from -100 to 100 (higher is better) and a negative score indicates more incorrect than correct answers. Mistral Small 4 scores ahead of gpt-oss-120B (high, -50), Qwen3.5 122B A10B (Reasoning, -40), and NVIDIA Nemotron 3 Super 120B A12B (Reasoning, -42) Key model details: ➤ Context window: 256K tokens (up from 128K on Small 3.2) ➤ Pricing: $0.15/$0.6 per 1M input/output tokens ➤ Availability: Mistral first-party API only. At native FP8 precision, Mistral Small 4's 119B parameters require ~119GB to self-host the weights (more than the 80GB of HBM3 memory on a single NVIDIA H100) ➤ Modality: Image and text input with text output only ➤ Licensing: Apache 2.0 license

English
89
31
1.9K
148.9K