Tony Merguez

2K posts

Tony Merguez banner
Tony Merguez

Tony Merguez

@ACoulbo

Teddy rugged me

Katılım Ağustos 2019
790 Takip Edilen42 Takipçiler
Tony Merguez retweetledi
τroy
τroy@TroyQuasar·
The continuous-time attention that will be used in our upcoming models already outperforms the GPT-OSS 20B model. It is the best attention for stable and faster long-context AI and can process millions of tokens.
Quasar@QuasarModels

This is Quasar Attention, the mechanism behind the upcoming Quasar models, designed to support context lengths of up to 5 million tokens. Attention has long been a bottleneck for processing extended context. Standard attention mechanisms struggle to scale beyond ~200k tokens in training, creating a ceiling on how much information models can reliably use. One approach to solving this has been linear attention methods, such as gated delta attention (used in Qwen 3.5) or Kimi delta attention. These improve efficiency and allow longer sequences, but introduce trade-offs: instability at extreme lengths, quality degradation, and in practice, they are not strictly linear. Quasar Attention takes a different approach. It uses a continuous-time formulation, implemented as a fully matrix-based system rather than relying on vector-state approximations. In practice, this improves stability, reduces cost, and maintains performance as sequence length increases. In internal stress tests at 50 million tokens, KDA-based approaches begin to lose stability, while Quasar Attention remains stable. This allows performance to hold as sequence length increases, rather than degrading beyond a fixed threshold. On BABILong, a Quasar-based model pretrained on 20B tokens and fine-tuned on 16k sequences was evaluated on contexts ranging from 1 million to 10 million tokens, maintaining consistent performance across that range. By contrast, models using gated delta attention show significant degradation at longer lengths, in some cases dropping to ~10% performance at 10 million tokens. (Note: results are indicative; setups are not directly comparable) On RULER benchmarks, a Quasar-10B model (built on Qwen 3.5 with frozen base weights and Quasar Attention added), pretrained on 200B tokens, achieved 87% at 1 million tokens, outperforming significantly larger baselines, including Qwen3 80B, under the same evaluation conditions. Taken together, this points to a shift in where long-context performance is won or lost: not in model size alone, but in the attention mechanism itself. Quasar Attention represents a step change in long-context modelling, setting a new standard for stability and performance at scale. We thank @TargonCompute for the compute and for being our compute provider and long-term partner in training the upcoming Quasar models Here is the link to our paper 👇

English
3
14
50
1.8K
Tony Merguez retweetledi
Quasar
Quasar@QuasarModels·
This is Quasar Attention, the mechanism behind the upcoming Quasar models, designed to support context lengths of up to 5 million tokens. Attention has long been a bottleneck for processing extended context. Standard attention mechanisms struggle to scale beyond ~200k tokens in training, creating a ceiling on how much information models can reliably use. One approach to solving this has been linear attention methods, such as gated delta attention (used in Qwen 3.5) or Kimi delta attention. These improve efficiency and allow longer sequences, but introduce trade-offs: instability at extreme lengths, quality degradation, and in practice, they are not strictly linear. Quasar Attention takes a different approach. It uses a continuous-time formulation, implemented as a fully matrix-based system rather than relying on vector-state approximations. In practice, this improves stability, reduces cost, and maintains performance as sequence length increases. In internal stress tests at 50 million tokens, KDA-based approaches begin to lose stability, while Quasar Attention remains stable. This allows performance to hold as sequence length increases, rather than degrading beyond a fixed threshold. On BABILong, a Quasar-based model pretrained on 20B tokens and fine-tuned on 16k sequences was evaluated on contexts ranging from 1 million to 10 million tokens, maintaining consistent performance across that range. By contrast, models using gated delta attention show significant degradation at longer lengths, in some cases dropping to ~10% performance at 10 million tokens. (Note: results are indicative; setups are not directly comparable) On RULER benchmarks, a Quasar-10B model (built on Qwen 3.5 with frozen base weights and Quasar Attention added), pretrained on 200B tokens, achieved 87% at 1 million tokens, outperforming significantly larger baselines, including Qwen3 80B, under the same evaluation conditions. Taken together, this points to a shift in where long-context performance is won or lost: not in model size alone, but in the attention mechanism itself. Quasar Attention represents a step change in long-context modelling, setting a new standard for stability and performance at scale. We thank @TargonCompute for the compute and for being our compute provider and long-term partner in training the upcoming Quasar models Here is the link to our paper 👇
Quasar tweet media
English
23
80
247
104.1K
doc🃏
doc🃏@jkrdoc·
shorting $tao here > purely speculative / hype atm > war aint over, bearish on stocks > bitcoin will trend down in april lets see if betting against the herd will pay out. 🫡
doc🃏 tweet media
English
76
8
171
40.1K
Tony Merguez retweetledi
Targon
Targon@TargonCompute·
We needed to run trusted workloads on untrusted host machines. So over a year ago, we started building the Targon Virtual Machine to enable Confidential TEEs in production. Today we're sharing our white paper written alongside @intel: Decentralized Compute on Untrusted Hardware Using Intel® TDX and Encrypted CVMs
Targon tweet media
English
59
207
900
280.6K
Tony Merguez retweetledi
Raphael Barberi📷
Raphael Barberi📷@itsrapha83·
“I want a house boat” “You mean a boat house?” “No.”
Raphael Barberi📷 tweet media
English
5
14
215
4K
Tony Merguez retweetledi
Fearnot
Fearnot@Fearnot_10·
Spring is arriving and so is $CYPHR. @CypherGangNFT 's, @CypherBrawls, @Blastpadxpr staking and the Bid and Win parties are just a few ways to earn. Can't forget the telegram where it sometimes rains $CYPHR. Come join the fun! t.me/CypherGang
Fearnot tweet media
English
0
8
17
500
0xPiracyData
0xPiracyData@0xPiracyData·
Smokey, your reaction actually proves the point I made in my post about ego. You’re interpreting this as a personal attack and a shift in position, while as a founder (the captain of the ship) you also carry responsibility for the people who helped - and are still helping - to build the ecosystem. A change of position? It’s not that simple. Would I like to continue supporting @Berachain? Yes. I still haven’t sold anything - not my $BERA, not $HENLO, not $YEET, and not a single NFT from the dozens of collections I hold. But can I continue investing in Berachain given my financial losses? Most likely not. We can talk a lot about ideas and higher values, but we all need money to live. And if I saw real changes for the better - real potential in the ecosystem - I would gladly continue doing so. A couple of weeks ago, the Berachain Foundation disqualified 121 reward vaults, including the Henlo community’s Henlockers. Can this be called support? Can this be considered a reasonable action aimed at helping the project survive? I don’t think so. Product–market fit? In my view, The Honey Jar did a tremendous amount for onboarding users. Their products clearly had PMF, as evidenced by the strong support from the community. Is Jani responsible for certain mistakes? Absolutely - and I apply the same criticism to his projects. The same goes for other projects that make mistakes. Let me repeat: the opportunity for improvement begins when criticism is not treated as a personal attack, but as a way to make things better.
English
2
1
10
1.1K
0xPiracyData
0xPiracyData@0xPiracyData·
My bet on @berachain turned out to be a complete failure. I lost about $100k across multiple projects such as Henlo, Mibera, ApiologyDAO, Smilee, @eatsleepyeet , Burrbear, and Badboyz by grifter @MaxLongCEO. None of these projects delivered what they promised. This is an example of how a really good project can be ruined by grifting, poor communication, and ego. And I don't really blame projects (except of Burrbear and BadBoyz), these projects were affected by decisions of @SmokeyTheBera and the team.
English
15
3
75
11.5K
Tony Merguez
Tony Merguez@ACoulbo·
@Raclure03 La boîte à gants ? Ah ben non elle est restée dans la voiture
Français
0
0
0
113
🅶🅾🆁🅺
🅶🅾🆁🅺@Raclure03·
J'adore ❤️❤️😎
Français
55
179
1.2K
120.5K
Zantec
Zantec@P7MarcP7·
@LeBoonnnCopainn @Psyhodelikus si tu ne sais pas faire la différence entre sponsor d'une émission et financer une chaine je ne peux rien pour toi dsl
Français
3
0
18
4.5K
Tony Merguez retweetledi
Score - Subnet 44
Score - Subnet 44@webuildscore·
MANIFEST is now live on the Score subnet. Miners will no longer work on single static tasks. They now operate using MANIFESTS - structured configurations that define what a vision model should detect, how performance is measured, latency requirements, and trust parameters. This enables scalable, real-world vision deployment across multiple industries simultaneously.
Score - Subnet 44 tweet media
English
7
18
136
21.7K
Tony Merguez retweetledi
Openτensor Foundaτion
Openτensor Foundaτion@opentensor·
This is exactly why decentralized AI wins. When you build in public, you don't get "stolen" — you get built upon. Bittensor is intelligence that belongs to everyone, and controlled by no one. τ
Anthropic@AnthropicAI

We’ve identified industrial-scale distillation attacks on our models by DeepSeek, Moonshot AI, and MiniMax. These labs created over 24,000 fraudulent accounts and generated over 16 million exchanges with Claude, extracting its capabilities to train and improve their own models.

English
31
125
626
30.7K
Jean-Luc Mélenchon
Jean-Luc Mélenchon@JLMelenchon·
Le père du logiciel de conception de l'ère moderne CATIA, Bernard Charles, a quitté la présidence de Dassault Systèmes qui le produisait. Une pensée de reconnaissance au nom des millions de lycéens professionnels en France qui ont appris leur métier en l'utilisant gratuitement. Bonne chance à son successeur !
Français
75
247
1.8K
150.4K
Tony Merguez retweetledi
Score - Subnet 44
Score - Subnet 44@webuildscore·
We just updated the @manakoai website. You can now explore real use cases across 7 verticals and see how businesses can use Manako to turn camera data into real time insights, alerts, and automation. See what you can build with Manako ⤵️ manako.ai
Score - Subnet 44 tweet media
English
2
14
102
5.4K
Tony Merguez retweetledi
Victor VL
Victor VL@Victor_crypto_2·
Bittensor Ecosystem Highlights of the Week #48 // SUBNET UPDATES & ACHIEVEMENTS ➤ @IOTA_SN9 SN9 @MacrocosmosAI’s Train at Home app is now available to everyone on Mac. (bit.ly/3Zy7gYh) ➤ @basilic_ai SN39 Basilica now has one-click deployments with native TAO deposits for @openclaw agents. (bit.ly/4rQVnIL) ➤ @basilic_ai SN39 x @gradients_ai SN56 Gradients is now running reinforcement learning evaluations on Basilica’s infrastructure. (bit.ly/4kygVYm) ➤ @yanez__ai SN54 Yanez signed their first client for their Biometric Data API subscription product, which is their second live product. (x.com/yanez__ai/stat…) ➤ @redteam SN61 RedTeam quietly shipped 26 releases / 2,000+ commits, landed major production integrations (100M+ DAU + 1inch as a paying customer), and committed to subnet buybacks starting at 1 TAO/day. (x.com/_redteam_/stat…) ➤ @chutes_ai SN64 Following their latest Novelty Search, they announced Chutes Search, their AI search engine. (x.com/chutes_ai/stat…) Thanks to ParaClaw, you can give your OpenClaw agent access to 60+ models. (x.com/chutes_ai/stat…) GLM-5 is now live on Chutes. (x.com/chutes_ai/stat…) ➤ @SynthdataCo SN50 Synth explained how to launch a Polymarket trading bot using their API with OpenClaw. (x.com/SynthdataCo/st…) The Synth MCP server is now live. (x.com/SynthdataCo/st…) ➤ @Loosh_ai SN78 The Loosh subnet portal is now live. (x.com/Loosh_ai/statu…) They’ve also been accepted into the NVIDIA Inception Program. (x.com/Loosh_ai/statu…) ➤ @numinous_ai SN6 Numinous is releasing their Eversight API. (x.com/numinous_ai/st…) ➤ @platform_tao SN100 They’re teasing their upcoming agentic IDE, Cortex, launching on February 20th. (x.com/platform_tao/s…) ➤ @v0idai SN106 Void 2.0, powered by Chainlink CCIP, is now live. (x.com/v0idai/status/…) ➤ @Bitcast_network SN93 Bitcast partnered with @FluenceMGMT, a crypto creator marketing agency. (x.com/Bitcast_networ…) ➤ @djinn_gg SN103 Djinn released their updated whitepaper. (x.com/djinn_gg/statu…) ➤ @SwarmSubnet SN124 They showcased their drone holding an autonomous position, powered by a model trained on their subnet. (x.com/SwarmSubnet/st…) ➤ @TatsuEcosystem SN84 Tatsu dropped a new website for their ChipForge subnet. (x.com/TatsuEcosystem…) ➤ @bitsecai SN60 Their V2.2 incentive mechanism is still improving, from 35% last week to 70% this week. (x.com/bitsecai/statu…) Bitsec found a lot of exploits in OpenClaw’s codebase. (x.com/bitsecai/statu…) ➤ @resilabsai SN46 Their new IM has been live for just a few days, and the top model on RESI improved from 46% to 78% accuracy. (x.com/resilabsai/sta…) ➤ @babelbit SN59 Babelbit announced a complete redesign of their website. (x.com/babelbit/statu…) // SUBNET LAUNCHES ➤ @handshake_58 SN58 Handshake is the 4th @bitstarterAI-incubated subnet, building trustless micropayments for autonomous AI agents. (x.com/handshake_58/s…) ➤ @GenomesDAO SN55 NIOME is the new Yuma-accelerated subnet, built to advance ethical, scalable research in drug response and personalized medicine. (x.com/YumaGroup/stat…) // BITTENSOR ECOSYSTEM ➤ @opentensor Const stepped down as CEO of OpenTensor to make Bittensor more decentralized and censorship-resistant, but his involvement and dedication to Bittensor will stay the same. (x.com/const_reborn/s…) ➤ @YumaGroup x @bitget Yuma partnered with Bitget to bring seamless $TAO staking to their users through their validator service provider. (x.com/YumaGroup/stat…) ➤ @crunchDAO Crunch introduced their subnet mining hub to simplify access to Bittensor’s decentralized AI mining ecosystem. (x.com/crunchDAO/stat…) ➤ @taostats You can now plug Bittensor data and docs into your OpenClaw agent thanks to the Taostats MCP Server. (x.com/mogmachine/sta…) // PODCASTS & ARTICLES ➤ @opentensor Novelty Search on Bittensor Governance (x.com/opentensor/sta…) ➤ @TAO_dot_com Episode 7 with @Carrot_____1 and @KeithSingery (x.com/TAO_dot_com/st…) ➤ @YumaGroup subnet spotlight with @numinous_ai (x.com/YumaGroup/stat…) ➤ Revenue Search 56 by @SiamKidd and @MarkCreaser with Numinous (x.com/numinous_ai/st…) ➤ Revenue Search 57 with @TPN_Labs (x.com/MarkCreaser/st…) ➤ @VenturaLabs Ep. 80 with @Sebyverse from RESI (x.com/VenturaLabs/st…) ➤ Hash Rate 157 by @markjeffrey (x.com/markjeffrey/st…) $TAO
Victor VL tweet media
English
6
19
79
13K
Tony Merguez retweetledi
Bittensor France 🇫🇷
Bittensor France 🇫🇷@bittensor_fr·
1 000 abonnés ! Merci à tous de nous lire au quotidien ❤️ Pour marquer le coup, on vous présente : bittensorfrance.fr 🇫🇷 L'objectif ? Rendre la découverte de Bittensor plus accessible aux francophones. On a pensé une interface simple et intuitive pour découvrir le réseau et explorer les subnets en français. 🚧 C'est une V1 : le site est amené à évoluer grâce à vos retours. Alors vos critiques et vos idées en commentaires sont les bienvenues pour une future V2 ! Merci encore et bonne exploration à tous 🙏
Bittensor France 🇫🇷 tweet media
Français
12
22
122
26.2K
Tony Merguez retweetledi
Openτensor Foundaτion
Openτensor Foundaτion@opentensor·
Reminder: Keep building. Think it. Pitch it. Build it. τ Bittensor Subnet Ideathon (Round 1) closes Feb 25. - $18K cash prize pool - Up to 𝞃1000 discretionary funding from @UnsupervisedCap - $5K @basilic_ai compute credits - Direct entry to @bitstarterai accelerator Registration live → build the next subnet on Bittensor. hackquest.io/hackathons/Bit… Own the future. #Bittensor $TAO #SubnetIdeathon @hackquest_
English
54
51
285
73.2K
Tony Merguez retweetledi
Zero Knowledge Network
Zero Knowledge Network@_zknet·
ZKNetwork Project Update – 2026-01-23 The Zero Knowledge Network continues advancing into 2026, shaping its trajectory as a full-stack, privacy-first infrastructure platform. This cycle brings progress on modular protocol packaging, privacy-preserving consensus research, and a closer look at how zero-knowledge access control transforms real-world security. ➡️ Platform Strategy: Privacy as Infrastructure As ZKNetwork matures toward broader market engagement, the team is packaging core capabilities into modular, composable offerings for developers, enterprises, and partners. This reflects a core conviction: privacy is a fundamental public utility, not a premium feature. Value flows through infrastructure participation: registration anchors identity, governance enables coordination, and service guarantees meet enterprise requirements, all while verification remains freely accessible to all. The platform is crystallizing around distinct layers: ZK Trust Registry for privacy-preserving cryptographic identity and attestation, ZK-Firewall for access control via zero-knowledge proofs, ZK Verification Services for programmable compliance logic, ZK Provenance Services for supply-chain integrity, and Secure Edge Hardware extending zero-knowledge guarantees into the physical world. These modules are designed for independent adoption or seamless composition, connecting ZKNetwork's privacy substrate to real-world deployment across consumers, developers, and regulated industries. ➡️ R&D: Privacy-Preserving Consensus The team continues pushing programmable privacy boundaries with Aztec and Noir. On-Chain Content Identifiers: A new Noir contract and TypeScript library enables storing variable-length content identifiers on-chain with Poseidon2 hashing, providing a foundation for verifiable references to off-chain data without revealing underlying content. Commit-Reveal Voting: Building on this primitive, the team is implementing commit-reveal patterns for ZKNetwork's distributed PKI consensus. Participants commit a vote hash, then reveal after commitment closes, protecting voting integrity while enabling trustless verification. Private Operator-Node Association: Development continues on secure associations between human Operators and their always-online Nodes, enabling nodes to maintain public operational identities while keeping operators cryptographically hidden. The voting research directly informs how these agents participate in automated systems while operators manage financial aspects securely. ➡️ Use-Case Spotlight: Access Without Identity Traditional access control creates a paradox: to prove you belong, you must reveal who you are. ZK-Firewall dissolves this tradeoff. Consider age verification. Today's methods require submitting IDs, creating databases of who accessed what and when. With ZK-Firewall, users prove they meet the threshold without revealing birthdate, name, or any identifying information. The service learns one bit: "eligible" or not. The pattern extends to residency verification without exposing addresses, credential checks without revealing identity, and enterprise resource gating without creating subpoena-vulnerable access logs. ZK-Firewall transforms access control from surveillance mechanism to privacy-preserving utility. Organizations reduce regulatory exposure while users interact without leaving exploitable trails. ➡️ Building Forward ZKNetwork's modular architecture reflects deliberate strategy: build foundational primitives the network requires, then package them as standalone protocols with enterprise-ready expressions. Like branches from a rooted trunk, each capability extends reach while sharing common cryptographic foundations. Stay tuned and stay private! BE THE N E T W O R K…
Zero Knowledge Network tweet media
English
9
121
46
1.6K
Tony Merguez retweetledi
Victor VL
Victor VL@Victor_crypto_·
J'ai lancé le live j'arrive à 21h : youtube.com/watch?v=vEM_X0… ➤ La limites à 128 subnets est-elle légitime ? ➤ Alpha & News subnets ➤ Dépression sur le marché (pour changer un peu) ➤ $TAO va-t-il à 10000000000000000$ en 2026 ?
YouTube video
YouTube
Victor VL tweet media
Français
1
4
22
1.7K