Mark McQuade

302 posts

Mark McQuade banner
Mark McQuade

Mark McQuade

@MarkMcQuade

CEO and founder of @arcee_ai | @huggingface 🤗 alum. AI and Data Obsessed. Fitness Fanatic. Tattoo Enthusiast.

Miami, FL Se unió Şubat 2021
1K Siguiendo826 Seguidores
Tweet fijado
Mark McQuade
Mark McQuade@MarkMcQuade·
Today we drop Trinity-Large-Thinking. SOTA on Tau2-Airline, frontier-class on Tau2-Telecom, and the #2 model on PinchBench, right behind Opus. On BCFLv4, we're in the mix with the best. 26 people with under $50M raised and a ruthless pursuit of greatness. What this team just pulled off is nothing short of incredible. One hell of an accomplishment and I couldn't be more proud of Arcee. And we've got more to prove.
Arcee.ai@arcee_ai

Today we're releasing Trinity-Large-Thinking. Available now on the Arcee API, with open weights on Hugging Face under Apache 2.0. We built it for developers and enterprises that want models they can inspect, post-train, host, distill, and own.

English
11
19
189
11.6K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
A quick update on Trinity-Large-Preview: we're deprecating our hosted preview endpoint on April 22, 2026. The response to Trinity-Large-Thinking has been incredible, and while we work to grow our long-term serving capacity, we need to prioritize reliability for paid users.
English
2
4
108
9.2K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
OpenClaw 2026.4.7 just dropped, and Arcee is officially a bundled provider. The new bundled provider plugin includes our full catalog as well as OpenRouter support. Select Trinity-Large-Thinking from the drop-down and give it a run. Check out Trinity-Large-Thinking x @openclaw in action in the video below. 👇
OpenClaw🦞@openclaw

OpenClaw 2026.4.7 🦞 🔮 openclaw infer 🎬 music + video editing 💾 session branch/restore 🔗 webhook-driven TaskFlows 🤖 Arcee, Gemma 4, Ollama vision 🧠 memory-wiki: persistent knowledge, not just vibes Because “trust me bro” is not a knowledge system. github.com/openclaw/openc…

English
3
7
75
5.5K
Mark McQuade retuiteado
John T. Chambers
John T. Chambers@JohnTChambers·
Trinity-Large-Thinking from JC2-backed @arcee_ai signals the shift from outsourcing intelligence to controlling it directly. Strategic autonomy + data you can stand behind will decide what AI actually gets deployed. AI is infrastructure now. venturebeat.com/technology/arc…
English
0
1
4
329
Mark McQuade retuiteado
Alex Cheema
Alex Cheema@alexocheema·
Yes, @arcee_ai should run well on Apple Silicon RDMA clusters with @exolabs. It’s a 398B model, 13B active parameters (very sparse so great for Apple Silicon). It’s natively 16-bit at ~800GB so you’d need 4 x 256GB Mac Studios to run the full model. You can run it in 6-bit on 128GB MBP + 256GB Mac Studio.
Beff (e/acc)@beffjezos

Ayo @exolabs can I run a @arcee_ai Trinity-Large-Thinking on a 128GB MBP and a 256GB Mac Studio?

English
1
3
68
9.6K
Mark McQuade retuiteado
clem 🤗
clem 🤗@ClementDelangue·
Let’s go @arcee_ai (and startups in general)!
clem 🤗 tweet media
English
9
9
150
11.7K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
Arcee x @NousResearch Trinity x Hermes Agent
English
6
20
200
41.9K
Mark McQuade retuiteado
VentureBeat
VentureBeat@VentureBeat·
Arcee's new, open source Trinity-Large-Thinking is the rare, powerful U.S.-made AI model that enterprises can download and customize venturebeat.com/ai/arcees-new-…
English
1
8
20
2.6K
Mark McQuade retuiteado
Sam Fraser
Sam Fraser@samfrannn·
Display the api call for any prompt executed in chat. Copy the call directly into your terminal & update with your key to get started.
English
1
2
19
1.7K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
Jensen showcased PinchBench (by @kilocode) on stage at NVIDIA GTC as the new standard for evaluating @openclaw agent capability. Trinity-Large-Thinking just hit #2 on @pinchbench globally (91.9%) behind only Claude Opus 4.6 (93.3%), which costs ~20x more per token. That's an expensive percentage point... To celebrate the ongoing Trinity partnership, @OpenRouter is running a promotion to make Trinity-Large-Thinking free to use for OpenClaw until Sunday, April 5th. Even when the promo ends, the Trinity economics are pretty incredible. Trinity is $0.25/m input tokens (OpenClaw uses A LOT of input tokens), but with a 60%-90% cache hit rate at $0.06/m cache tokens the average input cost nets out to $0.087/m. You no longer have to compromise on logic to keep your agent infrastructure affordable. We're excited to see what you build.
Arcee.ai tweet media
English
14
19
118
12.7K
Mark McQuade retuiteado
Paddy Srinivasan
Paddy Srinivasan@paddix·
Inference is evolving fast—and so are the models powering it. With @arcee_ai's Trinity, we’re seeing a new class of open-weight reasoning models: massive capability, radically more efficient economics. 👉 #2 on PinchBench (Kilo’s benchmark for agentic capability) 👉 ~96% lower cost than the top model That changes the equation. Now you can run advanced reasoning + agentic workloads on @digitalocean. at scale. This is the next wave of AI: thinking + doing, in one place. Check it out 👇 digitalocean.com/blog/run-advan… #AI #Inference #AgenticAI #OpenSource #DigitalOcean #LLMs
English
2
6
36
2K
Mark McQuade retuiteado
DigitalOcean
DigitalOcean@digitalocean·
Now in Public Preview: @arcee_ai's Trinity Large-Thinking. ✨ The #1 U.S. model on OpenRouter. 3.4T tokens processed. 400B parameters. Live on DigitalOcean’s Serverless Inference. 🌊
English
5
2
18
3K
Mark McQuade retuiteado
Morgan
Morgan@morganlinton·
Everyone is talking about Kimi and Qwen, but I'm honestly surprised more people aren't talking about models like Trinity from Arcee. I've been doing a deeper dive here and it's pretty interesting, here's a few differences that I'm not sure ppl fully realize. - Qwen and Kimi both have Apache licenses with restrictions in them. I'm honestly shocked most people don't know this. Once you hit over 100M monthly users, things change. - Arcee's models like Trinity use Apache 2.0 with NO restrictions, none, get over 100M monthly users, you're still in a good place. - The training data for Qwen and Kimi is undisclosed, for Arcee it's both disclosed and legally vetted. - One genuinely unique thing Arcee does: they released Trinity-Large-TrueBase, a raw 10-trillion-token checkpoint that hasn't undergone learning rate anneals or instruction tuning, letting researchers in regulated industries start from scratch for authentic audits and custom alignment. As I've been tinkering around with more small(ish) local LLMs, I've been doing a deeper dive, and honestly, I think I'm going to be doing more with Arcee vs. Kimi or Qwen because I can't help but think about scale. If I build something, I'd love to think that some day hundreds of millions, or billions of people, might use it, and I don't want to suddenly be dealing with license restrictions in this case and need to rearchitect. That being said, still pretty new to the Arcee models so I'll have to see how they perform. More to come, as always, I have a lot more to learn, and will share what I learn on here.
English
5
1
29
2.2K
Mark McQuade retuiteado
Lucas Atkins
Lucas Atkins@latkins·
A big advantage of open weights is that companies can adapt models to their users. Cursor not disclosing its base models is frustrating, but it is not “embarrassing”, and it does not diminish how impressive its RL scaling has been in such a short time.
English
2
5
62
5.7K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
We’re looking forward to the @PrimeIntellect Day event tomorrow (3.14) in SF (always love a good Pi pun). Arcee CTO @latkins will be giving some updates on Trinity and the realities of scaling large-scale pre-training. He’ll be speaking alongside an impressive lineup of builders and researchers from Perplexity, Cognition, Hugging Face, and the team at Prime Intellect. The venue is officially booked out, but the waitlist is still open. If you’re in the Bay Area, it's worth jumping on the list in case a spot opens up. We’d love to see you there. luma.com/zjpepfzn
English
2
2
36
5.8K
Mark McQuade retuiteado
Arcee.ai
Arcee.ai@arcee_ai·
OpenClaw exploded into the ecosystem late January, right around the exact same time we launched Trinity Large Preview. It’s been incredible to watch these two projects scale in parallel over the last few months. Looking at the latest @openclaw usage on @OpenRouter it’s surreal to see Trinity Large Preview processing more volume than heavyweights like Claude, Gemini, GPT, and Grok. A huge thank you to the community for trusting Arcee as the model layer for this inventive new agentic crustacean.🦞
Arcee.ai tweet mediaArcee.ai tweet media
English
4
4
42
4.4K