NVIDIA AI

12.6K posts

NVIDIA AI banner
NVIDIA AI

NVIDIA AI

@NVIDIAAI

Teaching your AI new tricks.

Santa Clara, CA Katılım Haziran 2016
853 Takip Edilen293.6K Takipçiler
Sabitlenmiş Tweet
NVIDIA AI
NVIDIA AI@NVIDIAAI·
Meet Nemotron 3 Nano Omni 👋 Our latest addition to the Nemotron family is the highest efficiency, open multimodal model with leading accuracy. 30B parameters. 256K context length. 🧵👇
English
89
188
1.3K
444.6K
Brett Adcock
Brett Adcock@adcock_brett·
This is crazy - 2 hours away from 24 hours of continuous humanoid work! The robots have sorted over 28,000 packages so far Bob, Frank, and Gary are all healthy
English
76
50
733
24.4K
0xSero
0xSero@0xSero·
My next big thing.
0xSero tweet media
English
4
1
78
4.5K
NVIDIA AI retweetledi
NVIDIA Data Center
NVIDIA Data Center@NVIDIADC·
💡 Why did @togethercompute choose NVIDIA Blackwell to serve DeepSeek-V4? Because NVIDIA Blackwell is built for the bottlenecks that matter most in long-context inference: → KV-cache pressure during decode → MoE weight bandwidth during prefill A single NVIDIA HGX B200 system can keep DeepSeek-V4’s compressed CSA/HCA/SWA cache layouts resident across many concurrent long-context requests, while native MXFP4 support enables efficient end-to-end quantized inference for V4’s MoE weights. The result? Higher throughput, lower overhead, and optimized serving efficiency at scale.
Together AI@togethercompute

x.com/i/article/2053…

English
5
11
86
14.2K
NVIDIA AI
NVIDIA AI@NVIDIAAI·
Thanks for building with us @CrusoeAI 💚 As context windows explode, tokenization is becoming a major hidden bottleneck in inference pipelines. fastokens is open source, already integrated with Dynamo & @lmsysorg, and designed for the next generation of 100K-token agent systems.
English
1
0
10
843
Crusoe
Crusoe@CrusoeAI·
The tokenizer inside @NVIDIA Dynamo? That's fastokens — built by Crusoe. On NVIDIA GB200 NVL72 it delivers 9.1x average speedup over Hugging Face and up to 40% faster TTFT. We don't just run open-source. We build it. Read the full breakdown: crusoe.ai/resources/blog…
English
1
1
19
1.5K
Prime Intellect
Prime Intellect@PrimeIntellect·
Introducing Renderers RL trainers work in tokens. Environments work in messages. Going back and forth corrupts sampled tokens, wasting compute on every agentic turn. With Renderers, we fix this mismatch. This unlocks >3x throughput on popular open models.
Prime Intellect tweet media
English
15
73
688
182.1K
LangChain
LangChain@LangChain·
Special guests at the Ask me Anything booth. @NVIDIA! Stop by during lunch!
LangChain tweet media
English
2
3
31
3.3K
0xSero
0xSero@0xSero·
I just received a 100,000$ grant from the Human Rights Foundation. In total I received: - 100K USD through HRF - 25.8K USD through donations site - 25K Brev credits through Nvidia - 4x B200s for a month - 5K from lambda - 4x RTX PRO 6000 private donor Open source must win
0xSero@0xSero

x.com/i/article/2034…

English
216
167
3.4K
154.1K
NVIDIA AI
NVIDIA AI@NVIDIAAI·
Delivering agentic inference at scale requires balancing efficiency across: 1) Models and algorithms 2) Software 3) Compute Our full-stack platform continuously optimizes for these inputs using extreme co-design across compute, networking, storage, and memory. Plus, software with broad ecosystem support across millions of developers. The result: lower cost per token, higher throughput, and more scalable AI systems.
English
12
14
107
7.6K
NVIDIA AI
NVIDIA AI@NVIDIAAI·
OpenShell v0.0.40 🔀 local-domain service routing in the gateway ☸️ k8s node scheduling + tolerations 🔒 CLI TLS now uses the OS trust store 🛡️ SecretResolver debug no longer leaks secrets Two security fixes ship alongside new routing and K8s scheduling control. github.com/NVIDIA/OpenShe…
English
5
16
143
9.4K
Joe Guglielmucci
Joe Guglielmucci@JoeGuglielmucci·
@NVIDIAAI Nemoclaw obviously. It's insane to even be testing anything else.
English
1
0
1
69
Sampath Bommakanti
Sampath Bommakanti@sambommakanti·
@NVIDIAAI I just built my home robot to use VSLAM and navigate the environment autonomously. Based on NVIDIA Jetson Orin Nano super dev kit and RGB-D camera. Next to pure voice interaction capability through LLM and NemoClaw to get it to do things and interact with users 💪🏽
English
1
0
1
124
Sudo su
Sudo su@sudoingX·
nobody is talking about how good nemotron 3 nano omni 30b-a3b actually is on local. very underrated. multimodal, reasoning, video understanding, image vision, all shipped in one open source release by nvidia. moe architecture 30b total params, 3b active per token, q8 is near lossless and fits comfortably on a single dgx spark with room to breathe. i have been running it for weeks now and the gap between what this model can do and what the conversation says is wide. nvidia is pushing hard on the open-source front. most builders haven't noticed yet because the discourse is locked on closed-source frontier benchmarks and the next viral chart. meanwhile this thing handles agentic loops, processes video inputs, reasons across image context, and stays responsive on consumer tier unified memory hardware. on dgx spark it flies. more content coming, showing all the modalities in action. if you have used it, what is your experience. drop your stack and your findings, curious what other builders are seeing across hardware tiers.
Sudo su tweet media
English
28
20
236
13.3K
poolside
poolside@poolsideai·
Poolside is hosting a 2-day model research hackathon in London. Join us to push an open-weight agent model as far as you can. RL and fine-tune Laguna XS.2, our latest-generation model, on Prime Intellect Lab. Dates: May 29–30 Partners: @nvidia + @PrimeIntellect + @huggingface Prize: NVIDIA DGX Spark Agents need better models. Better models need cracked researchers. Link below.
English
25
40
210
77.6K