Ksenia_TuringPost

18.7K posts

Ksenia_TuringPost banner
Ksenia_TuringPost

Ksenia_TuringPost

@TheTuringPost

Newsletter exploring AI&ML - AI 101, Agentic Workflow, Business insights. From ML history to AI trends. Led by @kseniase_ Know what you are talking about👇🏼

Join over 102,000 readers انضم Haziran 2020
11.4K يتبع82.9K المتابعون
Ksenia_TuringPost أُعيد تغريده
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
NVIDIA's Nemotron 3 is an architectural response to the 2 pressures: - Long-context cost as agentic interactions scale - Repeated reasoning cost from invoking full models for small subtasks Nemotron 3 proposes several design decisions to solve this: ▪️ Hybrid architecture: Transformer + Mamba 2 layers for efficient long-context processing ▪️ Mixture-of-Experts (MoE) and LatentMoE on top of it to get cheaper experts ▪️ Multi-token prediction ▪️ NVFP4 precision = 4.75 bits used for inference and pre-training, allowing Nemotron pre-training dataset achieve up to 4× faster convergence than standard open web datasets. This is all about one key idea – "Acceleration is intelligence" Here is the tech stack explained and what the Nemotron Coalition is – NVIDIA has just announced that this alliance of leading players like Cursor, Mistral, Black Forest Labs, etc., is gathering to develop the Nemotron family of open models → turingpost.com/p/nemotroncoal…
Ksenia_TuringPost tweet media
English
4
14
105
6.1K
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
Have a feeling that Google is also working towards something like that uniting all the experiences around antigravity, Google AI studio, Gemini web experience and maybe even AI mode
Berber Jin@berber_jin1

SCOOP - OpenAI is planning to simplify its product experience and launch one "superapp" -- part of its broader effort to instill more discipline and focus into the business, and beat back the threat posed by Anthropic more here in our @WSJ story wsj.com/tech/openai-pl…

English
2
0
6
572
clem 🤗
clem 🤗@ClementDelangue·
Nvidia just crossed Google as the biggest org on @huggingface with 3,881 team members on the hub. I'm officially calling it: Nvidia is the new American king of open-source AI!
clem 🤗 tweet media
English
43
84
765
113.5K
Ksenia_TuringPost أُعيد تغريده
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
It was a busy week @NVIDIAGTC! Celebrating my birthday on the road 🎉
Ksenia_TuringPost tweet media
English
4
1
33
2.5K
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
A new paper from @ylecun and others – V-JEPA 2.1 It changes the recipe of V-JEPA so the model learns both: • Global semantics – what is happening in the scene • Dense spatio-temporal structure – where things are and how they move The idea is to supervise not just masked tokens but the visible ones too There are 4 key ingredients for V-JEPA 2.1: - Dense prediction loss on both masked and visible tokens - Deep self-supervision across intermediate layers - Modality-specific tokenizers (2D for images, 3D for videos) within a shared encoder - Model + data scaling The workflow turns into: masked image/video → encode visible tokens → predict latent representations for both masked and visible tokens → supervise at multiple layers Here are the details:
Ksenia_TuringPost tweet media
English
7
45
248
41.2K
Ksenia_TuringPost أُعيد تغريده
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
NemoClaw – NVIDIA’s contribution to the emerging OpenClaw ecosystem and one of the biggest announcements at NVIDIA GTC It's a framework for long-running autonomous agents. ▪️ The idea: Install OpenClaw together with Nemotron models and OpenShell (NVIDIA’s new security runtime) in a single command. NemoClaw gives agents a sandboxed execution environment that: - runs OpenClaw inside a secure container – OpenShell - enforces policies on network, filesystem, and processes - routes all model calls via NVIDIA cloud - provides CLI tools to manage agents In other words, NVIDIA is no longer aiming only to power the model. It wants to sit under the agent itself.
Ksenia_TuringPost tweet media
English
10
22
131
9.1K
Ksenia_TuringPost
Ksenia_TuringPost@TheTuringPost·
8. So V-JEPA 2.1 looks strong across both prediction and dense visual understanding (even with the encoder kept frozen) Some of the results: • +20% robot grasping success over V-JEPA 2 in zero-shot real-world manipulation • 10× faster navigation planning, with 5.687 ATE on Tartan Drive And new SOTA: • 7.71 mAP on Ego4D short-term object interaction anticipation • 40.8 Recall@5 on EPIC-KITCHENS action anticipation
Ksenia_TuringPost tweet mediaKsenia_TuringPost tweet mediaKsenia_TuringPost tweet media
English
1
0
1
576