Alessandro

2.7K posts

Alessandro

Alessandro

@alew3

Tech & Sports Enthusiast // Chief AI Officer @ https://t.co/Y5xWTG5Ka3

Palo Alto, CA Katılım Aralık 2007
1.3K Takip Edilen760 Takipçiler
Felix Rieseberg
Felix Rieseberg@felixrieseberg·
We're shipping a new feature in Claude Cowork as a research preview that I'm excited about: Dispatch! One persistent conversation with Claude that runs on your computer. Message it from your phone. Come back to finished work. To try it out, download Claude Desktop, then pair your phone.
English
973
1.5K
17.4K
6.2M
Alessandro
Alessandro@alew3·
@felixrieseberg I have Max 20 , how do I get this working? Already update the desktop and mobile.
English
2
0
0
1K
Felix Rieseberg
Felix Rieseberg@felixrieseberg·
Your desktop has to be running. Like Cowork itself, we’re shipping an early version - you can expect more to come here within the next few days and weeks. Rolling out now to Max subscribers, with Pro coming in the next few days. Try it and let me know what you think. Download the mobile app and pair it with your desktop app: claude.com/download
English
72
18
896
226.6K
Alessandro retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)
Andrej Karpathy tweet media
English
1.1K
3.7K
28.3K
10.9M
Thariq
Thariq@trq212·
I did a 2 hour workshop on the Claude Agent SDK at AI engineer! We are still so early to agents, I hope this is useful if you’re thinking of making one.
AI Engineer@aiDotEngineer

🆕 Claude Agent SDK [Full Workshop] youtube.com/watch?v=TqC1qO… For our first big drop of the year, excited to bring you @trq212's full 2 hour workshop covering all of @AnthropicAI's agentic SDK (formerly known as Claude Code SDK). By far the most popular workshop of AIE CODE! Now published online for free (sorry for AV/delay issues)... long story

English
69
135
1.5K
209.8K
Awni Hannun
Awni Hannun@awnihannun·
mlx-lm is becoming quite a powerful little inference framework! The latest release adds tensor-parallel LLM inference for use with the new low-latency JACCL back-end in MLX (h/t @angeloskath). Also updated to support Transformers V5!
Awni Hannun tweet media
English
11
13
145
7.4K
Alessandro retweetledi
Rachel Thomas
Rachel Thomas@math_rachel·
"People who go all in on AI agents now are guaranteeing their obsolescence. If you outsource all your thinking to computers, you stop upskilling, learning, and becoming more competent. AI is great at helping you learn." @jeremyphoward @NVIDIAAI youtube.com/watch?v=zDkHJD… 2/
YouTube video
YouTube
English
3
10
66
18.5K
Alessandro
Alessandro@alew3·
@awnihannun Looking good! BTW, any reason MLX can't leverage the Neural Engine?
English
0
0
0
783
Awni Hannun
Awni Hannun@awnihannun·
Running Ring 1T reasoning model on a single M3 Ultra with mlx-lm. It's quantized to 3.5 bits-per-weight. Uses 440GB and generated ~6k tokens at 18.2 toks/sec. Getting closer to GPT-5 at home.
English
16
50
511
55.1K
Awni Hannun
Awni Hannun@awnihannun·
The new batch generation in MLX LM is pretty fast. Here's 4 simultaneous generations with Qwen3 4B on my M4 max:
English
20
22
254
60.2K
Alessandro retweetledi
OpenAI
OpenAI@OpenAI·
ChatGPT agent is ready to introduce itself. openai.com/live
English
382
544
4.1K
1.5M
Alessandro retweetledi
Charlie Marsh
Charlie Marsh@charliermarsh·
You can set `UV_TORCH_BACKEND=auto` and uv will automatically install the right CUDA-enabled PyTorch for your machine, zero configuration
Charlie Marsh tweet media
English
73
224
2.5K
191K
Prince Canuma
Prince Canuma@Prince_Canuma·
I’ve left @arcee_ai! I really love what we have achieved together across research and product. From model fusion, offline distillation of Llama 405B all the way to building, leading and launching Arcee Orchestra from scratch within 4 months. Already miss everyone, more than colleagues they are my friends ❤️ Nevertheless, I’m very excited to soon announce what’s next. Meanwhile, I’m happy to share that I’ll be working full-time on MLX (mlx, mlx-lm, mlx-vlm, mlx-audio and more) to help build the best on-device R&D experience and products by bringing the latest OS models and features to Apple Silicon.
English
37
6
253
16.7K
Scott
Scott@scottstts·
@maximelabonne There are rarely apps that have more than 700M users, other than that the rest are not hard constraints
English
2
0
15
3.7K
Maxime Labonne
Maxime Labonne@maximelabonne·
Llama 4's new license comes with several limitations: - Companies with more than 700 million monthly active users must request a special license from Meta, which Meta can grant or deny at its sole discretion. - You must prominently display "Built with Llama" on websites, interfaces, documentation, etc. - Any AI model you create using Llama Materials must include "Llama" at the beginning of its name - You must include the specific attribution notice in a "Notice" text file with any distribution - Your use must comply with Meta's separate Acceptable Use Policy (referenced at llama.com/llama4/use-pol…) - Limited license to use "Llama" name only for compliance with the branding requirements
Maxime Labonne tweet media
English
78
97
749
203.4K
Alessandro retweetledi
OpenRouter
OpenRouter@OpenRouter·
Excited to announce our first-ever “stealth” model... Quasar Alpha 🥷 It’s a prerelease of an upcoming long-context foundation model from one of the model labs: - 1M token context length - specifically optimized for coding, but general-purpose as well - available for free
OpenRouter tweet media
OpenRouter@OpenRouter

A stealth model has entered the chat... 🥷

English
72
104
1.2K
374.3K
New
New@newsystems_·
It's finally here: Brampton Brampton is the world's most intelligent, creative, and fastest model. Brampton dramatically outperforms Grok 3, Claude 3.7 Sonnet, and GPT 4.5. Reply with "brampton" for early access.
English
4K
216
3.8K
2M