Greg Lin Tanaka

806 posts

Greg Lin Tanaka banner
Greg Lin Tanaka

Greg Lin Tanaka

@GregTanaka

@Prompt_Driven, serial founder (@a16z, @GVteam, @MenloVentures), former @CityofPaloAlto Councilman, e/acc #AI @Caltech @UCBerkeley

Palo Alto, CA Katılım Ağustos 2009
763 Takip Edilen7.1K Takipçiler
Sean Grove
Sean Grove@sgrove·
Does the world need a new programming language? Quite possibly!
English
4
0
4
714
Greg Lin Tanaka retweetledi
Prompt Driven
Prompt Driven@Prompt_Driven·
Ever get that "grenade in the codebase" feeling from agentic coders like Claude Code? You're never sure what they'll add, delete, or duplicate. I started exploring a new approach: what if prompts themselves were the source of truth instead of merely being used to patch the code?
English
1
1
1
609
Ivan Fioravanti ᯅ
Ivan Fioravanti ᯅ@ivanfioravanti·
Another MLX Boom! 💥 Qwen3-235B-A22B-Thinking-2507-3bit-DWQ! Group Size 32 to keep quality high enough! arc_easy test: 80.3 (testing 3bit, 4bit, 8bit to compare!) M3 Ultra performance: Prompt: 20.2 tokens-per-sec Generation: 31.6 tokens-per-sec Peak memory: 103.010 GB Ready for our Mac with 128GB+!
Ivan Fioravanti ᯅ tweet media
English
8
12
151
11.9K
Greg Lin Tanaka retweetledi
Qwen
Qwen@Alibaba_Qwen·
>>> Qwen3-Coder is here! ✅ We’re releasing Qwen3-Coder-480B-A35B-Instruct, our most powerful open agentic code model to date. This 480B-parameter Mixture-of-Experts model (35B active) natively supports 256K context and scales to 1M context with extrapolation. It achieves top-tier performance across multiple agentic coding benchmarks among open models, including SWE-bench-Verified!!! 🚀 Alongside the model, we're also open-sourcing a command-line tool for agentic coding: Qwen Code. Forked from Gemini Code, it includes custom prompts and function call protocols to fully unlock Qwen3-Coder’s capabilities. Qwen3-Coder works seamlessly with the community’s best developer tools. As a foundation model, we hope it can be used anywhere across the digital world — Agentic Coding in the World! 💬 Chat: chat.qwen.ai 📚 Blog: qwenlm.github.io/blog/qwen3-cod… 🤗 Model: hf.co/Qwen/Qwen3-Cod… 🤖 Qwen Code: github.com/QwenLM/qwen-co…
Qwen tweet media
English
382
1.5K
9.4K
2.3M
Greg Lin Tanaka retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
I attended a vibe coding hackathon recently and used the chance to build a web app (with auth, payments, deploy, etc.). I tinker but I am not a web dev by background, so besides the app, I was very interested in what it's like to vibe code a full web app today. As such, I wrote none of the code directly (Cursor+Claude/o3 did) and I don't really know how the app works, in the conventional sense that I'm used to as an engineer. The app is called MenuGen, and it is live on menugen.app. Basically I'm often confused about what all the things on a restaurant menu are - e.g. Pâté, Tagine, Cavatappi or Sweetbread (hint it's... not sweet). Enter MenuGen: you take a picture of a menu and it generates images for all the menu items and presents them in a nice list. I find it super useful to get a quick visual sense of the menu. But the more interesting part for me I thought was the exploration of vibe coding around how easy/hard it is to build and deploy a full web app today if you are not a web developer. So I wrote up the full blog post on my experience here, including some takeaways: karpathy.bearblog.dev/vibe-coding-me… Copy pasting just the TLDR: "Vibe coding menugen was exhilarating and fun escapade as a local demo, but a bit of a painful slog as a deployed, real app. Building a modern app is a bit like assembling IKEA future. There are all these services, docs, API keys, configurations, dev/prod deployments, team and security features, rate limits, pricing tiers... Meanwhile the LLMs have slightly outdated knowledge of everything, they make subtle but critical design mistakes when you watch them closely, and sometimes they hallucinate or gaslight you about solutions. But the most interesting part to me was that I didn't even spend all that much work in the code editor itself. I spent most of it in the browser, moving between tabs and settings and configuring and gluing a monster. All of this work and state is not even accessible or manipulatable by an LLM - how are we supposed to be automating society by 2027 like this?" See the post for full detail, and maybe give MenuGen a go the next time you're at a restaurant!
Andrej Karpathy tweet media
English
435
646
7.6K
780K
Greg Lin Tanaka retweetledi
AI at Meta
AI at Meta@AIatMeta·
Today is the start of a new era of natively multimodal AI innovation. Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick — our most advanced models yet and the best in their class for multimodality. Llama 4 Scout • 17B-active-parameter model with 16 experts. • Industry-leading context window of 10M tokens. • Outperforms Gemma 3, Gemini 2.0 Flash-Lite and Mistral 3.1 across a broad range of widely accepted benchmarks. Llama 4 Maverick • 17B-active-parameter model with 128 experts. • Best-in-class image grounding with the ability to align user prompts with relevant visual concepts and anchor model responses to regions in the image. • Outperforms GPT-4o and Gemini 2.0 Flash across a broad range of widely accepted benchmarks. • Achieves comparable results to DeepSeek v3 on reasoning and coding — at half the active parameters. • Unparalleled performance-to-cost ratio with a chat version scoring ELO of 1417 on LMArena. These models are our best yet thanks to distillation from Llama 4 Behemoth, our most powerful model yet. Llama 4 Behemoth is still in training and is currently seeing results that outperform GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM-focused benchmarks. We’re excited to share more details about it even while it’s still in flight. Read more about the first Llama 4 models, including training and benchmarks ➡️ go.fb.me/gmjohs Download Llama 4 ➡️ go.fb.me/bwwhe9
AI at Meta tweet media
English
831
2.4K
12.8K
3.7M
Tom Dörr
Tom Dörr@tom_doerr·
Bought a 55-inch OLED TV as a monitor - best purchase I've made in a long time
Tom Dörr tweet media
English
36
3
257
36.4K
Novita AI
Novita AI@novita_labs·
Our powerful DeepSeek R1 Turbo is live on @HuggingFace! 🤗 It's improved in every way: 64K context + 16K max output + 30 throughput + 99.9% stability + 💰 CHEAPER THAN EVER Try it out on Hugging Face's model page 🫶 @julien_c
English
3
14
97
23.6K
Greg Lin Tanaka retweetledi
Qwen
Qwen@Alibaba_Qwen·
Today, we release QwQ-32B, our new reasoning model with only 32 billion parameters that rivals cutting-edge reasoning model, e.g., DeepSeek-R1. Blog: qwenlm.github.io/blog/qwq-32b HF: huggingface.co/Qwen/QwQ-32B ModelScope: modelscope.cn/models/Qwen/Qw… Demo: huggingface.co/spaces/Qwen/Qw… Qwen Chat: chat.qwen.ai This time, we investigate recipes for scaling RL and have achieved some impressive results based on our Qwen2.5-32B. We find that RL training con continuously improve the performance especially in math and coding, and we observe that the continous scaling of RL can help a medium-size model achieve competitieve performance against gigantic MoE model. Feel free to chat with our new models and provide us feedback!
Qwen tweet media
English
474
1.5K
8.8K
3.6M
Paul Gauthier
Paul Gauthier@paulgauthier·
GPT-4.5 Preview scored 45% on aider's polyglot coding benchmark. 65% Sonnet 3.7, 32k think tokens (SOTA) 60% Sonnet 3.7, no thinking 48% DeepSeek V3 45% GPT 4.5 Preview 27% ChatGPT-4o 23% GPT-4o aider.chat/docs/leaderboa…
Paul Gauthier tweet media
English
59
43
513
56.3K
NVIDIA AI Developer
NVIDIA AI Developer@NVIDIAAIDev·
Securely experiment and build your own specialized agents, as the 671-billion-parameter DeepSeek-R1 model is now available as an NVIDIA NIM microservice in preview on build.nvidia.com. Learn more ➡️ nvda.ws/4grQaBq
English
111
253
1.1K
642.9K
Greg Lin Tanaka
Greg Lin Tanaka@GregTanaka·
Everyone’s debating how @deepseek_ai R1 trained for <$6M, but it’s simple: same reason @Google Search is ‘free.’ Google sells ads; DeepSeek sells intent. Knowing what the world is thinking fuels their edge in quantitative trading. Data is the new currency. 💡📊 #AI #QuantTrading
English
0
0
4
356