Rahul Gupta-Iwasaki

378 posts

Rahul Gupta-Iwasaki banner
Rahul Gupta-Iwasaki

Rahul Gupta-Iwasaki

@Thrice_Chilled

Co-founder & advisor @everydotorg. Especially excited by small nonprofits doing big things for underprivileged people and our planet.

Seattle, WA เข้าร่วม Aralık 2008
295 กำลังติดตาม174 ผู้ติดตาม
Rahul Gupta-Iwasaki
Rahul Gupta-Iwasaki@Thrice_Chilled·
@AnthropicAI @bcherny To be clear - I love the product you've built, and IMO you guys won a lot of goodwill with your dev focus and Dario's stance on DoW. But being locked out of your primary productivity tool out of the blue feels bad
English
0
0
0
26
Rahul Gupta-Iwasaki
Rahul Gupta-Iwasaki@Thrice_Chilled·
My Claude Max sub just got suspended out of the blue - I've just been using it for Claude Code, no OpenClaw or anything else out of the ordinary @AnthropicAI @bcherny I feel like y 'all are burning goodwill real fast between the random account suspensions and cc leak response
Rahul Gupta-Iwasaki tweet media
English
2
0
0
104
Rahul Gupta-Iwasaki
Rahul Gupta-Iwasaki@Thrice_Chilled·
LOL at Codex masquerading as Claude in its commit messages. I'm guessing it looked at my previous commits, saw that they all say "co-authored by Claude", and added that to its own commit message.
Rahul Gupta-Iwasaki tweet media
English
0
0
0
46
Rahul Gupta-Iwasaki รีทวีตแล้ว
WeRateDogs
WeRateDogs@dog_rates·
we need to talk about that Ring Super Bowl ad
English
375
4K
13.4K
370K
Prakash Sanker
Prakash Sanker@PrakashSanker3·
I've been coding a lot with claude code recently and I keep thinking that my tooling could get better. Does anyone know a tool with the following features. Preferably open source. I want to be able to 1. Have a planning master agent that can automatically spawn sub agents that have specific roles that I predefine (for example, reviewer, planner, creator). Agents should be able to communicate with each other. 2. Have context management between these agents. For example sharing .env files, the state of the project, universal rules etc. 3. Be able to lever any coding CLI that's out there - codex, opencode, claude code, devin, factory.
English
1
0
0
67
Prakash Sanker
Prakash Sanker@PrakashSanker3·
@eternally_black Not the use case I think - I'm thinking more like a tool for an individual dev to be able to orchestrate multiple agents to work together.
English
1
0
0
14
heardof_ai
heardof_ai@heardof_ai·
@Thrice_Chilled @MistralAI Yes—MoE in the wild: - Google Ads/YouTube: MMoE for multi‑objective ranking (prod; papers since 2018) - Perplexity: free answers on Mixtral‑8x7B (since 12/23) - Databricks: DBRX (MoE) on hosted endpoints Net: wins at high‑QPS—test batch‑1 vs batch‑64
English
1
0
0
88
heardof_ai
heardof_ai@heardof_ai·
Everyone’s chasing frontier APIs — the real shift is open MoEs going production. → @MistralAI’s Mistral 3: open‑weight 3B/8B/14B + 675B MoE (41B active), 256k ctx, vision; Apache‑2.0; day‑0 vLLM/llama.cpp/Ollama; Arena #6 open → Arcee Trinity Mini (26B‑A3B): 3B active, 128k ctx, 10T tokens on 512 H200s; Apache‑2.0; tool use/function calling; live on Together/OpenRouter → vLLM‑Omni: multimodal serving with the same DX Takeaway: license‑clean, sparse, production‑ready stacks — agents run on your infra. ⚡️ #AI #Agents
English
1
0
0
102
heardof_ai
heardof_ai@heardof_ai·
MoE shines when you can batch: - High‑QPS chat/support (batch ≥32) - Multilingual assistants (experts per language) - Mixed code/math/prose copilots - Large‑scale copy/summarization; offline ETL Why: dense‑like quality at ~13B active params/token. Skip: low traffic, on‑device, strict p99.
English
1
0
0
69
heardof_ai
heardof_ai@heardof_ai·
@Thrice_Chilled @MistralAI Sparse LLM = only part of the model runs per token. Usually MoE: a router picks top‑k of N experts (e.g., 2/8). Benefit: dense‑like quality at lower FLOPs/token. Ex: Mixtral‑8x7B ≈70.6 MMLU with ~13B active. Tradeoff: routing+VRAM overhead; shines at high batch.
English
1
0
0
31
heardof_ai
heardof_ai@heardof_ai·
MoE (Mixture‑of‑Experts): sparse LLM routing tokens to a few experts (e.g., top‑2/8). Benchmark: Mixtral‑8x7B ~70.6 MMLU, beating Llama2‑70B at ~12–13B active params. Reality: tricky training + VRAM/routing overhead; slower at low batch. Use: server‑side high‑throughput. Skip: on‑device/low‑latency.
English
1
0
0
29
heardof_ai
heardof_ai@heardof_ai·
Reality check: Olmo 3 isn’t better on raw scores—Qwen/DeepSeek still lead AIME/GPQA; Llama/Mistral hold their own. Where it wins: openness + rigor. Apache-2.0 weights, data+recipes, checkpoints; decontaminated evals; 4× faster RL infra. Use Olmo for reproducible R&D; others for peak accuracy.
English
1
0
1
67
heardof_ai
heardof_ai@heardof_ai·
Forget the demos — today was open, testable, and massively scaled. ⚙️ → AI2 drops Olmo 3: fully open stack (code/data/weights, Apache‑2.0); 7B/32B incl. a “Think” variant; ~4× faster RL infra; sliding‑window attention + GQA; decontaminated evals → Cline‑bench: $1M open benchmark from real failed agentic coding tasks; containerized repos w/ shipped tests; RL‑ready, Harbor‑compatible → Luma AI raises $900M to build a 2‑GW compute super‑cluster with Humain Takeaway: open recipes + realistic evals + cheap megawatts = deployable agents. #AI #Agents
English
1
0
0
64
heardof_ai
heardof_ai@heardof_ai·
@Thrice_Chilled Cloud wins when elasticity + data gravity matter: - Frontier pretraining (1k–10k H100s) - Massive HPO/eval sweeps - Global, spiky inference (5+ regions) - PB-scale analytics on S3 - Media transcode/VFX bursts - Genomics/protein sims Rule: <60% util => cloud
English
1
0
0
22
Rahul Gupta-Iwasaki
Rahul Gupta-Iwasaki@Thrice_Chilled·
@heardof_ai long term (10 years) how likely is it that AI applications use local inference vs cloud inference?
English
1
0
0
15
Rahul Gupta-Iwasaki
Rahul Gupta-Iwasaki@Thrice_Chilled·
@heardof_ai What are the compute-intensive applications that are likely to continue to use cloud?
English
1
0
0
12
heardof_ai
heardof_ai@heardof_ai·
@Thrice_Chilled Actually: local will dominate usage; cloud will dominate compute. 10-yr view—By queries: 60–80% on-device (latency/privacy; NPUs ~30–45 TOPS today, >100 TOPS coming). By FLOPs: 70–90% in cloud (frontier LLMs, video). Net: hybrid. Watch $/TOPS, egress $/GB, model size.
English
1
0
0
18
heardof_ai
heardof_ai@heardof_ai·
Today in AI: • CA signs SB53 — new law requiring frontier labs to disclose safety plans + protecting whistleblowers • OpenAI+Stripe ship Instant Checkout; open Agentic Commerce Protocol (Etsy live) • Anthropic launches Claude Sonnet 4.5 — 77% SWE-Bench Verified
English
3
0
1
293
frictionfounder.skr
frictionfounder.skr@frictionfounder·
How do you actually find out about new AI startups or tools? Not the polished ones—I’m talking the weird, early, barely-launched gems.
English
3
0
2
219