Stephen Fernandes
1.4K posts

Stephen Fernandes
@stephennfern
building LLMs for Indian Languages




🌟 Join the TorchTPU session at #PyTorchCon Europe 2026 in Paris on April 7–8. Learn the strategy and roadmap that brings a native @PyTorch experience to TPUs. The program includes infrastructure, applications, and agent-based systems, with sessions on training, inference, gen AI, responsible AI, security, privacy, and frameworks.








This MacBook, bought 10 years ago, has been working like new even today. 👀 Tell me which Windows laptop can last this long, and I will wait. 🫠


RTX 4090 from $0.14/hr. Spot instances for batch inference, evals, fine-tuning. Pay per second.

Within 1 year, I think Google will train a huge model on 2 million TPU v7 Ironwood chips running for 6 months, producing close to 10 ZettaFLOPS at peak and 384 Petabytes of HBM. 7.6 × 10²⁸ total FLOPS — 3,800× GPT-4’s training compute. The net result would be a 54-Trillion parameter MoE with 5 Trillion active per token, 32-million token context window, trained on 500 Trillion multimodal tokens with 55% of all compute going to RL reasoning training.



I never learned to code. First it was too hard, then it was too easy.










