

Ivan Chong (張建生)
4.5K posts

@ichong
Imperfect Christian. Managing Partner at Concitor TCG, specializing in AI Business Transformation











WATCH: Taiwanese grandmothers aged 89 and 91 train at the gym. An increasing number of elderly people in Taiwan’s super-aged society are hitting the gym to stay healthy, both physically and mentally.


MiniMax M2.7 costs money to access. Kimi K2 costs money. GLM-4.7 costs money. DeepSeek V3.2 costs money. NVIDIA is giving you all of them. Right now. For free. No credit card. No trial period. No expiry date. Just a free API key and immediate access to some of the most powerful AI models on the planet. NVIDIA has quietly made its NIM — NVIDIA Inference Microservices — APIs available to the public through build.nvidia.com/models. You receive an actual API key, choose a model, send requests, and pay nothing. And the models are not toys. MiniMax M2.7 is a 230 billion parameter model with a Sparse Mixture-of-Experts architecture — 256 local experts, 8 activated per token — with a 204,800 token context window, excelling in coding, reasoning, and complex office tasks. This is a model companies are paying per token to access through MiniMax's own API. NVIDIA is serving it for nothing. GLM-5.1 is a flagship LLM for agentic workflows, coding, and long-horizon reasoning tasks. GLM-4.7 is a multilingual agentic coding partner with stronger reasoning, tool use, and UI skills. DeepSeek V3.2 — the model that caused a global market panic in January 2025 when it proved Chinese AI could match American labs for a fraction of the cost — is in the catalog. Free. The full list keeps going. GPT-OSS-120B. Sarvam-M. Llama 4 Maverick. Mistral Large. Qwen3-Coder. The full catalogue lives at build.nvidia.com/models and grows regularly. Here is how to set it up in 60 seconds. Grab your API key at build.nvidia.com. Set your base URL to integrate.api.nvidia.com/v1. Set your API key to your NVIDIA key starting with nvapi-. Select your model — for example, minimaxai/minimax-m2.7. That is the entire setup. Because it uses the standard OpenAI SDK format, it plugs directly into every tool you already use. Cursor, Zed, OpenCode, Hermes agent, Claude Code — all of them work without any code changes. Now here is the part nobody is saying out loud. NVIDIA is not doing this out of generosity. The catalog is a top-of-funnel play for NVIDIA AI Enterprise, their paid inference platform. The path is designed to be frictionless: prototype on the free API, test on GPU sandbox instances, then deploy self-hosted NIM containers in your own data center with a paid license. Every developer who builds on NVIDIA's free tier is a developer who learns NVIDIA's API conventions, runs experiments on NVIDIA hardware, and builds deployment pipelines around NVIDIA's infrastructure. When they need to scale to production — they already know which chips to buy. The free tier is not the product. The enterprise contract that follows is. It is the smartest customer acquisition strategy in enterprise technology. Let you try the best hardware in the world for free. Make it trivially easy to integrate. Then sell you the infrastructure when you need to scale. Here are the honest limitations. Developers get 1,000 free inference credits on signup with a rate limit of 40 requests per minute — enough for meaningful prototyping before committing to self-hosted deployment. The larger models eat through credits surprisingly fast. 40 requests per minute is a prototyping budget — it is not enough to run a production application. But for evaluation, development, learning, personal projects, and running production-grade frontier models without needing an H100 cluster in your garage? Since integrate.api.nvidia.com/v1 is OpenAI-compatible, OpenClaw, OpenCode, Zed, and Cursor can call it directly. Swapping in NVIDIA's endpoint is a base URL change and an API key. Nothing more. 100+ models. Real API key. No credit card. No expiry. The X post that went viral asking "why is nobody talking about this?" hit 31,000 reposts in 48 hours. Now you know. build.nvidia.com/models Source: NVIDIA NIM · build.nvidia.com · Medium/Coding Nexus · Lilting.ch · April 2026