BluePi

11.5K posts

BluePi banner
BluePi

BluePi

@BluePi_In

Founder @BluePi | GCP, Vertex AI & AI Agents | Transforming businesses through data engineering & migration | Writing threads on AI, cloud & future tech

Gurgaon, India Bergabung Şubat 2013
1.5K Mengikuti1.5K Pengikut
Tweet Disematkan
BluePi
BluePi@BluePi_In·
The Great Transition Honestly, I don’t think the human brain is wired for what’s coming over the next 36 months. We are about to enter a period of profound cognitive dissonance that will be felt in every corner of the globe. open.substack.com/pub/pronamchat…
BluePi tweet media
English
0
1
1
200
BluePi me-retweet
Furkan Gözükara
Furkan Gözükara@FurkanGozukara·
Absolute bombshell. Data reveals someone made a massive 580 MILLION dollar trade on oil exactly 15 minutes BEFORE Donald Trump posted his tweet about pausing the Iran war. Someone on the inside just made a life changing fortune. The corruption is blatant.
English
2K
28K
105.3K
5.2M
BluePi me-retweet
Tianyu Liu
Tianyu Liu@rogerliuty·
This score of Kimi K2.5 looks a bit off to me — @nextjs @vercel, could someone please reach out? My DMs are open. We’ve been actively collaborating with open-source benchmark communities, submitting PRs to help improve evaluation quality. Especially for software engineering tasks, runtime environments and configs can be quite complex — even the best models or teams may face challenges ensuring fully accurate or stable results. We're eager to keep working with the community to strengthen evaluation systems — better benchmarks lead to better models.
Elie Steinbock — oss/acc@elie2222

Vercel Next.js benchmark 👇 Kimi K2.5 scores 19% Compose 2 scores 76% Post-training matters

English
10
9
222
36.4K
BluePi me-retweet
Jay
Jay@jayair·
Yeah it’s called composer cos it’s composed of other models
English
50
73
2.4K
63.6K
BluePi me-retweet
Zixuan Li
Zixuan Li@ZixuanLi_·
Don't panic. GLM-5.1 will be open source.
English
276
425
7.7K
989.9K
BluePi me-retweet
BridgeMind
BridgeMind@bridgemindai·
Composer 2 is just Kimi K2.5 with reinforcement learning. Someone sniffed the API calls. The model ID is "kimi-k2p5-rl-0317-s515-fast" hosted under Anysphere's account. Cursor isn't training their own model from scratch. They're fine-tuning Kimi K2.5 with RL and calling it Composer 2. That blog post said "our first continued pretraining run." It's continued pretraining on someone else's model. Now the hallucination problems make a lot more sense.
BridgeMind tweet media
English
26
6
190
14.4K
BluePi me-retweet
Aakash Gupta
Aakash Gupta@aakashgupta·
Cursor is raising at a $50 billion valuation on the claim that its “in-house models generate more code than almost any other LLMs in the world.” Less than 24 hours after launching Composer 2, a developer found the model ID in the API response: kimi-k2p5-rl-0317-s515-fast. That’s Moonshot AI’s Kimi K2.5 with reinforcement learning appended. A developer named Fynn was testing Cursor’s OpenAI-compatible base URL when the identifier leaked through the response headers. Moonshot’s head of pretraining, Yulun Du, confirmed on X that the tokenizer is identical to Kimi’s and questioned Cursor’s license compliance. Two other Moonshot employees posted confirmations. All three posts have since been deleted. This is the second time. When Cursor launched Composer 1 in October 2025, users across multiple countries reported the model spontaneously switching its inner monologue to Chinese mid-session. Kenneth Auchenberg, a partner at Alley Corp, posted a screenshot calling it a smoking gun. KR-Asia and 36Kr confirmed both Cursor and Windsurf were running fine-tuned Chinese open-weight models underneath. Cursor never disclosed what Composer 1 was built on. They shipped Composer 1.5 in February and moved on. The pattern: take a Chinese open-weight model, run RL on coding tasks, ship it as a proprietary breakthrough, publish a cost-performance chart comparing yourself against Opus 4.6 and GPT-5.4 without disclosing that your base model was free, then raise another round. That chart from the Composer 2 announcement deserves its own paragraph. Cursor plotted Composer 2 against frontier models on a price-vs-quality axis to argue they’d hit a superior tradeoff. What the chart doesn’t show is that Anthropic and OpenAI trained their models from scratch. Cursor took an open-weight model that Moonshot spent hundreds of millions developing, ran RL on top, and presented the output as evidence of in-house research. That’s margin arbitrage on someone else’s R&D dressed up as a benchmark slide. The license makes this more than an attribution oversight. Kimi K2.5 ships under a Modified MIT License with one clause designed for exactly this scenario: if your product exceeds $20 million in monthly revenue, you must prominently display “Kimi K2.5” on the user interface. Cursor’s ARR crossed $2 billion in February. That’s roughly $167 million per month, 8x the threshold. The clause covers derivative works explicitly. Cursor is valued at $29.3 billion and raising at $50 billion. Moonshot’s last reported valuation was $4.3 billion. The company worth 12x more took the smaller company’s model and shipped it as proprietary technology to justify a valuation built on the frontier lab narrative. Three Composer releases in five months. Composer 1 caught speaking Chinese. Composer 2 caught with a Kimi model ID in the API. A P0 incident this year. And a benchmark chart that compares an RL fine-tune against models requiring billions in training compute without disclosing the base was free. The question for investors in the $50 billion round: what exactly are you buying? A VS Code fork with strong distribution, or a frontier research lab? The model ID in the API answers that. If Moonshot doesn’t enforce this license against a company generating $2 billion annually from a derivative of their model, the attribution clause becomes decoration for every future open-weight release. Every AI lab watching this is running the same math: why open-source your model if companies with better distribution can strip attribution, call it proprietary, and raise at 12x your valuation? kimi-k2p5-rl-0317-s515-fast is the most expensive model ID leak in the history of AI licensing.
Harveen Singh Chadha@HarveenChadha

things are about to get interesting from here on

English
247
549
4.4K
1.4M
BluePi me-retweet
Ajeet ( opensox.ai )
Ajeet ( opensox.ai )@ajeetprssingh·
imagine being a $380B company and nuking a small, community-built Open Source project just because it's better than yours and costs less.
Ajeet ( opensox.ai ) tweet media
English
65
79
3.3K
330.1K
BluePi me-retweet
Hugging Models
Hugging Models@HuggingModels·
Meet Sarvam-30B: a massive 30-billion parameter language model built for multilingual conversations. It's a specialized, custom-coded model designed to understand and create text in English, Hindi, Bengali, and Tamil. A true polyglot AI.
Hugging Models tweet media
English
12
206
1.5K
25.8K
BluePi me-retweet
Chubby♨️
Chubby♨️@kimmonismus·
Indian model Sarvam-105b is really really good Sarvam AI has open-sourced two India-built reasoning models, Sarvam 30B and 105B, positioning them as globally competitive open models. The big unlock is not just benchmark scores like 98.6 on Math500 for 105B or strong local deployment efficiency for 30B, but the full-stack story: in-house data, training, RL, tokenizer design, and inference optimization built for both frontier GPUs and consumer devices.
Chubby♨️ tweet media
English
43
247
2.1K
111.3K
BluePi me-retweet
Adam Mockler
Adam Mockler@adammocklerr·
Did Trump just trigger the most self defeating chain of events in modern history? > Israel drags US into war with Iran > Oil surges in price, troops die > Trump panics and lifts Russia sanctions so India can buy more oil > Russia uses that money to help Iran > We just funded our own enemy
English
693
6.2K
35K
1.2M
BluePi
BluePi@BluePi_In·
This is directionally right, but “AI taking over broking” is still overstated. What’s really happening is interface collapse. Search, portfolio analysis, disclosures, and first-pass advice are moving into a conversational layer. Execution, risk, compliance, and liability are still very much human + system driven. Most of what we’re seeing today—MCPs, assistants, digests—is about reducing friction, not outsourcing judgment. The hard problems aren’t LLMs: suitability and mis-selling explainability under scrutiny accountability when advice goes wrong SEBI’s stance makes this clear: AI is allowed, but responsibility doesn’t move. The broker still owns outcomes. The real disruption won’t come from better chat. It’ll come when AI systems are trusted to act under constraints—capital limits, audit trails, reversibility—without breaking regulatory trust. That’s still ahead of us.
English
1
0
2
2.5K
Rahul Mathur
Rahul Mathur@Rahul_J_Mathur·
AI is slowly taking over retail stock broking in India: The launch of Perplexity Finance for India in Aug ‘25 took incumbents & startups by surprise - it was supposed to be simple banter between Arvind & Nikhil Kamath of Zerodha - but a week later the real product came out Here’s how the industry has adopted AI ⤵️
English
21
82
694
168.9K
BluePi
BluePi@BluePi_In·
7/ Models matter. Harnesses decide who actually ships. Curious where others disagree — especially people running multi-file changes, not demos.
English
1
0
0
16
BluePi
BluePi@BluePi_In·
3/ Important clarification: Claude Opus 4.5 and GLM-4.7 are not the same class. Opus is the stronger model, full stop. And yet—
English
4
0
0
43
BluePi
BluePi@BluePi_In·
1/ After shipping real code with these tools, one thing is clear to me: For coding agents, the harness beats the model. A weak loop will waste even top-tier intelligence.
English
1
0
0
23
BluePi
BluePi@BluePi_In·
6/ UI is a separate axis. For interaction quality and long sessions, Gemini models still win for me. They’re just easier to live inside. (@GoogleDeepMind @geminicli)
English
0
0
0
16
BluePi
BluePi@BluePi_In·
5/ Right now, the setup that compounds fastest for me: Codex CLI + Codex 5.2 Xtra High. Planning, execution, rollback, and retry stay coherent across long sessions. That’s rare. (@OpenAI)
English
0
0
0
37
BluePi
BluePi@BluePi_In·
4/ In practice, I still get better outcomes from: Claude Code + GLM-4.7 over Opus 4.5 + weaker harnesses. That gap is entirely explained by tooling. (@AnthropicAI @zhipu_ai)
English
0
0
0
37