Truth

30 posts

Truth

Truth

@NumCupboard

World Katılım Eylül 2024
16 Takip Edilen11 Takipçiler
Truth
Truth@NumCupboard·
@varunram You can't compare the revenue of a business with ~0 token inference costs with one that has massive token inference costs.
English
0
0
0
51
Varunram Ganesh
Varunram Ganesh@varunram·
Wait, Vercel has lower "ARR" than Lovable? Is this real?
Varunram Ganesh tweet media
English
25
2
223
105.1K
Truth
Truth@NumCupboard·
@thdxr ... and then Qwen 4 Coder comes out, and paired with OpenCode it obliterates all of them.
English
1
0
0
153
dax
dax@thdxr·
they said cursors data flywheel would make them unstoppable but then claude code came out they said claude codes data flywheel would make them unstoppable but then codex came out they said codex's data flywheel would make them unstoppable then composer 2 came out
English
56
8
703
56.2K
Truth
Truth@NumCupboard·
@sethbannon This is not correct. They simply play the spray and pray game. It is NOT about "giving as many founders as possible a launching pad". No idea who ingrained this idealistic fallacy into you, but it is what it is: A fallacy.
English
0
0
0
8
Seth Bannon
Seth Bannon@sethbannon·
It's not fair to blame YC for backing apparent frauds like Delve. They only spend 10 minutes with founders over zoom before investing. It's a cursory screen. They're not trying to deeply vet founders -- they're trying to give as many founders as possible a launching pad.
English
55
26
383
34K
Truth
Truth@NumCupboard·
@ZixuanLi_ But I appreciate that you mentioned "luck". So it appears that you recognize that any positive returns would converge to negative returns if one lets the law of large numbers play out reality.
English
0
0
0
56
Truth
Truth@NumCupboard·
@ZixuanLi_ This isn't how financial markets work. It's the ever-returning fallacy of comparing publicly accessible financial markets with other markets. Financial markets are the most competitive markets by far. Any edge that a public AI model gives becomes worthless almost instantly.
English
1
0
0
898
Zixuan Li
Zixuan Li@ZixuanLi_·
Whether driven by luck or analytical capabilities, GLM-5 is currently the only model outperforming the Human Baseline on predictionarena.ai. Anyone using GLM-5 for trading? Does it feel capable to you?
Zixuan Li tweet media
English
39
53
641
52K
Truth
Truth@NumCupboard·
@toddsaunders I think the YC model is obsolete. Not - will be. It is already. LLMs changed the whole game.
English
0
0
0
19
Todd Saunders
Todd Saunders@toddsaunders·
YC’s model works for a specific founder. Technical, Bay Area adjacent founders who want to play a zero sum game. But there’s a massive founder group being ignored. The operator who built a $5M-$50M services business and knows exactly what software their industry needs. They know the pain and workflows better than anyone else. And that founder has an industry following / friends to sell the product to. Claude Code just gave them all the power. And I’m going to surround them with the best people I know to help turn their domain expertise into a giant software company.
English
60
15
368
29.1K
Truth
Truth@NumCupboard·
@paulg Incorrect non-sense. The hidden system prompt always contains the instructions to use markdown formatting along with examples. That's why.
English
0
0
0
27
Paul Graham
Paul Graham@paulg·
The fact that AIs tend to answer you in bulleted lists tells us something important, though somewhat depressing: people can't read. They don't do this by accident. What you're seeing is an implicit portrait of the median user.
English
466
141
2.6K
217.1K
Marcel Pociot 🧪
Marcel Pociot 🧪@marcelpociot·
After using Composer 2 pretty much all day, I gotta say... this is an amazing model! ❤️
English
17
3
217
28.6K
Truth
Truth@NumCupboard·
@ctjlewis Stanford's non-PhD programs have been far from research-level for quite some time now. They shifted towards practically oriented skill learning. Most CS undergraduates don't even do any proof heavy math anymore, and theoretical CS is also cut short severely. It's not elite.
English
1
0
32
4.4K
Lewis 🇺🇸
Lewis 🇺🇸@ctjlewis·
It feels almost like YC has hijacked incentives to the point where other institutions are being damaged. We expect Stanford to produce brilliant researchers, not box-checking faggots who are all clones of each other, doing “prompt to sales leads” bullshit.
Lewis 🇺🇸 tweet media
English
26
26
980
55.1K
Truth
Truth@NumCupboard·
@TeksEdge @novita_labs @OpenRouter Yeah. Novita may be ZDR indeed. However does Novita confirm that all their models are hosted by themselves? Because obviously they cannot control logging of third party inference providers. That said, it may well be that Novita has a non-public commercial agreement with MiniMax.
English
1
0
0
126
David Hendrickson
David Hendrickson@TeksEdge·
📉 Bad News: MiniMax M2.7 is closed weights. No HuggingFace release = no local inferencing. 🛡️ Good News, you can still use it with 100% privacy. Use @novita_labs (US SF-based) via @OpenRouter. They operate under a Zero Retention agreement: 🚫 No logging inputs 🚫 No storing chat history 🚫 No training on your prompts 📜 “You acknowledge and agree that we shall not log, store, or retain any User Content, including but not limited to your Inputs, prompts, chat histories, or any Outputs generated by the Services, unless we have obtained your prior written consent for specific instances.” If we can't get open weights, zero retention is better than nothing. 🤝
David Hendrickson tweet media
English
29
17
218
38.2K
Amir Farahani
Amir Farahani@TheAmirFarahani·
@bennpeifert Crazy how y'all are obessed w bro. What about the 1000+ other successes from YC?
English
6
0
6
5K
Benn Eifert 🥷🏴‍☠️
new rule: jail time for Garry Tan every time a YC startup is convicted of fraud
English
57
152
3.5K
158.5K
Truth
Truth@NumCupboard·
@jessfraz Kimi appears to be VERY generous here that they permit this Composer 2 is not available in the publicly accessible Fireworks AI Model Library. By my understanding, this would mean that Cursor uses Fireworks as a compute provider and so Cursor would have to expose it themselves.
English
0
0
0
105
Jessie Frazelle
Jessie Frazelle@jessfraz·
This is nuts to me, the one thing Moonshot (kimi creators) asks you do is say that they are the base. Like just say it, what's the big deal, everyone already knows! It's insane. At this point not saying it makes you look like the baddies.
English
6
1
62
10.1K
Truth
Truth@NumCupboard·
@leerob They are VERY generous here. But congratulations. Based purely on the license wording, legally speaking, they could probably demand much more out of this, since you are primarily using Fireworks as a compute provider for a closed weight model derived from Kimi K2.5.
English
0
0
8
1.4K
Lee Robinson
Lee Robinson@leerob·
Here's confirmation the license is correct from the Kimi team. Agree with the feedback we should have mentioned the base up front, we will do that for the next model! x.com/Kimi_Moonshot/…
Kimi.ai@Kimi_Moonshot

Congrats to the @cursor_ai team on the launch of Composer 2! We are proud to see Kimi-k2.5 provide the foundation. Seeing our model integrated effectively through Cursor's continued pretraining & high-compute RL training is the open model ecosystem we love to support. Note: Cursor accesses Kimi-k2.5 via Fireworks' hosted RL and inference platform as part of an authorized commercial partnership.

English
23
14
429
96.3K
Lee Robinson
Lee Robinson@leerob·
Yep, Composer 2 started from an open-source base! We will do full pretraining in the future. Only ~1/4 of the compute spent on the final model came from the base, the rest is from our training. This is why evals are very different. And yes, we are following the license through our inference partner terms.
Fynn@fynnso

was messing with the OpenAI base URL in Cursor and caught this accounts/anysphere/models/kimi-k2p5-rl-0317-s515-fast so composer 2 is just Kimi K2.5 with RL at least rename the model ID

English
358
197
2.8K
1.4M
Truth
Truth@NumCupboard·
@dzhulgakov @FireworksAI_HQ @leerob Thanks for clarifying. Kimi appears to be VERY generous here tbh. Legally speaking they could demand much more out of this based on the license wording most likely.
English
0
0
0
56
Dmytro Dzhulgakov
Dmytro Dzhulgakov@dzhulgakov·
@NumCupboard @FireworksAI_HQ @leerob We've been working with Kimi and Cursor from the beginning, it's all good x.com/lqiao/status/2…
Dmytro Dzhulgakov tweet media
Lin Qiao@lqiao

I’m extremely excited about working together with @cursor_ai and @Kimi_Moonshot to pushing the most cutting edge quality and inference performance – Kimi-k2.5 served as the foundation for Cursor continued pretraining & high-compute RL, powered by @FireworksAI_HQ RL infra & ultra-fast inference stack, delivering frontier coding performance.

English
1
0
0
164
Dmytro Dzhulgakov
Dmytro Dzhulgakov@dzhulgakov·
Data flywheel is the moat. Open base models + fine-tuning enable frontier quality/speed/cost tailored specifically for *your* AI product @FireworksAI_HQ RL platform takes care of GPUs/algo, you focus on crafting the best model quality. We powered Composer 2 RL run at a huge scale and would love to help your use case.
clem 🤗@ClementDelangue

Looks like it’s confirmed Cursor’s new model is based on Kimi! It reinforces a couple of things: - open-source keeps being the greatest competition enabler - another validation for chinese open-source that is now the biggest force shaping the global AI stack - the frontier is no longer just about who trains from scratch, but who adapts, fine-tunes, and productizes fastest (seeing the same thing with OpenClaw for example).

English
2
2
44
3.8K
Truth
Truth@NumCupboard·
@scottwww @adcock_brett @mcuban > They are ... No they are not. > It makes zero sense ... It does make sense. You are satisfied with finding local optima among infinitely many possible local optima. I want to find the global optimum. Big difference.
English
0
0
0
18
Scott Wainner
Scott Wainner@scottwww·
@NumCupboard @adcock_brett @mcuban They are for any tasks done by humans. Because the world was already designed for humans. It makes zero sense to create 1,000 specialized robots for different tasks that a human can do now.
English
1
0
0
23
Truth
Truth@NumCupboard·
@ericzakariasson That is nothing. Running an interpreted code editor in Chromium in a CLI-first agentic AI era is insane. I know you'll make the move at some point. But it surprises me that it hasn't happened yet. This could also fix your `com(.)todesktop(.)xyz` ID and Hilary issue.
English
0
0
0
81
Truth
Truth@NumCupboard·
@eliebakouch I don't believe that it is this what he is saying. He is using ambiguous wording to say the following: 25% of THEIR compute was spent on fine-tuning the base, and 75% on RL.
English
0
0
4
770
elie
elie@eliebakouch·
so 3x the training compute gets you 1% improvement on swe bench multilingual and 21% on terminal bench 2.0 but k2.5 is in non thinking mode? if those benchmarks are useless, it's weird that they are the ones reported in cursor blog then? something is wrong
elie tweet media
Lee Robinson@leerob

Yep, Composer 2 started from an open-source base! We will do full pretraining in the future. Only ~1/4 of the compute spent on the final model came from the base, the rest is from our training. This is why evals are very different. And yes, we are following the license through our inference partner terms.

English
31
21
708
137.2K
Truth
Truth@NumCupboard·
@sceptical_panda @teortaxesTex You are a freaking top-moron. You never looked at hardware company financials. You may be even more of a pseudo-intellectual moron than @teortaxesTex . If you'd surpass a 130 IQ threshold (which is not high), you'd realize that most of his tweets are slop.
English
0
0
0
133