cedricporter

509 posts

cedricporter

cedricporter

@Stupid_ET

Explorer

Katılım Eylül 2012
1.8K Takip Edilen90 Takipçiler
cedricporter retweetledi
cedricporter retweetledi
cedricporter retweetledi
章工
章工@435hz·
真正赚到大钱的人,从来不是聪明、人脉、运气,而是静下来专注做一件事,长期积攒复利等待爆发。
中文
19
41
370
14.5K
cedricporter retweetledi
陳威廉
陳威廉@williamlab·
我每天看各种报告,刷各种文章,真的深有同感。。 能说人话不说人话,全是各种AI味道很浓的故作高深的辞藻堆叠,真服了。 真是所谓“为赋新词强说愁”。
陳威廉 tweet media
中文
9
2
16
5.2K
cedricporter retweetledi
Elon Musk
Elon Musk@elonmusk·
This is still 0.5T, but a more recent training checkpoint. 1T model is ~5 days away from finishing initial training. Will be a major step change improvement in coding, long context and skills. The SpaceXAI model factory is finally working. Should be an improved base model landing every ~2 weeks.
English
264
325
4.2K
372.3K
cedricporter retweetledi
Elon Musk
Elon Musk@elonmusk·
Universal HIGH INCOME via checks issued by the Federal government is the best way to deal with unemployment caused by AI. AI/robotics will produce goods & services far in excess of the increase in the money supply, so there will not be inflation.
English
46.6K
22.8K
195.2K
69M
cedricporter retweetledi
向阳乔木
向阳乔木@vista8·
书中读到这个段落,感觉哈萨比斯确实聪明! 2013年,扎克伯格意识到 Facebook 在 AI 领域已经落后,开始疯狂追赶。 为阻止 DeepMind 落入谷歌之手,Facebook 提出了极具诱惑力的条件,还邀请哈萨比斯到家里参加私人晚宴。 席间,哈萨比斯对扎克伯格实施了一次“微妙的测试”: 哈萨比斯先讨论了 AI 的潜力,然后故意转换话题,聊起了当时其他一些热门的技术趋势,包括VR、AR以及3D打印。 面对完全不同的技术方向,扎克伯格表现得“同样兴奋”。 说明小扎只是单纯地追逐每个技术热点,并没能真正理解 AI 为什么比其他任何技术都更伟大、更具决定性。 最终,哈萨比斯拒绝了扎克伯格,选择了谷歌。
中文
86
41
592
121.5K
cedricporter retweetledi
Ramin Nasibov
Ramin Nasibov@RaminNasibov·
Ramin Nasibov tweet media
ZXX
112
1.8K
12.1K
216.4K
cedricporter retweetledi
shenxiao
shenxiao@Shenxiao123971·
这是最接近四维的视角了😱
中文
1.1K
13.6K
145.6K
13.6M
cedricporter retweetledi
Midas Trend
Midas Trend@RealMidasTrend·
张雪是越挖越有料啊。张雪分享自我激励的心法,这在心理学上是有科学依据的,也是很多心理学家推荐的方法。 不得不说,张雪悟性是真的很高,自己悟出了这么好的方法。 可惜我们学校的教育方法总是挫折教育,老师天天都是盯着学生的缺点问题在管,很难看到孩子的优点和成绩。 反正孩子上了10年学,每次老师打电话都是告状,有家长说,每当一接到学校老师电话,心里就一紧,心脏莫名难受,都担心长期下去自己会犯心脏病了。
中文
82
627
4.1K
334K
cedricporter retweetledi
Cos(余弦)😶‍🌫️
还是继续提醒下吧,不要用你的 IDE 打开任何陌生的目录。这种 1Click 利用手法,虽然在这些 IDE 新版都做了缓解措施,但还可以绕过。 项目方注意自己的研发人员,别成为这种投毒攻击的靶子。Drift 2.8 亿美金被盗教训就在眼前。
Cos(余弦)😶‍🌫️ tweet media
Cos(余弦)😶‍🌫️@evilcos

特意处理成的极简无害的 PoC,可以用你的 IDE 玩下(如 Cursor/VS Code/Antigravity/TRAE): github.com/evilcos/vscode… 纯演示,具体步骤看仓库的 README。 再再次提醒:在野攻击已经不少,身边已经有几个中招真实案例,比如愿意公开这个遭遇的 @brucexu_eth 所幸止损及时无任何损失。😎

中文
10
26
182
55.1K
cedricporter retweetledi
Fuli Luo
Fuli Luo@_LuoFuli·
Two days ago, Anthropic cut off third-party harnesses from using Claude subscriptions — not surprising. Three days ago, MiMo launched its Token Plan — a design I spent real time on, and what I believe is a serious attempt at getting compute allocation and agent harness development right. Putting these two things together, some thoughts: 1. Claude Code's subscription is a beautifully designed system for balanced compute allocation. My guess — it doesn't make money, possibly bleeds it, unless their API margins are 10-20x, which I doubt. I can't rigorously calculate the losses from third-party harnesses plugging in, but I've looked at OpenClaw's context management up close — it's bad. Within a single user query, it fires off rounds of low-value tool calls as separate API requests, each carrying a long context window (often >100K tokens) — wasteful even with cache hits, and in extreme cases driving up cache miss rates for other queries. The actual request count per query ends up several times higher than Claude Code's own framework. Translated to API pricing, the real cost is probably tens of times the subscription price. That's not a gap — that's a crater. 2. Third-party harnesses like OpenClaw/OpenCode can still call Claude via API — they just can't ride on subscriptions anymore. Short term, these agent users will feel the pain, costs jumping easily tens of times. But that pressure is exactly what pushes these harnesses to improve context management, maximize prompt cache hit rates to reuse processed context, cut wasteful token burn. Pain eventually converts to engineering discipline. 3. I'd urge LLM companies not to blindly race to the bottom on pricing before figuring out how to price a coding plan without hemorrhaging money. Selling tokens dirt cheap while leaving the door wide open to third-party harnesses looks nice to users, but it's a trap — the same trap Anthropic just walked out of. The deeper problem: if users burn their attention on low-quality agent harnesses, highly unstable and slow inference services, and models downgraded to cut costs, only to find they still can't get anything done — that's not a healthy cycle for user experience or retention. 4. On MiMo Token Plan — it supports third-party harnesses, billed by token quota, same logic as Claude's newly launched extra usage packages. Because what we're going for is long-term stable delivery of high-quality models and services — not getting you to impulse-pay and then abandon ship. The bigger picture: global compute capacity can't keep up with the token demand agents are creating. The real way forward isn't cheaper tokens — it's co-evolution. "More token-efficient agent harnesses" × "more powerful and efficient models." Anthropic's move, whether they intended it or not, is pushing the entire ecosystem — open source and closed source alike — in that direction. That's probably a good thing. The Agent era doesn't belong to whoever burns the most compute. It belongs to whoever uses it wisely.
English
173
229
1.8K
752.1K
cedricporter retweetledi
Carl Zha
Carl Zha@CarlZha·
The Chinese language is highly contextual. Know your context:
Carl Zha tweet media
English
95
170
2.1K
192.8K
cedricporter retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an "idea file". The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it for your specific needs. So here's the idea in a gist format: gist.github.com/karpathy/442a6… You can give this to your agent and it can build you your own LLM wiki and guide you on how to use it etc. It's intentionally kept a little bit abstract/vague because there are so many directions to take this in. And ofc, people can adjust the idea or contribute their own in the Discussion which is cool.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
1.1K
2.8K
26.4K
6.8M
魔都老猿
魔都老猿@AriXZone·
小米集团股价腰斩,知名私募怒怼! 来源:上海证券报 日斗投资董事长王文在微博上直言:“小米做了那么多个行业,其实没有几个能做到行业前三,既然手上的活都没有做好,沉下心来把手上的先干好不行吗?现在汽车放缓了,又讲AI和机器人。可是AI和机器人,在这个时代,最聪明的人和钱都在那里,你小米凭啥竞争呢?我甚至怀疑小米到底有没有长期规划,往往哪个成热点,它就进入哪里。”
魔都老猿 tweet media魔都老猿 tweet media魔都老猿 tweet media
中文
80
16
153
124.2K
cedricporter retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.8K
6.9K
57.5K
20.6M
cedricporter retweetledi
dontbesilent
dontbesilent@dontbesilent·
Mac 上的豆包输入法好像把闪电说干掉了。。 我本来在闪电说用的就是豆包的大模型
中文
24
0
51
25.2K
SimbaLee
SimbaLee@lipeng0820·
豆包语音输入法!YYDS!其他语音输入法可以退下了。 前两天跟朋友聊起来,人类的信息输出方式里,语言是最快的,AI时代打字非常影响效率,脑机接口实现之前,语音输入是最好的过渡形式。 此刻,豆包应该能暂时成为这个过渡阶段最好的选择了。
中文
48
7
103
122.5K
cedricporter retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Software horror: litellm PyPI supply chain attack. Simple `pip install litellm` was enough to exfiltrate SSH keys, AWS/GCP/Azure creds, Kubernetes configs, git credentials, env vars (all your API keys), shell history, crypto wallets, SSL private keys, CI/CD secrets, database passwords. LiteLLM itself has 97 million downloads per month which is already terrible, but much worse, the contagion spreads to any project that depends on litellm. For example, if you did `pip install dspy` (which depended on litellm>=1.64.0), you'd also be pwnd. Same for any other large project that depended on litellm. Afaict the poisoned version was up for only less than ~1 hour. The attack had a bug which led to its discovery - Callum McMahon was using an MCP plugin inside Cursor that pulled in litellm as a transitive dependency. When litellm 1.82.8 installed, their machine ran out of RAM and crashed. So if the attacker didn't vibe code this attack it could have been undetected for many days or weeks. Supply chain attacks like this are basically the scariest thing imaginable in modern software. Every time you install any depedency you could be pulling in a poisoned package anywhere deep inside its entire depedency tree. This is especially risky with large projects that might have lots and lots of dependencies. The credentials that do get stolen in each attack can then be used to take over more accounts and compromise more packages. Classical software engineering would have you believe that dependencies are good (we're building pyramids from bricks), but imo this has to be re-evaluated, and it's why I've been so growingly averse to them, preferring to use LLMs to "yoink" functionality when it's simple enough and possible.
Daniel Hnyk@hnykda

LiteLLM HAS BEEN COMPROMISED, DO NOT UPDATE. We just discovered that LiteLLM pypi release 1.82.8. It has been compromised, it contains litellm_init.pth with base64 encoded instructions to send all the credentials it can find to remote server + self-replicate. link below

English
1.4K
5.4K
28.1K
66.5M