DaemonEye

2.3K posts

DaemonEye

DaemonEye

@DaemonEye

围观改变中国!!!

China Shanghai Katılım Ocak 2010
536 Takip Edilen222 Takipçiler
DaemonEye
DaemonEye@DaemonEye·
@interjc 2026了除了本地mac跑模型好点,看不出win作为开发宿主机有什么问题,docker拉不起来还是wsl2不够你用?
中文
0
0
1
64
Justin
Justin@interjc·
程序员入职发 Windows 电脑的公司基本可以避雷
Justin tweet media
中文
178
3
128
57.1K
DaemonEye
DaemonEye@DaemonEye·
@geniusvczh 用户态线程调度python圈子10几年前就有greenlet那些玩意了。
中文
0
0
0
540
karminski-牙医
karminski-牙医@karminski3·
惊闻微软把BOC搞出来了???? 好久没关注编程语言的进展了. 大家知道CSP(比如golang)是显式传递指针的,而BOC是隐式传递指针的, 这玩意最牛逼的点是, 传统Actor是一维的锁拓扑, 而这玩意是个二维的DAG! 所以锁粒度可以做到极致. 理论吞吐量嗷嗷高. 那么老问题来了, 就py那个垃圾GIL, 现在实现BOC必然要涉及到内存拷贝. boc将一个指针从线程A传递到线程B, 如果指针内部引用了py的原生数据结构, 就必然要拷内存 (不拷就直接segment fault了). 所以为了方便无拷贝传递指针, 我估计它还得在堆里自己搞个C实现的原始数据结构(FFI时代的老经验谈). 至于这玩意是否兼容py3.13那就不知道了... 总之这个是特别值得关注的, 算是PL圈子的爆炸性新闻了. 但无奈现在应该没有手工编程艺人了. microsoft.github.io/bocpy/
中文
15
42
389
85.7K
DaemonEye
DaemonEye@DaemonEye·
@lovelyshuimao 他整个3.0 3.1的tool call就有问题,看看3.2能否解决,推理能力和世界知识真的没问题。
中文
1
0
2
506
lovely水
lovely水@lovelyshuimao·
Gemini最大的问题就是搜索不积极 作为一个谷歌的AI真的感觉到匪夷所思
中文
56
2
329
46.2K
DaemonEye
DaemonEye@DaemonEye·
@python_xxt 如果只拿来查知识,补充推理那gemini太强了。他的问题只有在堆叠了一堆tools跑很大流程的coding时候才会暴雷...
中文
0
0
5
737
Robinson · 鲁棒逊
Robinson · 鲁棒逊@python_xxt·
原推这个,和我体感高度一致 并不是所有人都有意愿,去探索AI工具的上限 比如说, 我反复和团队的人, 安利 Claude 的 Opus 4.6, 安利 OpenAI 的 Chatgpt 5.5 , 然后,只安利成功了 Gemini, 盖因为 Google的AI Studio 免费... 结果现在 很多人还在用 gemini 3.1 pro, 还有一部分人在用豆包 和 deepseek ... 虽然我认可只要用AI 模型就好过不用AI模型, 但是推友们都能理解并接受 “用力所能及能负担的最强模型”,身边的成员却无法践行,真的让我挺受挫的,其中不乏一些我认为非常优秀的人... 这也让我深刻意识到,并不是所有人都有意愿,去探索工具的上限,这终归是少数人的需求。 对于大多人来说,“够用”就行了。
Tracy@CTracy0803

虽然大家普遍认为Claude是最牛的大模型公司,周围也都在用Claude。 但实际上,Claude的月活只有2000-3000 万人,是Gemini的1/30,更是Chatgpt的1/40。

中文
61
5
95
47K
DaemonEye
DaemonEye@DaemonEye·
@mranti 错了,用户需要用最贵的模型去跑通流程,跑通后固定成skill再去用低价格的模型压长期调用成本,上来就用差的,光是跑起来不出错就折腾死人
中文
0
0
0
448
DaemonEye
DaemonEye@DaemonEye·
@liumengxinfly 你逼自己human in the loop 光是看那些thinking trace都能学到一捆知识。看agent怎么玩aws/gh/kubectl那几个命令都能看到稀奇古怪的玩法。
中文
0
0
0
89
Mengxin Liu
Mengxin Liu@liumengxinfly·
我发现 Vibe Coding 容易有失落感的原因了。之前做项目哪怕最终结果完全失败,过程中也能学会很多东西,有很多随机的发现,体验感是很足的。而现在哪怕项目完全成功,自己可能也啥也没学会,还容易否定自己。
中文
16
4
82
20.5K
DaemonEye
DaemonEye@DaemonEye·
@Rikka_Ela 太正常了,你这东西又不是数学/写代码可以快速的RL左脚踩右脚螺旋上升。对他来说都是语料,20年前的教材,5年前写的wiki和今年的论文都是文本材料。
中文
0
0
3
146
Rika☁️
Rika☁️@Rikka_Ela·
暴露:目前市面上所有的AI在化学/生物方面的语料库几乎都是一坨大的,很多时候都是一本正经的胡说八道。我平常用的Deepseek, GPT和Gemini,在高等有机/群论与光谱,遗传基因学/细胞免疫学上这三玩意简直是谁拉这了.jpg,老是左右脑互搏,还经常拿十几年前就已经淘汰的理论拿来解释
Rika☁️ tweet media
中文
50
18
310
26K
DaemonEye
DaemonEye@DaemonEye·
@m0d8ye 他那玩意最大的问题就是toolcall,完全不可用
中文
0
0
0
762
Max Lv
Max Lv@m0d8ye·
gemini 3.1 pro 的 benchmark 分数那么高,但感觉很少有人提 gemini-cli?
中文
25
0
10
13.8K
DaemonEye
DaemonEye@DaemonEye·
@geniusvczh 你这准备外挂Claude 4.7写设计文档 todo之后塞给co pilot嫖1x 的5.4xhigh吗?
中文
0
0
0
823
geniusvczh
geniusvczh@geniusvczh·
用了一整天的gpt 5.5,感觉比起5.4有一个巨大的飞跃,就是现在任务可以布置得更复杂。这不是说他可以解决比5.4难得多的问题,而是本来已经差点撑爆5.4的任务,5.5可以连着干好几个。 7.5x就7.5x吧。只要习惯了长程任务,一个5.5 request可以当几个5.4用,也不是不行,pro+半年后才过期,先榨干🤪
中文
20
5
75
14.1K
geniusvczh
geniusvczh@geniusvczh·
@DaemonEye 我已经熟练掌握让AI一次干超多活不等他的技术了,慢一点就慢一点,反正我也不等他🤪
中文
1
0
0
149
geniusvczh
geniusvczh@geniusvczh·
以前写代码要强劲的机器是因为修bug修的太快需要频繁跑test。现在AI做一件事花的时间是我的好几倍(如果只算决定代码怎么写的话,可以到几十倍),build的性能已经不重要了。我已经把整个开发环境移到了本来给咖啡烘豆机当UI的Yoga 7上,$4500配的台式机除了打游戏已经什么都不做了🤪
中文
5
0
25
7.3K
DaemonEye
DaemonEye@DaemonEye·
@berryxia 这类思路见过好几篇了,基本都论证比如看10m级别的长文让ai grep跳着读效果好得多
中文
0
0
0
731
Berryxia.AI
Berryxia.AI@berryxia·
MIT一招,直接把AI巨头过去5年的百亿“上下文窗口军备竞赛”打成笑话! 所有大模型最头疼的“Context Rot”(上下文腐烂)终于被干掉了! MIT CSAIL三位研究员刚放出的RLM(Recursive Language Models)直接把规则改了: ✅ 超长文档不再塞进AI窗口,而是以外部Python变量存着 ✅ AI像顶尖程序员一样写代码:正则搜索、结构导航、精准切片 ✅ 需要哪块就只拉哪块,读完后递归生成子AI并行分析,再主AI合成答案 ✅ 无总结、无丢失、无性能衰减 实测结果炸裂: • 最难的长上下文基准上,传统前沿模型接近0分 • RLM直接双位数百分比提升 • 可处理1000万token(主流模型原生窗口的100倍!) • 成本甚至更低 代码已开源,GitHub一键替换现有LLM调用,零改动就能支持以前完全不可能的超长任务。 上下文窗口战争?结束了。 MIT不是把窗口做更大,而是直接走出了战场。 这才是2026年真正的具身智能/长期Agent杀手级方案。 (原论文+完整代码已附👇)
Berryxia.AI tweet media
Elias Al@iam_elias1

MIT just made every AI company's billion dollar bet look embarrassing. They solved AI memory. Not by building a bigger brain. By teaching it how to read. The paper dropped on December 31, 2025. Three MIT CSAIL researchers. One idea so obvious it hurts. And a result that makes five years of context window arms racing look like the wrong war entirely. Here is the problem nobody solved. Every AI model on the planet has a hard ceiling. A context window. The maximum amount of text it can hold in working memory at once. Cross that line and something ugly happens — something researchers have a clinical name for. Context rot. The more you pack into an AI's context, the worse it performs on everything already inside it. Facts blur. Information buried in the middle vanishes. The model does not become more capable as you feed it more. It becomes more confused. You give it your entire codebase and it forgets what it read three files ago. You hand it a 500-page legal document and it loses the clause from page 12 by the time it reaches page 400. So the industry built a workaround. RAG. Retrieval Augmented Generation. Chop the document into chunks. Store them in a database. Retrieve the relevant ones when needed. It was always a compromise dressed up as a solution. The retriever guesses which chunks matter before the AI has read anything. If it guesses wrong — and it does, constantly — the AI never sees the information it needed. The act of chunking destroys every relationship between distant paragraphs. The full picture gets shredded into fragments that the AI then tries to reassemble blindfolded. Two bad options. One broken industry. Three MIT researchers and a deadline of December 31st. Here is what they built. Stop putting the document in the AI's memory at all. That is the entire idea. That is the breakthrough. Store the document as a Python variable outside the AI's context window entirely. Tell the AI the variable exists and how big it is. Then get out of the way. When you ask a question, the AI does not try to remember anything. It behaves like a human expert dropped into a library with a computer. It writes code. It searches the document with regular expressions. It slices to the exact section it needs. It scans the structure. It navigates. It finds precisely what is relevant and pulls only that into its active window. Then it does something that makes this recursive. When the AI finds relevant material, it spawns smaller sub-AI instances to read and analyze those sections in parallel. Each one focused. Each one fast. Each one reporting back. The root AI synthesizes everything and produces an answer. No summarization. No deletion. No information loss. No decay. Every byte of the original document remains intact, accessible, and queryable for as long as you need it. Now here are the numbers. Standard frontier models on the hardest long-context reasoning benchmarks: scores near zero. Complete collapse. GPT-5 on a benchmark requiring it to track complex code history beyond 75,000 tokens — could not solve even 10% of problems. RLMs on the same benchmarks: solved them. Dramatically. Double-digit percentage gains over every alternative approach. Successfully handling inputs up to 10 million tokens — 100 times beyond a model's native context window. Cost per query: comparable to or cheaper than standard massive context calls. Read that again. One hundred times the context. Better answers. Same price. The timeline of the arms race makes this sting harder. GPT-3 in 2020: 4,000 tokens. GPT-4: 32,000. Claude 3: 200,000. Gemini: 1 million. Gemini 2: 2 million. Every generation, every company, billions of dollars spent, all betting on the same assumption. More context equals better performance. MIT just proved that assumption was wrong the entire time. Not slightly wrong. Fundamentally wrong. The entire premise of the last five years of context window research — that the solution to AI memory was a bigger window — was the wrong answer to the wrong question. The right question was never how much can you force an AI to hold in its head. It was whether you could teach an AI to know where to look. A human expert handed a 10,000-page archive does not read all 10,000 pages before answering your question. They navigate. They search. They find the relevant section, read it deeply, and synthesize the answer. RLMs are the first AI architecture that works the same way. The code is open source. On GitHub right now. Free. No license fees. No API costs. Drop it in as a replacement for your existing LLM API calls and your application does not even notice the difference — except that it suddenly works on inputs it used to fail on entirely. Prime Intellect — one of the leading AI research labs in the space — has already called RLMs a major research focus and described what comes next: teaching models to manage their own context through reinforcement learning, enabling agents to solve tasks spanning not hours, but weeks and months. The context window wars are over. MIT won them by walking away from the battlefield. Source: Zhang, Kraska, Khattab · MIT CSAIL · arXiv:2512.24601 Paper: arxiv.org/abs/2512.24601 GitHub: github.com/alexzhang13/rlm

中文
22
118
502
101.4K
DaemonEye
DaemonEye@DaemonEye·
@raycat2021 不做旅游攻略惨案了。我包车一天40刀卧槽司机还兼职导游,爽得要死。 前年我在luxor包车包专业导游那价格低的男默女泪
中文
0
0
1
224
徐老猫
徐老猫@raycat2021·
这位英国游客在埃及时叫了出租去机场,谈好付250埃镑,约合4.83美元。 司机在高速公路半道停车,坚持要250美元,不然就不走了。 埃及到处是大大小小的各种坑,这位游客怎么不做点功课呢?
中文
126
33
307
89.2K
DaemonEye
DaemonEye@DaemonEye·
@CarbonizedC 什么老好人,小心德州和fl画炸了因为Latino倒戈倒输回去
中文
0
0
0
107
DaemonEye
DaemonEye@DaemonEye·
@jiongnasen 德州这轮重画,佛罗里达这轮重画州里投票了吗?你别只吃肉啊。还是大家一起坐下来讨论大家都不搞这套?不过到时候你又要打滚抛弃美国先贤了对吧
中文
1
0
16
571
约拿坦
约拿坦@jiongnasen·
奥巴马2020版:“为了挽救民主,我们必须终结gerrymandering!!” 奥巴马2026版:“为了挽救民主,我们必须加强gerrymendering!!!” 图二是2024年选举结果,图三是今天投票通过后将要重划的选区地图。 不要脸三个字就是民主党的座右铭。
约拿坦 tweet media约拿坦 tweet media约拿坦 tweet media
Barack Obama@BarackObama

Congratulations, Virginia! Republicans are trying to tilt the midterm elections in their favor, but they haven’t done it yet. Thanks for showing us what it looks like to stand up for our democracy and fight back.

中文
28
27
146
15.6K
DaemonEye
DaemonEye@DaemonEye·
@geniusvczh 但是昨天你们微软的gpt5.4爆炸了吧,那个速度慢的我怀疑1s 10token都没了
中文
1
0
0
338
geniusvczh
geniusvczh@geniusvczh·
没降智是因为微软自己买的API给你当中转站。我的pro+还有opus 4.7,但是这个7.5x比起1x的gpt 5.4 xhigh那就是路边一条了,狗都不用🤪
Oasis Feng@oasisfeng

前几天还在庆幸 GitHub Copilot 的 Opus 4.6 没有降智,结果今天 GitHub 就直接撤回了所有 Pro 订阅对 Claude Opus 全系的访问权……(仅提供给 Pro+ 及以上订阅套餐了)🫠 看来还真是,全行业都已经烧不起了。 #opus-models-removed-from-pro" target="_blank" rel="nofollow noopener">github.blog/changelog/2026…

中文
4
1
21
7.3K