Fred D. | 一铭

3.1K posts

Fred D. | 一铭 banner
Fred D. | 一铭

Fred D. | 一铭

@freddmts

- Agent coding addict : https://t.co/ZwbyJ8ZbKQ - Co-organizer, Vibe Coding Community Paris : https://t.co/y20yrcs1nn

Paris Katılım Eylül 2009
855 Takip Edilen335 Takipçiler
Fred D. | 一铭
Fred D. | 一铭@freddmts·
Si vous êtes à Paris le 14 avril, venez faire un tour au Meetup Vibe Coding Paris #2 : Controlled Autonomy chez @YesWeScale.
Vincent Le Gallic@vincentLg

Meetup Vibe Coding Paris #2 : Controlled Autonomy le 14/04 chez @YesWeScale 🚀 🔹 Talk #1 : @titouan_benoit (@DotfileApp) : Comment adapter sa DX pour des agents autonomes ? Architecture de systèmes où le bottleneck n'est plus l'écriture, mais le cycle d'exécution. 🔹 Talk #2 : @freddmts (@Cometh ) : Sortir du test manuel : mise en place de benchmarks automatisés et évaluation de la résilience des prompts avec @promptfoo. meetup.com/vibe-coding-co…

Français
0
1
0
20
Andrej Karpathy
Andrej Karpathy@karpathy·
One common issue with personalization in all LLMs is how distracting memory seems to be for the models. A single question from 2 months ago about some topic can keep coming up as some kind of a deep interest of mine with undue mentions in perpetuity. Some kind of trying too hard.
English
1.7K
1.1K
21.1K
2.6M
Kimi.ai
Kimi.ai@Kimi_Moonshot·
Zhilin at GTC: Introducing Attention Residuals Learning selective memory, rather than mechanically accumulating everything, is the beauty of attention. Many of you have probably read Attention Is All You Need, the 2017 Transformer paper that brought “human-like” attention into the model’s field of view. From that point on, models no longer simply read everything in a mechanical way. Instead, they began to develop a sense of what matters more and what matters less across the text, choosing to retain the more important information. Recently, Kimi applied this idea of attention to the temporal dimension, then rotated it 90 degrees into the model’s depth dimension. This allows the model to have attention not only over time, but also throughout the process of information transmission across layers—giving it a more intelligent way to understand and process information.
English
47
153
1.4K
99.9K
Fred D. | 一铭
Fred D. | 一铭@freddmts·
This matches my experience with OpenClaw. Since I share more personal details with it than with other assistants, even an offhand mention can turn into something it keeps bringing up.
Andrej Karpathy@karpathy

One common issue with personalization in all LLMs is how distracting memory seems to be for the models. A single question from 2 months ago about some topic can keep coming up as some kind of a deep interest of mine with undue mentions in perpetuity. Some kind of trying too hard.

English
0
0
0
230
Wei Ping
Wei Ping@_weiping·
🚀 Introducing Nemotron-Cascade 2 🚀 Just 3 months after Nemotron-Cascade 1, we’re releasing Nemotron-Cascade 2: an open 30B MoE with 3B active parameters, delivering best-in-class reasoning and strong agentic capabilities. 🥇 Gold Medal-level performance on IMO 2025, IOI 2025, and ICPC World Finals 2025: • Capabilities once thought achievable only by frontier proprietary models (e.g. Gemini Deep Think) or frontier-scale open models (i.e. DeepSeek-V3.2-Speciale-671B-A37B). • Remarkably high intelligence density with 20× fewer parameters. 🏆 Best-in-class across math, code reasoning, alignment, and instruction following: • Outperforms the latest Qwen3.5-35B-A3B (2026-02-24) and even larger Qwen3.5-122B-A10B (2026-03-11). 🧠 Powered by Cascade RL + multi-domain on-policy distillation: • Significantly expand Cascade RL across a much broader range of reasoning and agentic domains than Nemotron-Cascade 1, while distilling from the strongest intermediate teacher models throughout training to recover regressions and sustain gains. 🤗 Model + SFT + RL data: 👉 huggingface.co/collections/nv… 📄 Technical report: 👉 research.nvidia.com/labs/nemotron/…
Wei Ping tweet media
English
42
143
898
157.1K
Fred D. | 一铭
Fred D. | 一铭@freddmts·
@crystalsssup 😂 跟他合作过的中国人给他起了这个名字,其实是在提醒我们:他的水平有问题
中文
0
0
0
328
Fred D. | 一铭 retweetledi
Enes Akar
Enes Akar@enesakar·
Announcing Context7 CLI! MCP isn't the only way anymore. Now any AI agent can pull docs with Context7 — just the CLI and the find-docs skill. One command: npx ctx7 setup
English
48
110
1.1K
128.8K