
Lewis Mun Chun
1.7K posts

Lewis Mun Chun
@Lewismcmc
Lorem Ipsum Dolor Sit Amet

















Prompt: Make a looping HTML animation of a small robot arm stacking objects in a pattern



Meet Kimi Agentic Slides! Now with Nano Banana Pro 🍌 🎁 Thanksgiving Gift: 48H FREE & UNLIMITED ACCESS 🔸 Agentic search (Kimi K2) 🔸 Files → Slides (PDFs, images, docs+) 🔸 Fully editable + PPTX export 🔸 Designer-level visuals (infographics, illustrations) Try now: kimi.com/slides

🚀Introducing Lumine, a generalist AI agent trained within Genshin Impact that can perceive, reason, and act in real time, completing hours-long missions and following diverse instructions within complex 3D open-world environments.🎮 Website: lumine-ai.org 1/6

This paper from Tsinghua University and Shanghai Jiao Tong University received perfect scores (6, 6, 6, 6) at NeurIPS 2025! It aims to answer a key question: Does reinforcement learning really make large language models better reasoners? The authors study Reinforcement Learning with Verifiable Rewards (RLVR) and find that while it improves accuracy for small k, it doesn’t create new reasoning patterns—meaning the base model still determines the upper limit of reasoning ability. Across six RLVR variants, performance gains plateau, suggesting that current RL setups mainly refine reasoning rather than reinvent it. Interestingly, it’s distillation, not RL, that shows genuine signs of emergent reasoning. This research points to the next frontier for truly self-improving large language models. Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? Paper: arxiv.org/abs/2504.13837 Page: limit-of-RLVR.github.io Our report: mp.weixin.qq.com/s/2-GDxs8j1QYh… 📬 #PapersAccepted by Jiqizhixin









