Wand AI

124 posts

Wand AI banner
Wand AI

Wand AI

@WandAI_

Stop building agents. Start building hybrid workforces.

Palo Alto, CA Katılım Şubat 2023
30 Takip Edilen503 Takipçiler
Wand AI
Wand AI@WandAI_·
Wand is the operating system for the hybrid workforce, where humans and AI agents work side by side to get real work done. Work is changing. Wand is how it happens. Here’s what we build — and why it matters 👇 wand.ai/blog/meet-wand…
English
0
0
0
321
Wand AI retweetledi
M. V. Cunha
M. V. Cunha@mvcinvesting·
$NBIS just dropped another successful customer story. 👌🏻 Wand AI, an innovative player in LLM optimization, is leveraging $NBIS high-performance infrastructure to reinvent how large language models reason and scale. The goal was to reduce output length without sacrificing accuracy — all while cutting compute costs and latency. The results: • Shorter, more concise outputs • Maintained or improved accuracy across STEM benchmarks • Lower latency and more efficient scaling — even under tight inference constraints Thanks to $NBIS, Wand AI is now delivering smarter, leaner LLMs — faster, cheaper, and with zero performance trade-offs.
M. V. Cunha tweet media
English
20
22
311
22K
elvis
elvis@omarsar0·
// Concise Reasoning via Reinforcement Learning // I've seen this conclusion with reasoning LLMs in different forms: "conciseness often correlates with better performance." "We argue that the observed accuracy gains from generating longer responses are primarily a result of reinforcement learning’s loss optimization process, rather than an inherent need for extended responses." I wonder if there is something there. This new paper proposes a new training strategy that promotes concise and accurate reasoning in LLMs using RL. It challenges the belief that long responses improve accuracy; they offer both theoretical and empirical evidence showing that conciseness often correlates with better performance. • Long ≠ better reasoning – The authors mathematically show that RL with PPO tends to generate unnecessarily long responses, especially when answers are wrong. Surprisingly, shorter outputs correlate more with correct answers, across reasoning and non-reasoning models. • Two-phase RL for reasoning + conciseness – They introduce a two-phase RL strategy: (1) train on hard problems to build reasoning ability (length may increase), then (2) fine-tune on occasionally solvable tasks to enforce concise CoT without hurting accuracy. The second phase alone dramatically reduces token usage—by over 50%—with no loss in accuracy. • Works with tiny data – Their method succeeds with as few as 4–8 training examples, showing large gains in both math and STEM benchmarks (MATH, AIME24, MMLU-STEM). For instance, on MMLU-STEM, they improved accuracy by +12.5% while cutting response length by over 2×. • Better under low sampling – Post-trained models remain robust even when the temperature is reduced to 0. At temperature=0, the fine-tuned model outperformed the baseline by 10–30%, showing enhanced deterministic performance. • Practical implications – Besides improving model output, their method reduces latency, cost, and token usage, making LLMs more deployable. The authors also recommend setting λ < 1 during PPO to avoid instability and encourage correct response shaping. Very cool paper and lots of implications to continue improving the accessibility and efficiency of reasoning LLMs.
elvis tweet media
English
8
55
330
31.5K
Wand AI retweetledi
DAIR.AI
DAIR.AI@dair_ai·
Here are the top AI Papers of the Week (April 7 - 13): - NoProp - The AI Scientist V2 - Concise Reasoning via RL - Rethinking Reflection in Pre-Training - Efficient KG Reasoning for Small LLMs - Agentic Knowledgeable Self-awareness (bookmark for later) Read on for more:
English
2
77
374
50.1K
Wand AI
Wand AI@WandAI_·
Huge thanks to @rasbt for recognizing our latest research! We’re lucky to have some of the most brilliant minds in AI helping us push the boundaries of what's possible. Curious how we’re making reasoning more concise? Dive into the blog here: wand.ai/blog/from-prol…
Sebastian Raschka@rasbt

As we all know by now, reasoning models often generate longer responses, which raises compute costs. Now, this new paper (arxiv.org/abs/2504.05185) shows that this behavior comes from the RL training process, not from an actual need for long answers for better accuracy. The RL loss tends to favor longer responses when the model gets negative rewards, which I think explains the "aha" moments and longer chains of thought that arise from pure RL training. I.e., if the model gets a negative reward (i.e., the answer is wrong), the math behind PPO causes the average per-token loss becomes smaller when the response is longer. So, the model is indirectly encouraged to make its responses longer. This is true even if those extra tokens don't actually help solve the problem. What does the response length have to do with the loss? When the reward is negative, longer responses can dilute the penalty per individual token, which results in lower (i.e., better) loss values (even though the model is still getting the answer wrong). So the model "learns" that longer responses reduce the punishment, even though they are not helping correctness. In addition, the researchers show that a second round of RL (using just a few problems that are sometimes solvable) can shorten responses while preserving or even improving accuracy. This has big implications for deployment efficiency.

English
0
1
1
446
Wand AI
Wand AI@WandAI_·
If you're attending the Web Summit Qatar, don’t miss the chance to connect with Jean-Paul Sacy, our Head of Middle East & Africa. With 15,000+ founders, investors, journalists & visionaries gathering in Doha, it’s the perfect place to exchange ideas and explore new opportunities.
Wand AI tweet media
English
0
0
1
356
Wand AI
Wand AI@WandAI_·
How are humans and the AI workforce interacting? Rotem Alaluf (CEO at Wand AI), Ramesh Raskar (Associate Professor at MIT Media Lab), and Steve Nouri (CEO at GenAI) discusses critical topics that are shaping the future of human-AI collaboration. hubs.li/Q037zzwS0
English
0
0
1
276
Wand AI
Wand AI@WandAI_·
We’ve all heard the skepticism—“AI doesn’t really work.” But when done right, #enterpriseAI can transform how teams operate. One of our customers built an #AIworkforce within days using our platform, driving sales & increasing sales team capacity. Curious? hubs.li/Q033ppB30
English
0
0
2
178
Wand AI
Wand AI@WandAI_·
🎉 Unwrapping the Joy of #AIAgents: 12 AI Use Cases to Kick Start 2025! 🎁 5️⃣ AI Agents for Insurance Risk Mitigation – Unlocking Smarter Decisions We empower risk managers to move beyond reactive processes and into predictive, data-driven intelligence. hubs.li/Q030K4W80
English
0
0
0
248
Wand AI
Wand AI@WandAI_·
🎉 Unwrapping the Joy of #AIAgents: 12 AI Use Cases to Kick Start 2025! 🎁 4️⃣ AI Agents for Sales Teams: Sleigh Your Sales Goals Gift your Sales Team their own #A workforce to start boosting sales immediately. Request a demo today: hubs.li/Q030JRth0
English
0
0
1
188
Wand AI
Wand AI@WandAI_·
🎉 Unwrapping the Joy of #AIAgents: 12 #AI Use Cases to Kick Start 2025! 🎄 1: Finance – Deck the Books with AI Agents Give your finance team the gift of time. AI eliminates repetitive tasks, reduces errors, and lets your humans focus on strategic tasks. hubs.li/Q0306S2P0
English
0
0
1
170
Wand AI
Wand AI@WandAI_·
Discover how #AI can be trained to reason critically. Our Senior AI Researcher, Allen Roush, will present his latest paper: OpenDebateEvidence: A Massive-Scale Argument Mining and Summarization Dataset—a groundbreaking resource that helps #AIagents think like champion debaters.
Wand AI tweet media
English
0
0
3
172
Wand AI
Wand AI@WandAI_·
Our team is at #NeurIPS2024, one of the most anticipated AI and ML conferences of the year! We're excited to connect with some of the brightest minds and dive into groundbreaking research shaping the future of AI. hubs.li/Q02_dWr20
Wand AI tweet media
English
0
0
3
154