Razvan Dumitru

15 posts

Razvan Dumitru

Razvan Dumitru

@RazvanDuu

Katılım Haziran 2023
55 Takip Edilen16 Takipçiler
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
@rjsabouhi @maenstru56 @Vikas_NLP_UA @PanLiangming Exactly. ConciseRL formalizes that intuition: encode “be concise yet correct” in the reward, let PPO follow the gradient, and the attractors are short-but-sufficient chains. Appreciate the gradient-field lens!
English
0
0
1
71
RJ Sabouhi
RJ Sabouhi@rjsabouhi·
@RazvanDuu @maenstru56 @Vikas_NLP_UA @PanLiangming You already knew the attractors were encoded in the field. This makes it formal. “Gradient fields” learn the subtask structure hidden in the reward. Latent constraint flow as behavior. Symbolic recursion in motion.
RJ Sabouhi tweet media
English
1
0
1
117
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
+2.2 accuracy points using ~12.5× fewer tokens on TheoremQA. Length adapts to difficulty; traces read cleaner; drops smoothly into existing RL pipelines. (5/6)
English
0
0
0
78
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
On MATH: up to 31× fewer tokens on easier problems with ~+7 accuracy points; on the hardest tier: ~3.6× fewer tokens with +7.5 accuracy points. (4/6)
English
0
0
0
77
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
Reward semantic conciseness (not just shortness). An LLM judge scores whether a chain is dense and sufficient; combine with correctness so we trim fat without losing substance. (3/6)
Razvan Dumitru tweet media
English
0
0
0
68
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
Reasoning models overthink—long chains past the answer. That burns tokens, slows inference, and muddies evaluation. (2/6)
English
0
0
0
78
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
Up to 3.08× faster (second-turn MT-Redundant) and +49% on top of speculative decoding; benefits grow with longer context. Details in the paper. (5/6)
English
0
0
0
61
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
Drop-in on top of speculative decoding: rolling hashes find matches → a copy-and-verify path runs alongside your drafter → target model accepts/rejects. No special memory tricks. (4/6)
English
0
0
0
45
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
Real workflows repeat: multi-turn chat, RAG follow-ups, summaries, code edits. That redundancy is free speed if you can safely reuse instead of regenerate. (3/6)
English
0
0
0
44
Razvan Dumitru
Razvan Dumitru@RazvanDuu·
CopySpec watches the running text, spots a chunk that already exists in the context/previous turn, proposes a “copy block,” and lets the target model check it. Same output, less generation. (2/6)
English
0
0
0
42
Razvan Dumitru retweetledi
Vikas Yadav
Vikas Yadav@Vikas_NLP_UA·
🎉 Our work “Variable Layerwise Quantization: A Simple and Effective Approach to Quantize LLMs” is accepted at #ACLFindings2025 📎 arxiv.org/abs/2406.17415 — Keep key layers high-precision, push others lower → compact LLMs w/ ~no accuracy loss — Simple LIM & ZD scores rank layers
English
1
3
6
876
Razvan Dumitru retweetledi
Darius Peteleaza
Darius Peteleaza@maenstru56·
🧵Excited to attend #ICML2024 and present our paper titled "Enhancing Transformer RNNs with Multiple Temporal Perspectives" at the "Next Generation of Sequence Modeling Architectures" workshop! #AI
Darius Peteleaza tweet media
English
1
1
4
464