Ramakanth Kavuluru

3.5K posts

Ramakanth Kavuluru banner
Ramakanth Kavuluru

Ramakanth Kavuluru

@BioNLProc

Faculty at UKY. Views my own, not of my employer(s). Work: #BioNLP, #NLProc, medical informatics, machine learning, LLMs, AI & fairness, health+socialdata

Lexington, KY Katılım Ekim 2016
298 Takip Edilen750 Takipçiler
Ramakanth Kavuluru retweetledi
Kiran Garimella
Kiran Garimella@gvrkiran·
We have some new results showing the effectiveness of peer fact checking, operationalized on WhatsApp. Unlike most studies which use artificial headlines, our intervention was done on real user-donated WhatsApp data. Paper and details in the thread below:
Sudhamshu@theSudhamshow

Addressing misinformation on messaging apps like WhatsApp is challenging. E2EE means apps can't moderate at scale. Fact-checking tiplines face challenges due to low awareness, trust. Despite millions in the Global South relying on WA for news, corrections often do not reach them.

English
0
2
17
2.4K
Ramakanth Kavuluru
Ramakanth Kavuluru@BioNLProc·
@SenMcConnell Dear senator, I read the full letter of Mr. Kent that l link below for others to see too. I am unable to interpret its contents as "virulent anti-semitism". He did not attack jewish people for their faith. I would appreciate any clarifications. x.com/joekent16jan19…
Joe Kent@joekent16jan19

After much reflection, I have decided to resign from my position as Director of the National Counterterrorism Center, effective today. I cannot in good conscience support the ongoing war in Iran. Iran posed no imminent threat to our nation, and it is clear that we started this war due to pressure from Israel and its powerful American lobby. It has been an honor serving under @POTUS and @DNIGabbard and leading the professionals at NCTC. May God bless America.

English
0
0
0
25
U.S. Senator Mitch McConnell
Joe Kent testified before the Senate one year ago that Iran and its terror proxies threatened U.S. servicemembers in the Middle East. He said it would be an honor to return to the fight against terrorism, and he pledged to lead with integrity and accountability. The virulent anti-Semitism of his resignation letter makes it clear that Mr. Kent is incapable of upholding these pledges, and those who mistake its baseless and incendiary conspiracies for brave truth-telling are only fooling themselves. Isolationists and anti-Semites have no place in either party, and certainly do not deserve places of trust in our government.
English
4.9K
3.6K
19.9K
1.9M
Ramakanth Kavuluru retweetledi
Sepp Hochreiter
Sepp Hochreiter@HochreiterSepp·
xLSTM Distillation: arxiv.org/abs/2603.15590 Near-lossless distillation of quadratic Transformer LLMs into linear xLSTM architectures enables cost- and energy-efficient alternatives without sacrificing performance. xLSTM variants of instruction-tuned Llama, Qwen, & Olmo models.
Sepp Hochreiter tweet mediaSepp Hochreiter tweet media
English
5
59
310
21.9K
Ramakanth Kavuluru
Ramakanth Kavuluru@BioNLProc·
It's like every few months there are mass layoffs in the industry. I hear that our CS department is getting top notch applicants this hiring cycle, thanks to this fiasco. Happy we are going to benefit from this 🙂 cnbc.com/2026/03/14/met…
English
0
0
1
190
Ramakanth Kavuluru
Ramakanth Kavuluru@BioNLProc·
Happy to receive this research mentorship Immy 🙂 award in our department today.
Ramakanth Kavuluru tweet mediaRamakanth Kavuluru tweet mediaRamakanth Kavuluru tweet mediaRamakanth Kavuluru tweet media
English
1
0
8
185
Ramakanth Kavuluru retweetledi
Adam Rodman
Adam Rodman@AdamRodmanMD·
Our study of AMIE at @BIDMC_Medicine is out! You can read about what we did in posts from my co-authors (including the Google post below). But I wanted to talk about some background for this study, and what I think are most interesting findings. 🧵⬇️ x.com/GoogleResearch…
Google Research@GoogleResearch

Today we announce results from a first-of-its-kind study with @BIDMC_Medicine on AMIE, our conversational AI for clinical reasoning. In a real-world clinical study, AMIE was found to be safe, feasible, and well-received by patients. Learn more: goo.gle/4sXCogz

English
5
29
98
55.1K
Ramakanth Kavuluru retweetledi
Bryan Catanzaro
Bryan Catanzaro@ctnzr·
Announcing NVIDIA Nemotron 3 Super! 💚120B-12A Hybrid SSM Latent MoE, designed for Blackwell 💚36 on AAIndex v4 💚up to 2.2X faster than GPT-OSS-120B in FP4 💚Open data, open recipe, open weights Models, Tech report, etc. here: research.nvidia.com/labs/nemotron/… And yes, Ultra is coming!
Bryan Catanzaro tweet media
English
62
206
1.2K
201.1K
Ramakanth Kavuluru retweetledi
Tu Vu
Tu Vu@tuvllms·
🚨 New paper 🚨 Excited to share PRISM, a new “DeepThink” method that uses step-level correctness signals from a process reward model to guide inference over candidate solutions. PRISM matches or beats SOTA methods, enabling gpt-oss-20b to exceed gpt-oss-120b.👇 📰: arxiv.org/abs/2603.02479 #AI #LLMs
Tu Vu tweet media
English
3
31
136
13.3K
Ramakanth Kavuluru retweetledi
Microsoft Research
Microsoft Research@MSFTResearch·
Vision-language models improve multimodal systems, but can make them slower, costlier, and harder to deploy. Learn how Phi-4-reasoning-vision-15B, a compact and fast multimodal reasoning model, blends strengths of different methods while reducing their limits: msft.it/6014Q5X0u
Microsoft Research tweet media
English
1
19
71
16.8K
Ramakanth Kavuluru retweetledi
Joël Niklaus
Joël Niklaus@joelniklaus·
We just released pre-mixed, pre-shuffled pretraining datasets at 100BT scale. @asankhaya tested 50+ different mixture strategies at 1B scale. The winner? A static 50% finePDFs + 30% DCLM + 20% FineWeb-Edu blend. No fancy curriculum needed. We scaled this up to 100BT and pre-shuffled everything so you don't have to burn compute on sampling. Just use it: from datasets import load_dataset ds = load_dataset("HuggingFaceFW/finepdfs_50BT-dclm_30BT-fineweb_edu_20BT-shuffled") Browse the full smol-data collection: huggingface.co/collections/Hu… Reproduce it yourself: github.com/huggingface/da… Read the methodology: huggingface.co/blog/codelion/…
Joël Niklaus tweet media
English
3
28
198
13.6K
Ramakanth Kavuluru retweetledi
Qwen
Qwen@Alibaba_Qwen·
🚀 Introducing the Qwen 3.5 Small Model Series Qwen3.5-0.8B · Qwen3.5-2B · Qwen3.5-4B · Qwen3.5-9B ✨ More intelligence, less compute. These small models are built on the same Qwen3.5 foundation — native multimodal, improved architecture, scaled RL: • 0.8B / 2B → tiny, fast, great for edge device • 4B → a surprisingly strong multimodal base for lightweight agents • 9B → compact, but already closing the gap with much larger models And yes — we’re also releasing the Base models as well. We hope this better supports research, experimentation, and real-world industrial innovation. Hugging Face: huggingface.co/collections/Qw… ModelScope: modelscope.cn/collections/Qw…
Qwen tweet media
English
913
2.9K
21.4K
8.9M
Ramakanth Kavuluru retweetledi
Tinker
Tinker@tinkerapi·
To support open and collaborative science, we offer Tinker grants for researchers advancing the field. This week we’re featuring publications by some of our early research grant recipients! thinkingmachines.ai/blog/tinker-re…
English
3
39
258
152.5K
Ramakanth Kavuluru retweetledi
Zhijian Liu
Zhijian Liu@zhijianliu_·
Reasoning LLMs generate very long chains-of-thought, so even small quantization errors add up. With AWQ, Qwen3-4B drops 71.0 → 68.2 on MMLU-Pro (~4% relative loss). 😬 ParoQuant fixes this! It keeps only the critical rotation pairs and fuses everything into a single kernel. Recovers most of the lost reasoning accuracy with minimal overhead — so 4-bit models stay strong at reasoning. 💪💪
English
31
146
1.4K
167.9K
Ramakanth Kavuluru
Ramakanth Kavuluru@BioNLProc·
Twitter is doing LinkedIn nonsense these days. Someone joins a big tech/frontier AI lab and fifty people tweet a welcome message to them 🤦‍♂️
English
0
0
1
102
Ramakanth Kavuluru retweetledi
Mayee Chen
Mayee Chen@MayeeChen·
Data mixing - determining ratios across your training datasets - matters a lot for model quality. While building Olmo 3, we learned it’s hard to set up a method that finds a strong mix, and hard to maintain that mix as datasets change throughout development. Introducing Olmix👇
Mayee Chen tweet media
English
13
70
261
46.7K
Ramakanth Kavuluru retweetledi
Hyunwoo Kim
Hyunwoo Kim@hyunw_kim·
🚨New paper to level up your 🦞#Clawdbot ?! Bots are now posting your sensitive info in real time. But privacy research is a desert with no data to train better models. That's about to change Enter 🏝️Privasis, the oasis where you can train strong privacy-forward AI with scale✨
Hyunwoo Kim tweet media
English
3
18
91
21.5K