Adyasha Maharana

649 posts

Adyasha Maharana

Adyasha Maharana

@adyasha10

Research Scientist @DbrxMosaicAI. Data curation, vision+language, AI+health. Previously @unccs PhD, @Snap, @allen_ai, @AdobeResearch, @IHME_UW

Chapel Hill, NC Katılım Temmuz 2012
713 Takip Edilen724 Takipçiler
Sabitlenmiş Tweet
Adyasha Maharana
Adyasha Maharana@adyasha10·
Our code for generating extremely long multi-session conversations with custom personas and evaluating LLMs on the LoCoMo dataset is now live at github.com/snap-research/…! Come talk to us about this work at Poster session 5 (Aug 13, 16:00 - 17:30 local time) at ACL 2024
Adyasha Maharana@adyasha10

Can LLMs keep track of very long conversations? We evaluate 'conversational memory' of LLMs via 3 tasks on our dataset of multi-session multimodal dialogs --> LLMs struggle to remember, reason over history, draw long-range temporal/causal connections arxiv.org/abs/2402.17753 🧵

English
1
12
32
4.2K
Adyasha Maharana retweetledi
Samip
Samip@industriaalist·
Introducing Q Labs, a research lab focused on solving generalization. Alongside others (SSI, Flapping Airplanes), we see data efficiency as the key problem, but we're taking an unconventional approach to solve it: a new learning algorithm approximating Solomonoff induction.
English
36
56
680
160.3K
Adyasha Maharana retweetledi
Jonathan Frankle
Jonathan Frankle@jefrankle·
Special Databricks swag for the first five people to send me a selfie with Ashu in the Databricks booth at NeurIPS!
Ashutosh Baheti@abaheti95

Will be at #NeurIPS2025 from 2nd to 6th Dec. Excited to chat about async RL, Environment Exploration, Agents/Tool use, User Simulator, Synthetic Data Generation or any other topic!! You can find me at the @databricks booth @ Tue 12 - 4pm

English
5
3
39
20.6K
Adyasha Maharana retweetledi
Mohit Bansal
Mohit Bansal@mohitban47·
🚨 🤯 Wow! Yi Lin is an amazing researcher, who works on very hard and important problems in LLM and VLM training, RL, PEFT, Quantization, etc. -- ironically, he had several other top offers just a few months ago! Hire him ASAP if you want to pick up a top talent (and several other affected amazing folks)! 👇👇
Yi Lin Sung@yilin_sung

Tough week! I also got impacted less than 3 months after joining. Ironically, I just landed some new RL infra features the day before. Life moves on. My past work spans RL, PEFT, Quantization, and Multimodal LLMs. If your team is working on these areas, I’d love to connect.

English
5
31
159
45.9K
Adyasha Maharana retweetledi
Yi Lin Sung
Yi Lin Sung@yilin_sung·
Tough week! I also got impacted less than 3 months after joining. Ironically, I just landed some new RL infra features the day before. Life moves on. My past work spans RL, PEFT, Quantization, and Multimodal LLMs. If your team is working on these areas, I’d love to connect.
Jiaxun Cui 🐿️@cuijiaxun

Meta has gone crazy on the squid game! Many new PhD NGs are deactivated today (I am also impacted🥲 happy to chat)

English
42
64
498
173.8K
Adyasha Maharana retweetledi
Awni Hannun
Awni Hannun@awnihannun·
I always thought the decline in fundamental AI research funding would happen because AI didn’t generate enough value to be worth the cost. But it seems like it’s happening because it generated too much value. And the race to capture that value is taking priority. Just remembering that a lot of this started in curiosity driven industry research labs.
English
32
36
363
71.3K
Adyasha Maharana retweetledi
Physical Intelligence
Physical Intelligence@physical_int·
We've added pi-05 to the openpi repo: pi05-base, pi05-droid, pi05-libero. Also added PyTorch training code!🔥 Instructions and code here: github.com/Physical-Intel… This is an updated version of the model we showed cleaning kitchens and bedrooms in April: pi.website/blog/pi05
English
23
132
875
367.6K
Adyasha Maharana retweetledi
Alex Trott
Alex Trott@alexrtrott·
Ever wonder what it'd look like if an LLM Judge and a Reward Model had a baby? So did we, which is why we created PGRM -- the Prompt-Guided Reward Model. TLDR: You get the instructability of an LLM judge + the calibration of an RM in a single speedy package (1/n)
Alex Trott tweet media
English
6
25
155
30.7K
Prateek Yadav
Prateek Yadav@prateeky2806·
I've officially joined Meta Superintelligence Labs (MSL) org in the Bay Area. I'll be working on critical aspects of pre-training, synthetic data and RL for the next generation of models. Humbled and eager to contribute to the quest for superintelligence. @AIatMeta
English
36
27
812
80K
Adyasha Maharana retweetledi
Jonathan Frankle
Jonathan Frankle@jefrankle·
I'm at ICML 🇨🇦 and I'm hiring at @databricks. Visit our booth if you're interested. My scientific focus: It's 1972 in AI, there's an AI crisis, Dijkstra isn't here to save us, and maybe RL can. Why Databricks? The long road to AGI is being paved here and we have the real evals 🧵
English
9
24
224
41.9K
Adyasha Maharana retweetledi
David Fan
David Fan@DavidJFan·
Can visual SSL match CLIP on VQA? Yes! We show with controlled experiments that visual SSL can be competitive even on OCR/Chart VQA, as demonstrated by our new Web-SSL model family (1B-7B params) which is trained purely on web images – without any language supervision.
David Fan tweet media
English
12
96
463
86.1K
Adyasha Maharana retweetledi
Gedas Bertasius
Gedas Bertasius@gberta227·
For those of you who know me, I've always been very excited to combine my two passions for basketball and CV. Our #CVPR2025 paper does this by introducing a large-scale video dataset for fine-grained skill estimation in 🏀. Paper, code & data available: sites.google.com/cs.unc.edu/bas…
Yulu Pan@YuluPan_00

🚨 New #CVPR2025 Paper 🚨 🏀BASKET: A Large-Scale Dataset for Fine-Grained Basketball Skill Estimation🎥 4,477 hours of videos⏱️ | 32,232 players⛹️ | 20 fine-grained skills🎯 We present a new video dataset for skill estimation with unprecedented scale and diversity! A thread👇

English
0
13
69
8.5K
Adyasha Maharana retweetledi
Jonathan Frankle
Jonathan Frankle@jefrankle·
The hardest part about finetuning LLMs is that people generally don't have high-quality labeled data. Today, @databricks introduced TAO, a new finetuning method that only needs inputs, no labels necessary. Best of all, it actually beats supervised finetuning on labeled data.
Jonathan Frankle tweet media
English
13
136
894
90.6K
Adyasha Maharana retweetledi
Ryan Marten
Ryan Marten@ryanmart3n·
Announcing the Open Thoughts project. We are building the best reasoning datasets out in the open. Building off our work with Stratos, today we are releasing OpenThoughts-114k and OpenThinker-7B.
Ryan Marten tweet media
English
14
60
372
38.5K
Adyasha Maharana retweetledi
Javier Rando
Javier Rando@javirandor·
Carlini’s website will be auto-generated daily by a different LLM for the next 12 days. We joked about this during a dinner after NeurIPS and Christmas made it happen 💫🎄 nicholas.carlini.com/writing/2025/l…
English
5
7
76
18.1K
Adyasha Maharana retweetledi
Jonathan Frankle
Jonathan Frankle@jefrankle·
Excited for our Data and AI Eras Tour at the @databricks booth at NeurIPS!
Jonathan Frankle tweet mediaJonathan Frankle tweet media
English
4
4
95
6.4K
Adyasha Maharana retweetledi
Zack Ankner
Zack Ankner@ZackAnkner·
There have been a lot of anectodes about the Llama3 series of models being harder to post-training quanitze (PTQ) than Llama2. As part of this paper, we investigated the hypothesis that the degradation from PTQ grows with the token-to-parameter ratio (TPR), .ie as you overtrain.
Zack Ankner tweet media
Tanishq Kumar@tanishqkumar07

[1/7] New paper alert! Heard about the BitNet hype or that Llama-3 is harder to quantize? Our new work studies both! We formulate scaling laws for precision, across both pre and post-training arxiv.org/pdf/2411.04330. TLDR; - Models become harder to post-train quantize as they are overtrained on lots of data, so that eventually more pretraining data can be actively harmful if quantizing post-training! - The effects of putting weights, activations, or attention in varying precisions during pretraining are consistent and predictable, and fitting a scaling law suggests that pretraining at high (BF16) and next-generation (FP4) precisions may both be suboptimal design choices! Joint work with @ZackAnkner @bfspector @blake__bordelon @Muennighoff @mansiege @CPehlevan @HazyResearch @AdtRaghunathan.

English
2
10
87
11.4K
Adyasha Maharana
Adyasha Maharana@adyasha10·
@anikembhavi I learnt so much from you as an intern, Ani. Always looking forward to the next wonderful thing that comes from your leadership! Wish you the very best :)
English
1
0
1
426
Ani Kembhavi
Ani Kembhavi@anikembhavi·
After an incredible 10 years at AI2, today marks my last day. I joined when we were a tiny company with just two dozen people. It has been very exciting to watch AI2 grow into a major force in AI and it's been an honor to contribute to that journey. Throughout this period of growth and success, we have tackled a wide range of challenging problems in AI with a consistent ability to punch above our weight, all while maintaining a collaborative and friendly culture in a truly open research environment. The PRIOR team at AI2 has become my second family. You have been brilliant, hardworking, relentless year after year and been incredibly supportive of me, and I cannot thank you enough for everything. Saying goodbye has been tough, but I’m eagerly looking forward to the next wayve of adventures coming my way.
English
27
3
281
52.7K