Pratiksha Thaker

25 posts

Pratiksha Thaker

Pratiksha Thaker

@prthaker_

Research @databricks, recently postdoc @mldcmu. tweets are my own

Beigetreten Ekim 2024
32 Folgt44 Follower
Pratiksha Thaker retweetet
kanyes
kanyes@KanyesThaker·
Launched Hyper on ProductHunt because we hate taking notes and we hate scheduling meetings. The greatest work happens huddled around a whiteboard or standing in the kitchen. Check it out! producthunt.com/products/hyper…
English
2
2
5
8.9K
Pratiksha Thaker retweetet
Jonathan Frankle
Jonathan Frankle@jefrankle·
Meet KARL, an RL'd model for document-centric tasks at frontier quality and open source cost/speed. Great for @databricks customers and scientists (77-page tech report!) As usual, this isn't just one model - it's an RL assembly line to churn out models for us and our customers 🧵
Jonathan Frankle tweet mediaJonathan Frankle tweet media
English
9
46
241
69.2K
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
I'm so excited that this work is available after a year of carefully curating open problems with our collaborators. It was inspired by real issues we faced applying research techniques to problems in child safety, and we hope this work can help amplify those lessons.
Neil Kale@neilkale

[1/n] Open Problems in AI Child Safety AI is misused to generate CSAM at alarming scale. 400% increase in AI-generated CSAM since 2024 (IWF). 1 in 17 teens are victimized by deepfake nudes. We outline 15 open problems where AI safety research can help. 🔗aichildsafety.github.io

English
0
1
6
1.7K
Pratiksha Thaker retweetet
ML@CMU
ML@CMU@mlcmublog·
We asked LLMs: Is Santa real? 🎅 GPT-4o says Yes at any age. Claude tells 5-year-olds the truth. What does this reveal about invisible assumptions in AI? Do LLMs believe in the tooth fairy or the Illuminati? New holiday post here: blog.ml.cmu.edu/2025/12/23/is-…
English
0
3
5
770
Pratiksha Thaker retweetet
Jonathan Frankle
Jonathan Frankle@jefrankle·
I'm hiring interns for next summer at @databricks! Specifically on (1) empirical RL at scale on non-verifiable tasks and (2) enabling real people specify the behaviors they want out of AI (e.g., through evals) on highly complex tasks. 🧵
English
17
47
524
92.3K
Pratiksha Thaker retweetet
Roy Rinberg
Roy Rinberg@RoyRinberg·
Prospective PhD students interested privacy research - here's a google sheet with professors you may be interested in applying to! Feel encouraged to suggest edits, and share openly! (link in thread because twitter doesn't like links in tweets🤷)
Roy Rinberg tweet media
English
5
18
66
33K
Pratiksha Thaker retweetet
Steven Kolawole
Steven Kolawole@_stevenkolawole·
🧵THREAD Can we automatically identify parallelizable structure within LLM queries for massive efficiency gains? 10% of real prompts are parallelizable-> ~5x speedups w/ >90% quality preserved. For ChatGPT's 1B+ queries: 100M+ optimization opportunities. Full suite below🧵
Steven Kolawole tweet media
English
2
18
57
14.2K
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
I'm very excited about this work and thinking about more realistic data access models for MIAs. It wouldn't have been possible without @neilkale @gingsmith @zstevenwu and our amazing collaborators at @thorn!
English
0
0
0
191
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
🕵️‍♀️These results have important implications for domains like child safety, where it's critical to detect harmful content in training data, but auditors legally can't access this content to train attack models. There's still more work to be done, but this is a key first step.
English
1
0
0
219
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
I'm very excited to share some new work arxiv.org/abs/2506.06488. This work started out in conversations with @thorn where we realized that shadow model MIAs couldn't be used to audit models for harmful content of children. See 🧵 for why, and our progress on solving this...
English
1
7
25
4.3K
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
(And many thanks to @neilkale for helping me draft my first thread 😄)
English
0
0
1
164
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
🚨 Are you using empirical benchmarks to evaluate your LLM unlearning method? Our new paper arxiv.org/pdf/2410.02879 investigates how success on these benchmarks can be misleading. A🧵: 1/n
English
1
5
13
2.2K