Minae Kwon

117 posts

Minae Kwon banner
Minae Kwon

Minae Kwon

@MinaeKwon

Katılım Aralık 2018
553 Takip Edilen1.1K Takipçiler
Minae Kwon retweetledi
Anthropic
Anthropic@AnthropicAI·
Introducing the Anthropic Science Blog. Increasing the pace of scientific progress is a core part of Anthropic’s mission. The Science Blog will feature new research and stories of how scientists are using AI to accelerate their work. Read the intro: anthropic.com/research/intro…
English
208
638
5.1K
423.9K
Minae Kwon retweetledi
Anthropic
Anthropic@AnthropicAI·
We’re publishing a new constitution for Claude. The constitution is a detailed description of our vision for Claude’s behavior and values. It’s written primarily for Claude, and used directly in our training process. anthropic.com/news/claude-ne…
English
523
970
7.7K
3.3M
Siddharth Karamcheti
Siddharth Karamcheti@siddkaramcheti·
Thrilled to share that I'll be starting as an Assistant Professor at Georgia Tech (@ICatGT / @GTrobotics / @mlatgt) in Fall 2026. My lab will tackle problems in robot learning, multimodal ML, and interaction. I'm recruiting PhD students this next cycle – please apply/reach out!
Siddharth Karamcheti tweet mediaSiddharth Karamcheti tweet media
English
72
23
563
60.9K
Minae Kwon retweetledi
Anthropic
Anthropic@AnthropicAI·
Introducing an upgraded Claude 3.5 Sonnet, and a new model, Claude 3.5 Haiku. We’re also introducing a new capability in beta: computer use. Developers can now direct Claude to use computers the way people do—by looking at a screen, moving a cursor, clicking, and typing text.
Anthropic tweet media
English
472
1.8K
10K
3.7M
Minae Kwon retweetledi
Anthropic
Anthropic@AnthropicAI·
Introducing Claude 3.5 Sonnet—our most intelligent model yet. This is the first release in our 3.5 model family. Sonnet now outperforms competitor models on key evaluations, at twice the speed of Claude 3 Opus and one-fifth the cost. Try it for free: claude.ai
Anthropic tweet media
English
421
1.5K
7.1K
2.5M
Minae Kwon retweetledi
Dorsa Sadigh
Dorsa Sadigh@DorsaSadigh·
At #ICRA24 we've a few papers on 𝗴𝗿𝗼𝘂𝗻𝗱𝗲𝗱 𝗿𝗲𝗮𝘀𝗼𝗻𝗶𝗻𝗴 of LLMs/VLMs. • Grounded common-sense reasoning via active perception - @MinaeKwon's 🧵👇 • Physically grounding VLMs - @jensen_gao's 🧵👇 • Learning from online language corrections - @lihanzha's 🧵👇
English
1
5
82
18.9K
Minae Kwon retweetledi
Alex Tamkin
Alex Tamkin@AlexTamkin·
Made a short video exploring tool use and subagents! (w/ @aaron_begg and @typochondriac) Goal: Find the “quickest quicksort” implementation on GitHub by having a larger model orchestrate 100 subagent models Here’s how it works: 1/ twitter.com/AnthropicAI/st…
Anthropic@AnthropicAI

Tool use is now available in beta to all customers in the Anthropic Messages API, enabling Claude to interact with external tools using structured outputs.

English
1
6
90
18.4K
Minae Kwon retweetledi
Jesse Mu
Jesse Mu@jayelmnop·
We’re hiring for the adversarial robustness team @AnthropicAI! As an Alignment subteam, we're making a big effort on red-teaming, test-time monitoring, and adversarial training. If you’re interested in these areas, let us know! (emails in 🧵)
Jesse Mu tweet media
English
4
70
446
72.6K
Minae Kwon retweetledi
Minae Kwon retweetledi
Hugh Zhang
Hugh Zhang@hughbzhang·
I also have no idea what Q* is, but given speculation that it’s a method of self-learning and Monte-Carlo Tree Search (MCTS) in language models, I thought I’d share some recent work on an adjacent idea.
Hugh Zhang tweet media
English
10
37
400
87.2K
Minae Kwon retweetledi
Alex Tamkin
Alex Tamkin@AlexTamkin·
Eliciting Human Preferences with Language Models Currently, people write detailed prompts to describe what they want a language model to do We explore *generative elicitation*—where models interactively ask for this information through open-ended conversation 1/
Alex Tamkin tweet media
English
4
84
447
82.9K
Minae Kwon retweetledi
Andy Shih
Andy Shih@andyshih_·
Excited about recent improvements of our NeurIPS Spotlight paper, now even faster with ⚡️⚡️multiprocessing⚡️⚡️! We now get 2x speedup on as low as 50-step DDIM, and 4x speedup on 200-step DDIM! The first version of our paper showed good results, but we wanted even better. - torch.DataParallel showed big gains for DDPM, but not for few-step DDIM (due to python GIL and DP's parameter copying) - torch.DDP does not work out of the box for our algorithm To address this, we implemented custom multiprocessing logic with a producer/consumer design. This resolves python GIL, and keeps parameters persistent on GPUs to avoid repeated copying. Our multiprocessing implementation gave 5x improvement over torch.DataParallel for 50-step sampling 🥳🥳 Excited to present more about this work at NeurIPS! Try it out now at: github.com/AndyShih12/par…
Andy Shih@andyshih_

Diffusion models are slow to sample from. Many methods propose to sample using *fewer* steps, but this can hurt sample quality. We introduce ParaDiGMS, a new method to speed up diffusion models by 2-4x while using the *same* number of steps! arxiv.org/abs/2305.16317

English
1
11
71
18.4K
Minae Kwon retweetledi
shreya rajpal
shreya rajpal@ShreyaR·
It's an absolute honor to be a guest on the @twimlai podcast! @samcharrington and I cover everything under the sun in LLMOps, from hallucinations, RAG to LLM safety. Check out the podcast on the link below!
The TWIML AI Podcast@twimlai

Today we’re joined by @ShreyaR to discuss LLM safety for production applications! We explore hallucinations, RAG, evaluation & tooling for LLMs as well as Guardrails, an open-source project enforcing correctness in LMs. 🎧/🎥 Check out the episode at twimlai.com/go/647

English
2
6
59
15K
Minae Kwon retweetledi
Sang Michael Xie
Sang Michael Xie@sangmichaelxie·
Releasing an open-source PyTorch implementation of DoReMi! github.com/sangmichaelxie… The pretraining data mixture is a secret sauce of LLM training. Optimizing your data mixture for robust learning with DoReMi can reduce training time by 2-3x. Train smarter, not longer!
Sang Michael Xie tweet media
Sang Michael Xie@sangmichaelxie

Should LMs train on more books, news, or web data? Introducing DoReMi🎶, which optimizes the data mixture with a small 280M model. Our data mixture makes 8B Pile models train 2.6x faster, get +6.5% few-shot acc, and get lower pplx on *all* domains! 🧵⬇️ arxiv.org/abs/2305.10429

English
2
65
262
68.2K
Minae Kwon retweetledi
Yuchen Cui
Yuchen Cui@YuchenCui1·
We use gestures all the time for specifying targets! How can robots make sense of “gimme that one”? We propose GIRAF, a framework for interpreting human gesture instructions using LLMs. Paper to appear in @corl_conf: arxiv.org/abs/2309.02721 Website: tinyurl.com/giraf23
English
7
15
82
35.7K