Timo Schick

194 posts

Timo Schick banner
Timo Schick

Timo Schick

@timo_schick

NLP Researcher

Munich Katılım Şubat 2013
199 Takip Edilen2.9K Takipçiler
Sabitlenmiş Tweet
Timo Schick
Timo Schick@timo_schick·
🎉 New paper 🎉 Introducing the Toolformer, a language model that teaches itself to use various tools in a self-supervised way. This significantly improves zero-shot performance and enables it to outperform much larger models. 🧰 🔗 Link: arxiv.org/abs/2302.04761
GIF
English
41
284
1.4K
509.4K
Timo Schick
Timo Schick@timo_schick·
Looking forward to my first non-virtual conference in almost 2 years. If you’re attending #ICML2024 and want to chat, drop me a message 😊
English
1
0
25
2.8K
Timo Schick retweetledi
Roberta Raileanu
Roberta Raileanu@robertarail·
🤖 Want an agent that can learn new tasks from only a handful of demonstrations and no weight updates? 🚀 Check out our new work on In-Context Learning for Sequential Decision-Making, where we show how we can use transformers to few-shot learn new Procgen and MiniHack tasks. 👋 If you want to learn more about it, come chat with us at the FMDM workshop @NeurIPSConf on Friday, December 15. 🙌 Kudos to @sharathraparthy who did an outstanding job leading this work, designing and running lots of experiments, and digging deep trying to understand the model’s behavior. 🧵👇
Sharath Raparthy@sharathraparthy

🚨 🚨 !!New Paper Alert!! 🚨 🚨 How can we train agents that learn new tasks (with different states, actions, dynamics and reward functions) from only a few demonstrations and no weight updates? In-context learning to the rescue! In our new paper, we show that by training transformers on large diverse datasets of sequences of demonstrations with certain properties, we can generalize to new Procgen or MiniHack tasks from only a few demonstrations and no weight updates! Paper: arxiv.org/pdf/2312.03801… Work with these amazing collaborators @erichammy @_roberkirk @HenaffMikael @robertarail 1/13

English
0
11
70
11K
Timo Schick retweetledi
Rowan Cheung
Rowan Cheung@rowancheung·
Inflection AI just announced Inflection-2, a HUGE new 175 billion parameter language model. Capabilities exceed Google and Meta's top models and “is very close” to catching GPT-4. The CEO also said the company’s next model will be 10x larger in six months.
Rowan Cheung tweet media
English
9
61
743
272.4K
Timo Schick retweetledi
Mustafa Suleyman
Mustafa Suleyman@mustafasuleyman·
Thrilled to announce that Inflection-2 is now the 2nd best LLM in the world! 💚✨🎉 It will be powering Pi.ai very soon. And available to select API partners in time. Tech report linked... Come run with us! inflection.ai/inflection-2
English
73
111
1K
546.7K
Timo Schick retweetledi
Inflection AI
Inflection AI@inflectionAI·
🎉 Introducing Inflection-2, the 2nd best LLM in the world! Get ready to experience the future of AI with us. bit.ly/3TaUpcD
English
49
166
891
314K
Timo Schick retweetledi
Mustafa Suleyman
Mustafa Suleyman@mustafasuleyman·
Utterly insane weekend. So sad. Wishing everyone involved the very best. In the meantime, we finished training Inflection-2 last night! ✨ It's now the 2nd best LLM in the world... & we're scaling MUCH further. Details v soon. Come run with us!
English
69
191
1.3K
375.7K
Timo Schick retweetledi
Pi
Pi@pi·
In just over 100 days since launching Pi, we’ve just hit one billion messages exchanged. A huge milestone 🤯 Any predictions on how long it will take us to get to 2 billion?!
GIF
English
14
6
100
8.4K
Timo Schick retweetledi
Jason Weston
Jason Weston@jaseweston·
🚨New Paper 🚨 Self-Alignment with Instruction Backtranslation - New method auto-labels web text with instructions & curates high quality ones for FTing - Our model Humpback 🐋 outperforms LIMA, Claude, Guanaco, davinci-003 & Falcon-Inst arxiv.org/abs/2308.06259 (1/4)🧵
Jason Weston tweet media
English
13
138
653
357.5K
Timo Schick retweetledi
Mustafa Suleyman
Mustafa Suleyman@mustafasuleyman·
Excited to announce that we’ve raised $1.3B to build one of the largest clusters in the world and turbocharge the creation of Pi, your personal AI. forbes.com/sites/alexkonr…
English
142
312
2.6K
928.7K
Timo Schick retweetledi
Inflection AI
Inflection AI@inflectionAI·
We’re proud to announce Inflection-1, the best-in-class LLM developed at Inflection! Inflection-1, which powers Pi.ai, outperforms GPT-3.5, Chinchilla, and LLaMA on a number of academic benchmarks. More details in our technical memo: inflection.ai/inflection-1
English
19
86
370
163.3K
Timo Schick retweetledi
Manoel
Manoel@manoelribeiro·
One of our key sources of human data is no longer fully “human"! We estimate that 33-46% of crowd workers on MTurk used large language models (LLMs) in a text production task - which may increase as ChatGPT and the like become more popular and powerful. arxiv.org/abs/2306.07899
Manoel tweet media
English
42
517
1.8K
857.8K
Timo Schick retweetledi
AI Pub
AI Pub@ai__pub·
// Deep Papers #3: Toolformer // LLMs like Bing and ChatGPT use external tools like calculators and web search to answer questions. How do you teach LLMs to *use* these external tools? Toolformer shows how! We interviewed the authors :) Spotify: open.spotify.com/episode/6uXohG…
AI Pub tweet mediaAI Pub tweet mediaAI Pub tweet mediaAI Pub tweet media
English
4
46
355
131.5K
Timo Schick retweetledi
AI Pub
AI Pub@ai__pub·
// Toolformer Podcast: Preview // Today I'm interviewing the Toolformer authors! LLMs like Bing (and soon, ChatGPT) can use external tools like calculators or internet search to answer questions. But how do language models *learn to use* these tools? 1/5
AI Pub tweet mediaAI Pub tweet media
English
7
72
441
170.5K
Timo Schick retweetledi
Victor Sanh
Victor Sanh@SanhEstPasMoi·
We are reproducing Flamingo, a vision and language model developed by Deepmind (arxiv.org/abs/2204.14198). We spent a good amount of time fighting training divergences (aka "instabilities"). Surprisingly, even at the ~2-3B scale. Some learnings from overcoming these 🧵:
Victor Sanh tweet media
English
26
236
1.4K
303.4K