Kate Baumli

714 posts

Kate Baumli banner
Kate Baumli

Kate Baumli

@katebaumli

Research Engineer @GoogleDeepMind👩‍💻 Texas Ex @UTAustin🤘 Interested in intelligence, understanding, science, and mindfulness Views my own

London, England เข้าร่วม Aralık 2019
731 กำลังติดตาม1.2K ผู้ติดตาม
ทวีตที่ปักหมุด
Kate Baumli
Kate Baumli@katebaumli·
Check out our new paper “Human-Timescale Adaptation in an Open-Ended Task Space” 📄 arxiv.org/abs/2301.07608 So happy to share what we’ve been up to the last year. I really loved working on this project to create adaptive agents with these amazing collaborators💛
Feryal@FeryalMP

I’m super excited to share our work on AdA: An Adaptive Agent capable of hypothesis-driven exploration which solves challenging unseen tasks with just a handful of experience, at a similar timescale to humans. sites.google.com/corp/view/adap… See the thread for more details 👇 [1/N]

English
2
0
34
4.9K
Ted Moskovitz
Ted Moskovitz@ted_moskovitz·
I joined the multimodal team at @AnthropicAI this week—really excited to build some cool stuff!
English
18
2
212
33.9K
Kate Baumli รีทวีตแล้ว
Avishkar
Avishkar@avishkar58·
Really proud to work with this team of awesome people on the mission to create a tutor for every learner and an assistant for every teacher! All done with participation and responsibility at its core.
Google@Google

Introducing LearnLM: our new family of models based on Gemini and fine-tuned for learning. LearnLM applies educational research to make our products — like Search, Gemini and YouTube — more personal, active and engaging for learners. #GoogleIO

English
1
1
10
1.2K
Kate Baumli รีทวีตแล้ว
Google DeepMind
Google DeepMind@GoogleDeepMind·
We’re sharing Project Astra: our new project focused on building a future AI assistant that can be truly helpful in everyday life. 🤝 Watch it in action, with two parts - each was captured in a single take, in real time. ↓ #GoogleIO
English
211
953
3.7K
1.4M
Kate Baumli รีทวีตแล้ว
📎
📎@mikesnosense·
im like the opposite of machine learning. human forgetting
English
103
29.7K
124.5K
3.6M
Kate Baumli รีทวีตแล้ว
AK
AK@_akhaliq·
Google presents Genie Generative Interactive Environments introduce Genie, the first generative interactive environment trained in an unsupervised manner from unlabelled Internet videos. The model can be prompted to generate an endless variety of action-controllable virtual worlds described through text, synthetic images, photographs, and even sketches. At 11B parameters, Genie can be considered a foundation world model. It is comprised of a spatiotemporal video tokenizer, an autoregressive dynamics model, and a simple and scalable latent action model. Genie enables users to act in the generated environments on a frame-by-frame basis despite training without any ground-truth action labels or other domain-specific requirements typically found in the world model literature. Further the resulting learned latent action space facilitates training agents to imitate behaviors from unseen videos, opening the path for training generalist agents of the future.
English
78
501
2.3K
684.1K
Kate Baumli รีทวีตแล้ว
Tim Rocktäschel
Tim Rocktäschel@_rockt·
I am really excited to reveal what @GoogleDeepMind's Open Endedness Team has been up to 🚀. We introduce Genie 🧞, a foundation world model trained exclusively from Internet videos that can generate an endless variety of action-controllable 2D worlds given image prompts.
GIF
English
133
544
2.4K
832.7K
Kate Baumli รีทวีตแล้ว
Demis Hassabis
Demis Hassabis@demishassabis·
We have a long history of supporting responsible open source & science, which can drive rapid research progress, so we’re proud to release Gemma: a set of lightweight open models, best-in-class for their size, inspired by the same tech used for Gemini blog.google/technology/dev…
Demis Hassabis tweet media
English
170
346
1.9K
465.4K
Kate Baumli รีทวีตแล้ว
Jeff Dean
Jeff Dean@JeffDean·
Gemini 1.5 Pro - A highly capable multimodal model with a 10M token context length Today we are releasing the first demonstrations of the capabilities of the Gemini 1.5 series, with the Gemini 1.5 Pro model. One of the key differentiators of this model is its incredibly long context capabilities, supporting millions of tokens of multimodal input. The multimodal capabilities of the model means you can interact in sophisticated ways with entire books, very long document collections, codebases of hundreds of thousands of lines across hundreds of files, full movies, entire podcast series, and more. Gemini 1.5 was built by an amazing team of people from @GoogleDeepMind, @GoogleResearch, and elsewhere at @Google. @OriolVinyals (my co-technical lead for the project) and I are incredibly proud of the whole team, and we’re so excited to be sharing this work and what long context and in-context learning can mean for you today! There’s lots of material about this, some of which are linked to below. Main blog post: blog.google/technology/ai/… Technical report: “Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context” goo.gle/GeminiV1-5 Videos of interactions with the model that highlight its long context abilities: Understanding the three.js codebase: youtube.com/watch?v=SSnsmq… Analyzing a 45 minute Buster Keaton movie: youtube.com/watch?v=wa0MT8… Apollo 11 transcript interaction: youtube.com/watch?v=LHKL_2… Starting today, we’re offering a limited preview of 1.5 Pro to developers and enterprise customers via AI Studio and Vertex AI. Read more about this on these blogs: Google for Developers blog: developers.googleblog.com/2024/02/gemini… Google Cloud blog: cloud.google.com/blog/products/… We’ll also introduce 1.5 Pro with a standard 128,000 token context window when the model is ready for a wider release. Coming soon, we plan to introduce pricing tiers that start at the standard 128,000 context window and scale up to 1 million tokens, as we improve the model. Early testers can try the 1 million token context window at no cost during the testing period. We’re excited to see what developer’s creativity unlocks with a very long context window. Let me walk you through the capabilities of the model and what I’m excited about!
YouTube video
YouTube
YouTube video
YouTube
YouTube video
YouTube
Jeff Dean tweet media
English
184
1.1K
6K
1.7M
Kate Baumli รีทวีตแล้ว
Tommy Collison
Tommy Collison@tommycollison·
What have you been reading this weekend? 📚
English
48
3
23
20.2K
brian🧃
brian🧃@b___brian·
I’m always in awe of the way my girlfriend clutches the next 4 pages of the book she is reading
brian🧃 tweet media
English
85
436
25.2K
544.7K
Kate Baumli รีทวีตแล้ว
Oriol Vinyals
Oriol Vinyals@OriolVinyalsML·
Exciting times, welcome Gemini (and MMLU>90)! State-of-the-art on 30 out of 32 benchmarks across text, coding, audio, images, and video, with a single model 🤯 Co-leading Gemini has been my most exciting endeavor, fueled by a very ambitious goal. And that is just the beginning! A long 🐍 post about our Gemini journey & state of the field. The biggest challenges in LLMs are far from trivial or obvious. Evaluation and data stand out to me. We've moved beyond the simpler "Have we won in Go/Chess/StarCraft?" to “Is this answer accurate and fair? Is this conversation good? Does this complex piece of text prove the theorem?” Exciting potential coupled with monumental challenges. The field is less ripe further down the model pipeline. Pretraining is relatively well understood. Instruction tuning and RLHF, less so. In AlphaGo and AlphaStar we spent 5% of compute in pre-training and the rest in the very important RL phase, where the model learns from its successes or failures. In LLMs, we spend most of our time on pretraining. I believe there’s huge potential to be untapped. Cakes with lots of cherries, please 🎂 @Google has demonstrated its ability to move fast. It has been an absolute blast to see the energy from my colleagues and the support received. A “random” highlight is coauthoring our tech report with a co-founder. Another is coleading with @JeffDean. But beyond individuals, Gemini is about teamwork: it is important to recognize the collective effort behind such achievements. Picture a room full of brilliant people, and avoid attributing success solely to one person. On a personal note, recently I celebrated my 10 year anniversary at Google, and it’s been 8 years since @quocleix and I co-authored “A Neural Conversational Model”, which gave us a glimpse of what was, has, and is yet to come. Back then, that line of work received a lot of skepticism. Lessons learned: whatever your passion is, push for it! Zooming back out, there’s lots of change in our field, and the stakes couldn’t be higher. Excited for what’s to come from Gemini, but humbled by the responsibility to “get it right”. 2024 will be drastic. Welcome Gemini! blog.google/technology/ai/…
GIF
English
63
267
2K
456.2K
Kate Baumli
Kate Baumli@katebaumli·
Gemini 1.0 is here! Very grateful to have been one small part of the massive team working towards advancing the state of the art for generally capable models. Here's to a bright future, to continuing to advance intelligence, in the quest towards AGI. #introductionvideo" target="_blank" rel="nofollow noopener">deepmind.google/technologies/g…
English
0
1
13
896
Kate Baumli รีทวีตแล้ว
Google DeepMind
Google DeepMind@GoogleDeepMind·
We’re excited to announce 𝗚𝗲𝗺𝗶𝗻𝗶: @Google’s largest and most capable AI model. Built to be natively multimodal, it can understand and operate across text, code, audio, image and video - and achieves state-of-the-art performance across many tasks. 🧵 dpmd.ai/announcing-gem…
English
162
1.5K
5.8K
1.3M
Kate Baumli รีทวีตแล้ว
Sundar Pichai
Sundar Pichai@sundarpichai·
Introducing Gemini 1.0, our most capable and general AI model yet. Built natively to be multimodal, it’s the first step in our Gemini-era of models. Gemini is optimized in three sizes - Ultra, Pro, and Nano Gemini Ultra’s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks. With a score of 90.0%, Gemini Ultra is the first model to outperform human experts on MMLU. blog.google/technology/ai/…
Sundar Pichai tweet media
English
930
3.6K
22.6K
5M
Kate Baumli รีทวีตแล้ว
Demis Hassabis
Demis Hassabis@demishassabis·
The Gemini era is here. Thrilled to launch Gemini 1.0, our most capable & general AI model. Built to be natively multimodal, it can understand many types of info. Efficient & flexible, it comes in 3 sizes each best-in-class & optimized for different uses blog.google/technology/ai/…
Demis Hassabis tweet media
English
387
1.9K
10.8K
3.2M
Kate Baumli รีทวีตแล้ว
Tom Zahavy
Tom Zahavy@TZahavy·
I'm super excited to share AlphaZeroᵈᵇ, a team of diverse #AlphaZero agents that collaborate to solve #Chess puzzles and demonstrate increased creativity. Check out our paper to learn more! arxiv.org/abs/2308.09175 A quick 🧵(1/n)
Tom Zahavy tweet media
English
4
66
326
94.5K