Google AI

3.2K posts

Google AI banner
Google AI

Google AI

@GoogleAI

Making AI helpful for everyone. Show thinking ↓

Mountain View, CA เข้าร่วม Nisan 2009
30 กำลังติดตาม2.4M ผู้ติดตาม
ทวีตที่ปักหมุด
Google AI
Google AI@GoogleAI·
Today we launched Gemini 3.1 Flash TTS, our most expressive and controllable text-to-speech model yet. This launch [excitement] includes audio tags! 🗣🏷 Audio tags [explanatory] are a seamless way to guide vocal style, pace, and delivery using natural language commands embedded directly in your text. Want a different tempo or tone? [amazement] Just tag the audio to steer the AI-speech output! The model supports 70+ languages (24 of which are high-quality evaluated languages, including: Japanese, Hindi, and Arabic). Watch the audio tags in action in the demo below ↓
English
95
296
2.2K
180.6K
Google AI
Google AI@GoogleAI·
What a week! Here’s everything we shipped: — Gemini 3.1 Flash TTS, our latest text-to-speech model, featuring native multi-speaker dialogue and improved controllability and audio tags for more natural, expressive voices in 70+ languages — Gemini Robotics-ER 1.6 by @GoogleDeepMind, an upgrade designed to help robots reason about the physical world — The @GeminiApp for Mac desktop (tip: Use Option + Space to access the app via shortcut) — Personal Intelligence in @GeminiApp has new integrations with @GooglePhotos and Nano Banana 2, making it easier to create relevant, personalized images. Available for AI Pro, Plus, and Ultra subscribers in the US — A couple fun additions in @GoogleAIStudio to make building easier, including Design previews and tab tab tab functionality — Skills in @GoogleChrome, which let you save and reuse your most helpful Gemini prompts and run them in your browser with a single click
English
78
75
855
86K
Google AI
Google AI@GoogleAI·
@theoledgers All audio generated by Gemini 3.1 Flash TTS is watermarked with SynthID!
English
1
0
13
2.9K
Theo Ledger
Theo Ledger@theoledgers·
@GoogleAI Great. Do you have watermarks in the audio to help identify for harmful use?
English
1
0
1
2.9K
Google AI
Google AI@GoogleAI·
Today we launched Gemini 3.1 Flash TTS, our most expressive and controllable text-to-speech model yet. This launch [excitement] includes audio tags! 🗣🏷 Audio tags [explanatory] are a seamless way to guide vocal style, pace, and delivery using natural language commands embedded directly in your text. Want a different tempo or tone? [amazement] Just tag the audio to steer the AI-speech output! The model supports 70+ languages (24 of which are high-quality evaluated languages, including: Japanese, Hindi, and Arabic). Watch the audio tags in action in the demo below ↓
English
95
296
2.2K
180.6K
Google AI
Google AI@GoogleAI·
@HoodyAndShorts Placeholders? No these are just examples of how the audio tags work 🙂
English
0
0
1
120
Google AI
Google AI@GoogleAI·
Gemini 3.1 Flash TTS is rolling out in Google Vids and is available today in preview via the Gemini API and in @GoogleAIStudio. Whether you’re creating a pitch deck or recording a passion project, transform your scripts into studio-quality narration: blog.google/innovation-and…
English
8
11
96
21.2K
Google AI
Google AI@GoogleAI·
Got a doodle for your next project laying around? Turn it into working software using @GoogleAIStudio and Nano Banana. Watch us vibe code a weather-responsive outfit selector app from a single, hand-drawn sketch:
English
25
36
218
44K
Google AI
Google AI@GoogleAI·
TGIF! Here are some of our favorite updates from the past week: — Notebooks in @GeminiApp, an integration with @NotebookLM that enables you to retrieve context from your private notebooks or convert your active chats into grounded sources for new research — The @GeminiApp on web now generates customizable interactive visualizations, including 2D and 3D models, directly in your chat to help you deconstruct complex concepts — The new AI-powered @GoogleFinance tool is shipping to 100+ countries, delivering market research, advanced charting, expanded real-time data, and more
English
32
44
275
49.5K
Google AI
Google AI@GoogleAI·
Share your Gemma 4 builds or the model variants you’re training in the replies below!
English
4
1
26
17.8K
Google AI
Google AI@GoogleAI·
We love seeing what you’ve built with Gemma 4, the open model family that we released last week. Here are a few fun examples, described by the builders in their own words (🧵):
English
28
42
692
99.8K
Google AI
Google AI@GoogleAI·
Curious about vibe coding? Or are you already shipping apps and just want an easier way to explain your new favorite hobby to your friends, parents, grandparents, etc.? Either way, this video is for you ⏯️⤵️
English
27
22
202
36.8K
Google AI
Google AI@GoogleAI·
Here’s everything we launched this week (we promise not a single one of these is a joke): — Gemma 4, bringing our most intelligent open models and breakthrough reasoning to your personal hardware and devices while outcompeting models 20x its size — Veo 3.1 Lite, our latest video generation model, which delivers the same speed as Veo 3.1 Fast but at half the cost — Two new service tiers in the Gemini API in @GoogleAIStudio, bringing you granular control over cost and reliability through a single, unified interface — Focus mode in @GoogleAIStudio, the fastest way to make targeted edits to specific parts of your apps — New AI features launched to Google Vids from @GoogleWorkspace, including high-quality video generation from Veo 3.1, available to all users at no cost
English
79
73
825
76.1K
Google AI
Google AI@GoogleAI·
And just in case you’re wondering, "..What’s an open model?", we’ve got you covered: Basically, open models are AI systems where the model weights are publicly available for anyone to download, study, fine-tune and use on your own hardware (phones, computers, etc.). Open models can live on your hardware where your data is completely private and never has to leave your machine. Once you download an open model onto your device, it can run anywhere regardless of internet connection or access to data centers. To name a few examples, Gemma models can run in your pocket, underwater, in outer space, from subway tunnels, and on high-altitude flights without needing a cell tower or WiFi signal. As base models are released (the 'blueprints'), people can then further modify them for specialized use cases via fine-tuning. We’ve seen this in the Gemmaverse, where developers have downloaded Gemma over 400 million times and built more than 100,000 variants. Have you used an open model before? Let us know if you have any other questions about this neat technology!
English
2
11
98
24K
Google AI
Google AI@GoogleAI·
Today, we’re launching Gemma 4, our most intelligent open models to date. Built with the same breakthrough technology as Gemini 3, Gemma 4 brings advanced reasoning to your personal hardware and devices. Here’s what Gemma 4 unlocks for developers: — Intelligence-per-parameter: Our 31B (Dense) and 26B (MoE) models deliver state-of-the-art performance for their size, outcompeting models 20x their size on @arena — Commercial flexibility: Released under a permissive Apache 2.0 license for complete developer flexibility and digital sovereignty — Agentic workflows: Native support for function-calling and structured JSON output allows you to build reliable, autonomous agents — Multimodal edge AI: The E2B and E4B models bring native vision, audio, and low latency to mobile and IoT devices — Long-context reasoning: Up to 256K context windows allow you to process entire repositories or large documents in a single prompt Whether you're building global applications in 140+ languages or local-first AI code assistants, Gemma 4 is built to be your foundation. Explore in @GoogleAIStudio or download the weights on @HuggingFace, @Kaggle, and @Ollama.
English
147
358
2.4K
976.2K