Gemini App Gains Audio Input, Search Expands Language Support
Table of Contents
- 1. Gemini App Gains Audio Input, Search Expands Language Support
- 2. Gemini App Now Accepts Audio Files
- 3. Google Search AI Mode Adds Five New Languages
- 4. NotebookLM Receives Report Generation Updates
- 5. The Evolution of AI-Powered Tools
- 6. Frequently Asked Questions About Gemini Updates
- 7. How does Gemini’s audio file analysis compare to ChatGPT’s capabilities?
- 8. Google’s Gemini app Now Supports Audio Files, and Search Expands to Include Five New Languages
- 9. Gemini’s Expanding Capabilities: A Deep Dive
- 10. Analyzing Audio Files wiht Gemini
- 11. Expanding Linguistic Reach: Five New Languages Added
- 12. Benefits of Multilingual Support
- 13. Gemini 2.0 and deep Research Integration
- 14. Practical Tips for Utilizing Gemini’s New Features
- 15. Gemini vs.Competitors: A Quick Comparison
Mountain View, California – Google announced a series of significant enhancements to its Gemini-powered applications on monday, September 9, 2025. The updates focus on improving user accessibility and functionality across the Gemini app,Google Search,and NotebookLM.
Gemini App Now Accepts Audio Files
Responding to widespread user demand, The Gemini application now supports audio file uploads. Josh Woodward,a Vice President at Google Labs and Gemini,confirmed via a social media post that audio input was the most requested feature from users. This will allow users to interact with Gemini using voice notes,podcasts,or other audio sources.
Free Tier users are limited to processing a maximum of 10 minutes of audio,alongside a daily allowance of five prompts. Subscribers to AI Pro and AI Ultra plans benefit from extended capabilities, enabling the upload and processing of audio files up to three hours in length.Gemini can accommodate up to 10 files per prompt,including those contained within compressed ZIP archives.
Google Search AI Mode Adds Five New Languages
Google Search’s AI-powered mode,leveraging the advancements of Gemini 2.5, has broadened its linguistic support to include Hindi, Indonesian, Japanese, Korean, and Brazilian Portuguese. A company blog post highlighted that this expansion aims to provide wider access to the benefits of AI-driven search, enabling users to explore information in thier native languages.
This expansion builds on Google’s broader efforts to make its AI tools more globally accessible,following a trend observed in recent months toward multilingual AI applications. A 2024 report by Statista indicated a 35% rise in the use of AI translation services, signaling a growing need for multilingual AI solutions.
NotebookLM Receives Report Generation Updates
NotebookLM, Google’s AI-powered research and note-taking tool, has been updated with new report generation capabilities. Users can now leverage the software to create various document formats, including study guides, briefing documents, blog posts, flashcards, and quizzes-all available in over 80 languages. The new tool allows for customization of report structure, tone, and style.
NotebookLM already supported audio file analysis, setting it apart as a tool capable of extracting insights from diverse content formats. The latest updates are expected to be fully operational by the end of the current week.
Did You Know? NotebookLM’s ability to process audio files predates this new functionality in the core Gemini app, making it a powerful tool for research and analysis.
| Feature | Gemini (Free) | Gemini (AI Pro/Ultra) |
|---|---|---|
| Max Audio Upload | 10 Minutes | 3 Hours |
| daily Prompts | 5 | Unlimited |
| files Per Prompt | Up to 10 | Up to 10 |
Pro Tip: When uploading audio to Gemini,ensure the files are clearly labeled and organized to maximize the accuracy of the AI’s analysis.
Will these updates significantly alter how users interact with Google’s AI tools? And how will the expanded language support in Google Search impact global information access?
The Evolution of AI-Powered Tools
The continuous evolution of AI-driven applications like Gemini and NotebookLM mirrors the rapid advancements in Natural Language Processing and machine Learning. These tools are transforming how individuals access, process, and utilize information. Ongoing improvements in areas like multilingual support and multimodal input are making these technologies increasingly powerful and accessible to a wider audience.
Frequently Asked Questions About Gemini Updates
- What is the primary new feature in the Gemini app? The Gemini app now supports audio file uploads,a highly anticipated feature for users.
- How does the new Google Search AI Mode benefit users? it allows users to ask complex questions in five additional languages: Hindi, Indonesian, Japanese, Korean, and Brazilian Portuguese.
- what types of reports can notebooklm now generate? NotebookLM can create study guides, briefing documents, blog posts, flashcards, and quizzes.
- Are there any usage limits for the audio feature in Gemini? Free users are limited to 10 minutes of audio and five prompts per day, while Pro/Ultra users have higher limits.
- When will the new NotebookLM report features be fully available? The company expects them to be 100% available by the end of the week.
Share your thoughts on these updates in the comments below!
How does Gemini’s audio file analysis compare to ChatGPT’s capabilities?
Google’s Gemini app Now Supports Audio Files, and Search Expands to Include Five New Languages
Gemini’s Expanding Capabilities: A Deep Dive
Google’s Gemini app is rapidly evolving, becoming an even more versatile AI tool for users worldwide. Recent updates bring two important enhancements: the ability to analyze audio files and an expansion of its search capabilities to encompass five new languages. These improvements solidify Gemini’s position as a leading AI chatbot and AI assistant, offering broader accessibility and functionality. This article will explore these updates in detail, outlining what they mean for users and how to leverage them effectively.
Analyzing Audio Files wiht Gemini
gemini now supports the upload and analysis of audio files.This opens up a wealth of possibilities, moving beyond text-based interactions. Here’s what you can do:
Transcription: Gemini can accurately transcribe audio recordings, converting speech to text.This is incredibly useful for meetings, lectures, interviews, and personal notes.
Summarization: Need a swift overview of a long podcast or meeting recording? Gemini can summarize the key points from the audio file.
Content Extraction: Extract specific facts from audio, such as names, dates, or action items. This streamlines workflows and saves time.
Translation: Translate audio content from one language to another, breaking down communication barriers.
Audio-Based Question Answering: Ask Gemini questions about the content of an audio file, receiving answers based on its analysis.
Supported Audio Formats: Currently, Gemini supports common audio formats like MP3, WAV, and M4A. Google is expected to expand format compatibility in future updates.
Expanding Linguistic Reach: Five New Languages Added
Gemini’s search capabilities have been significantly broadened with the addition of support for five new languages:
Arabic
Hindi
Indonesian
Portuguese
Spanish
This expansion dramatically increases Gemini’s accessibility to a global audience. Users can now interact with Gemini and access information in their native language, fostering a more inclusive and user-kind experience. This is a major step towards democratizing AI technology and making it available to a wider range of users.
Benefits of Multilingual Support
Increased Accessibility: Reaches users who may not be proficient in English.
Improved Accuracy: Native language processing often yields more accurate results.
Enhanced User Experience: users feel more comfortable and engaged when interacting in their preferred language.
Global Collaboration: Facilitates communication and collaboration across language barriers.
Wider Information Access: Unlocks access to a broader range of information sources.
Gemini 2.0 and deep Research Integration
While not directly related to the audio and language updates, it’s significant to note the recent rollout of Gemini 2.0 and the “Deep Research” feature. As of today, September 9th, 2025, the Gemini 2.0 Flash experimental model is available to all Gemini users. The Deep Research function, available in Gemini Advanced, leverages advanced reasoning and long-context capabilities to act as a research assistant. This means Gemini can now:
explore complex topics in depth.
Synthesize information from multiple sources.
Generate thorough reports.
Assist with academic research and professional analysis.
This integration with gemini 2.0 further enhances Gemini’s capabilities as a powerful AI research tool.
Practical Tips for Utilizing Gemini’s New Features
Audio File Optimization: Ensure your audio files are clear and have minimal background noise for optimal transcription accuracy.
Specific Prompts: When asking questions about audio files, be as specific as possible to get the most relevant answers.
Language Selection: Always verify that Gemini is set to the correct language for your queries.
Experiment with deep Research: Utilize the Deep Research feature for complex topics requiring in-depth analysis.
Stay Updated: Google is continuously improving Gemini. Regularly check for updates and new features.
Gemini vs.Competitors: A Quick Comparison
| Feature | Gemini | ChatGPT | Bard |
|—|—|—|—|
| Audio File Analysis | Yes | Limited | No |
| Multilingual Support | Expanding (Now