Google I/O 2024 Developer Conference was held as scheduled, with this year’s focus on the latest artificial intelligence technology and product updates. In the two-hour keynote speech, Google unveiled a series of innovative technological breakthroughs and updates aimed at providing developers and consumers with more convenience and innovative experiences.
Table of Contents
Toggle
Next-generation development tool: Firebase Genkit
AI application is the focus
AI innovation in the education field: LearnLM
New feature for YouTube educational videos: AI-generated quizzes
Enhanced AI capabilities: Updates to Gemma 2
New discovery feature for Google Play
Protecting user safety: Detecting fraud activities in calls
Innovative search and interaction method: Ask Photos
Gemini AI applications
Gemini applications in Gmail
Gemini 1.5 Pro: Doubled processing power
Gemini Live: Real-time interactive experience
Gemini Nano: Integrated micro AI on Chrome
Gemini applications on Android
Gemini applications on Google Maps
Expanded AI capabilities: Performance improvement of Tensor Processing Units
In this year’s conference, Google introduced a new platform called Firebase Genkit. This is an open-source framework designed to simplify the development of artificial intelligence applications using JavaScript/TypeScript, with support for Go language updates coming soon. The launch of Firebase Genkit aims to accelerate the implementation of AI functionalities in both new and existing applications, covering various use cases such as content generation, summarization, text translation, and image generation.
During the keynote speech, Google mentioned AI technology a total of 121 times, demonstrating the company’s deep involvement in this field. CEO Sundar Pichai even emphasized that this is part of Google’s effort to showcase its dedication to AI to the outside world.
Google also introduced a new generation of generative AI model called LearnLM, specifically trained for the education field. This series of models, developed in collaboration between Google’s DeepMind AI research division and Google Research, aims to support student learning through conversational teaching. LearnLM has been tested in Google Classroom and will be further integrated into curriculum planning and optimization in the future.
YouTube has added an AI-generated quiz feature, allowing viewers to interact while watching educational videos, such as asking questions, getting explanations, or participating in quizzes. This feature provides a new way of learning for users who need to watch lengthy educational videos.
To meet the needs of developers, Google will soon add a new model with 2.7 billion parameters to Gemma 2. This model, optimized by Nvidia, can efficiently run on next-generation GPUs.
Google Play has updated its discovery feature, making it easier for users and developers to promote and find applications. This includes new user acquisition methods, updates to Play Points, and enhancements to other tools and APIs for developers.
Google previewed a new feature that can detect potential fraud activities in calls in real-time, which will be integrated into future versions of Android. By analyzing conversation patterns in calls, this system can effectively warn users about security risks.
Google Photos will soon introduce an experimental feature called “Ask Photos,” which uses AI to understand the content and metadata of photos. Users can use natural language queries to make the search process more intuitive and require less manual intervention.
In Gmail, users will be able to search, summarize, and compose emails using Gemini AI technology. Additionally, Gemini AI can perform more complex tasks such as handling e-commerce returns, including searching the inbox, finding receipts, and filling out online forms.
The upgrade of Gemini 1.5 Pro allows it to analyze longer documents, code repositories, videos, and recordings than ever before. In the latest private preview, the processing power of this flagship model has increased to handle up to 2 million tokens, double its previous capacity.
Google previewed a new feature called Gemini Live, which allows users to have deep voice conversations with Gemini through smartphones. Users can interrupt Gemini in the conversation, and the system will adapt to the user’s voice patterns in real-time. Furthermore, Gemini can identify and respond to the user’s surrounding environment through the smartphone camera.
Gemini Nano, the smallest member of the Google AI model family, will be directly integrated into the Chrome desktop client starting from Chrome version 126. This will allow developers to implement their own AI functionalities using models on their devices, such as the “Smart Compose” tool in Gmail.
Google’s Gemini AI will replace Google Assistant and be deeply integrated into the Android system. Users will be able to directly drag and drop AI-generated images into applications such as Gmail and Google Messages. YouTube users will also be able to use the “Ask this video” feature to search for specific information in videos.
Gemini’s capabilities will be applied to Google Maps’ developer platform, starting with the Places API. Developers will be able to showcase location and area summaries analyzed by Gemini in their own applications and websites. Developers will no longer need to write their own custom location descriptions.
Google also announced the sixth generation of its Tensor Processing Units (TPU) AI chips, named Trillium, which will be released later this year with significant performance improvements.
Google I/O 2024 showcased Google’s latest achievements in AI and technological innovation. From educational tools to developer resources, each update aims to improve efficiency, enhance interactivity, and ensure user safety.
AI
Gemini AI
Google I/O