21.05.2024
Project Astra
For Prelims: About Project Astra, Key facts about the multimodal model AI
|
Why in the news?
Recently, Google at the company’s annual developer conference presented an early version of Project Astra.
About Project Astra:
- It is a new multimodal AI agent developed by Google.
- It is capable of answering real-time questions fed to it through text, video, images and speech by pulling up the relevant information.
- It can see the world, remember where one has left a thing and even answer if a computer code is correct by looking at it through the phone’s camera.
- It is more straight-forward, there is no range of emotional diversity in its voice.
- It is not limited to smart phones. Google also showed it being used with a pair of smart glasses.
- It can learn about the world, making it as close as possible to a human-assistant-like experience.
Key facts about the multimodal model AI:
- It is a ML (machine learning) model that is capable of processing information from different modalities, including images, videos and text.
○For example, Google's multimodal model, Gemini, can receive a photo of a plate of cookies and generate a written recipe as a response and vice versa.
- This model expands on generative capabilities, processing information from multiple modalities, including images, videos, and text. Multimodality can be thought of as giving AI the ability to process and understand different sensory modes.
Source: Indian Express
Q - Project Astra, a multimodal AI agent, recently in news is developed by:
A. Google
B.OpenAI
C.C-DAC
D.IndiaAI
Answer A