
Google has announced the public preview of Gemini 1.5 Pro on its Vertex AI platform, marking a significant advancement in its generative AI capabilities. The model, which was launched in February, can process between 128,000 and 1 million tokens, making it one of Google's most capable AI models to date. This release is part of Google's broader effort to integrate advanced AI solutions into its cloud services, as highlighted during the #GoogleCloudNext event. The Gemini 1.5 Pro model supports a wide range of functionalities including audio (speech) understanding and a new File API for easier file handling. Additionally, Google has introduced new integrations and accelerators through a deepened partnership with DataRobot, aimed at facilitating the adoption of AI solutions by organizations. The model is now available in over 180 countries via the Gemini API in public preview, and it also features the world's largest context window for processing audio streams, including speech. Furthermore, businesses can now enhance Gemini models with Google Search capabilities on Vertex AI, providing a more comprehensive AI toolset. The TPU v5p is also now generally available, further enhancing Google's cloud offerings.









It can "hear" -> Google’s Gemini 1.5 Pro can now process audio w/out needing a transcript "The model can now listen to uploaded audio files and churn out information from things like earnings calls or audio from videos without the need to refer to a written transcript."… https://t.co/Rk5yIObmoZ
🚨 Google Gemini biggest update is here, Gemini API in a public preview and along with native audio (speech) 🔥 Now it can reason with both image frames and speech within video files uploaded in Google AI Studio. https://t.co/qQQDtzTfWK
Google Gemini 1.5 Pro's 1,000,000+ token context length is not talked about enough. At launch, 1.5 Pro was overshadowed by OpenAI's Sora and only accessible to select users. But now, anyone can try it for free. Here's a quick tutorial 🧵: https://t.co/lIjeHE8W9u