Roberto Panai
Member since 2020
Member since 2020
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Learners will get hands-on practice using Vertex AI Feature Store's streaming ingestion at the SDK layer.
Unite Google’s expertise in search and AI with Gemini Enterprise, a powerful tool designed to help employees find specific information from document storage, email, chats, ticketing systems, and other data sources, all from a single search bar. The Gemini Enterprise assistant can also help brainstorm, research, outline documents, and take actions like inviting coworkers to a calendar event to accelerate knowledge work and collaboration of all kinds. (Please note Gemini Enterprise was previously named Google Agentspace, there may be references to the previous product name in this course.)
This course introduces Google Cloud's AI and machine learning (ML) capabilities, with a focus on developing both generative and predictive AI projects. It explores the various technologies, products, and tools available throughout the data-to-AI lifecycle, empowering data scientists, AI developers, and ML engineers to enhance their expertise through interactive exercises.
This course is dedicated to equipping you with the knowledge and tools needed to uncover the unique challenges faced by MLOps teams when deploying and managing Generative AI models, and exploring how Vertex AI empowers AI teams to streamline MLOps processes and achieve success in Generative AI projects.
This course explores the benefits of using Vertex AI Feature Store, how to improve the accuracy of ML models, and how to find which data columns make the most useful features. This course also includes content and labs on feature engineering using BigQuery ML, Keras, and TensorFlow.