Dhruvil Patel
Member since 2025
Silver League
3212 points
Member since 2025
As the use of enterprise Artificial Intelligence and Machine Learning continues to grow, so too does the importance of building it responsibly. A challenge for many is that talking about responsible AI can be easier than putting it into practice. If you’re interested in learning how to operationalize responsible AI in your organization, this course is for you. In this course, you will learn how Google Cloud does this today, together with best practices and lessons learned, to serve as a framework for you to build your own responsible AI approach.
This is an introductory-level microlearning course aimed at explaining what responsible AI is, why it's important, and how Google implements responsible AI in their products. It also introduces Google's 3 AI principles.
This is an introductory level micro-learning course that explores what large language models (LLM) are, the use cases where they can be utilized, and how you can use prompt tuning to enhance LLM performance. It also covers Google tools to help you develop your own Gen AI apps.
This is an introductory level microlearning course aimed at explaining what Generative AI is, how it is used, and how it differs from traditional machine learning methods. It also covers Google Tools to help you develop your own Gen AI apps.
This course introduces concepts of responsible AI and AI principles. It covers techniques to practically identify fairness and bias and mitigate bias in AI/ML practices. It explores practical methods and tools to implement Responsible AI best practices using Google Cloud products and open source tools.
This course equips machine learning practitioners with the essential tools, techniques, and best practices for evaluating both generative and predictive AI models. Model evaluation is a critical discipline for ensuring that ML systems deliver reliable, accurate, and high-performing results in production. Participants will gain a deep understanding of various evaluation metrics, methodologies, and their appropriate application across different model types and tasks. The course will emphasize the unique challenges posed by generative AI models and provide strategies for tackling them effectively. By leveraging Google Cloud's Vertex AI platform, participants will learn how to implement robust evaluation processes for model selection, optimization, and continuous monitoring.
This course is dedicated to equipping you with the knowledge and tools needed to uncover the unique challenges faced by MLOps teams when deploying and managing Generative AI models, and exploring how Vertex AI empowers AI teams to streamline MLOps processes and achieve success in Generative AI projects.
Moving to the cloud creates numerous opportunities to start working in a new way and it empowers the workforce to better collaborate and innovate. But it’s also a big change. Sometimes the success of the change hinges not on the change itself, but on how it’s managed. This course will help people managers to understand some of the key challenges associated with cloud adoption, and provide them with a verified in-the-field framework that will assist them in supporting their teams on the change journey. By addressing the human factor of moving to the cloud, organizations increase their chances of realizing business objectives and investing in their future talent.
This course is for developers interested in learning how to use TPUs for inference—from architecture to deployment, and how to solve common implementation challenges.
This course is designed for developers looking to build an optimized AI inference stack on Google Cloud. Whether you’re working with GPUs or TPUs, you’ll explore the fundamental components of an inference stack, learn design principles for maximizing performance and reliability, and explore practical techniques to take your workloads from 0 to 1.