Maggie Nanyonga
Member since 2026
Silver League
12674 points
Member since 2026
In this Google DeepMind course, you will learn the fundamentals of language models and gain a high-level understanding of the machine learning development pipeline. You will consider the strengths and limitations of traditional n-gram models and advanced transformer models. Practical coding labs will enable you to develop insights into how machine learning models work and how they can be used to generate text and identify patterns in language. Through real-world case studies, you will build an understanding around how research engineers operate. Drawing on these insights you will identify problems that you wish to tackle in your own community and consider how to leverage the power of machine learning responsibly to address these problems within a global and local context.
Welcome to the "AI Infrastructure: Networking Techniques" course. In this course, you'll learn to leverage Google Cloud's high-bandwidth, low-latency infrastructure to optimize data transfer and communication between all the components of your AI system. By the end, you'll grasp the critical role networking plays across the entire AI pipeline from data ingestion and training to inference and be able to apply best practices to ensure your workloads run at maximum speed.
In this course, you’ll take a comprehensive journey through the storage solutions available on Google Cloud, specifically tailored for AI and high-performance computing (HPC) workloads. You’ll learn how to choose the right storage for each stage of the ML lifecycle. You’ll explore how to optimize for I/O performance during training, manage massive datasets for data preparation, and serve model artifacts with low latency. Through practical examples and demonstrations, you’ll gain the expertise to design robust storage solutions that accelerate your AI innovation.
This course provides a comprehensive guide to deploying, managing, and optimizing AI and high-performance computing (HPC) workloads on Google Cloud. Through a series of lessons and practical demonstrations, you’ll explore diverse deployment strategies, ranging from highly customizable environments using Google Compute Engine (GCE) to managed solutions like Google Kubernetes Engine (GKE). Specifically, you’ll learn how to create clusters and deploy GKE for inference.
Welcome to the Cloud TPUs course. We'll explore the advantages and disadvantages of TPUs in various scenarios and compare different TPU accelerators to help you choose the right fit. You'll learn strategies to maximize performance and efficiency for your AI models and understand the significance of GPU/TPU interoperability for flexible machine learning workflows. Through engaging content and practical demos, we'll guide you step-by-step in leveraging TPUs effectively.
Curious about the powerful hardware behind AI? This module breaks down performance-optimized AI computers, showing you why they're so important. We'll explore how CPUs, GPUs, and TPUs make AI tasks super fast, what makes each one unique, and how AI software gets the most out of them. By the end, you'll know exactly how to pick the right GPU for your AI projects, helping you make smart choices for your AI workloads.
Ready to get started with AI Hypercomputers? This course makes it easy! We'll cover the basics of what they are and how they help AI with AI workloads. You'll learn about the different components inside a hypercomputer, like GPUs, TPUs, and CPUs, and discover how to pick the right deployment approach for your needs.
This course introduces concepts of responsible AI and AI principles. It covers techniques to practically identify fairness and bias and mitigate bias in AI/ML practices. It explores practical methods and tools to implement Responsible AI best practices using Google Cloud products and open source tools.
This course introduces important topics of AI privacy and safety. It explores practical methods and tools to implement AI privacy and safety recommended practices through the use of Google Cloud products and open-source tools.
Artificial Intelligence (AI) offers transformative possibilities, but it also introduces new security challenges. This course equips security and data protection leaders with strategies to securely manage AI within their organizations. Learn a framework for proactively identifying and mitigating AI-specific risks, protecting sensitive data, ensuring compliance, and building a resilient AI infrastructure. Pick use cases from four different industries to explore how these strategies apply in real-world scenarios.
This course introduces concepts of AI interpretability and transparency. It discusses the importance of AI transparency for developers and engineers. It explores practical methods and tools to help achieve interpretability and transparency in both data and AI models.
This course reviews the essential security features of Model Armor and equips you to work with the service. You’ll learn about the security risks associated with LLMs and how Model Armor protects your AI applications.
As the use of enterprise Artificial Intelligence and Machine Learning continues to grow, so too does the importance of building it responsibly. A challenge for many is that talking about responsible AI can be easier than putting it into practice. If you’re interested in learning how to operationalize responsible AI in your organization, this course is for you. In this course, you will learn how Google Cloud does this today, together with best practices and lessons learned, to serve as a framework for you to build your own responsible AI approach.
This course equips learners with the essential knowledge and practical tools to develop and implement artificial intelligence (AI) responsibly. Through an exploration of ethical considerations, best practices, and governance procedures, participants will gain an understanding of how to navigate the complex landscape of AI while upholding ethical standards and minimizing potential risks.
Gen AI Agents: Transform Your Organization is the fifth and final course of the Gen AI Leader learning path. This course explores how organizations can use custom gen AI agents to help tackle specific business challenges. You gain hands-on practice building a basic gen AI agent, while exploring the components of these agents, such as models, reasoning loops, and tools.
Transform Your Work With Gen AI Apps is the fourth course of the Gen AI Leader learning path. This course introduces Google’s gen AI applications, such as Google Workspace with Gemini and NotebookLM. It guides you through concepts like grounding, retrieval augmented generation, constructing effective prompts and building automated workflows.
Gen AI: Navigate the Landscape s the third course of the Gen AI Leader learning path. Gen AI is changing how we work and interact with the world around us. But as a leader, how can you harness its power to drive real business outcomes? In this course, you explore the different layers of building gen AI solutions, Google Cloud’s offerings, and the factors to consider when selecting a solution.
Gen AI: Unlock Foundational Concepts is the second course of the Gen AI Leader learning path. In this course, you unlock the foundational concepts of generative AI by exploring the differences between AI, ML, and gen AI, and understanding how various data types enable generative AI to address business challenges. You also gain insights into Google Cloud strategies to address the limitations of foundation models and the key challenges for responsible and secure AI development and deployment.
Gen AI: Beyond the Chatbot is the first course of the Gen AI Leader learning path and has no prerequisites. This course aims to move beyond the basic understanding of chatbots to explore the true potential of generative AI for your organization. You explore concepts like foundation models and prompt engineering, which are crucial for leveraging the power of gen AI. The course also guides you through important considerations you should make when developing a successful gen AI strategy for your organization.
Create your first Gemini Enterprise application to earn a skill badge! Connect diverse data sources to your application to build a powerful, unified search and analysis engine. Master advanced capabilities like deep research agents, multi-agent ideation, and NotebookLM for focused analysis.
Discover how AI agents drive business impact. You’ll map agent types to your KPIs and explore use cases that solve real bottlenecks. Then, learn how Gemini Enterprise empowers you to build and orchestrate the right agents—from no-code to high-code solutions.
AI Agents represent a major shift beyond traditional large language models (LLMs): instead of simply generating text-based solutions, they can also act autonomously to execute them. This course introduces the fundamentals of AI Agents, how they differ from LLM APIs, and where they add value in the real world. Based on Google’s agents whitepaper, it provides the theoretical foundation needed before writing your first lines of agent code—ideal for developers, architects, and technical decision-makers who want to understand AI systems through the lens of autonomous, goal-directed behavior (and not just text generation). Join the community forum for questions and discussions.
Gain a conceptual overview of AI Agents. Discover how AI Agents use autonomous action and reasoning to solve complex problems. You’ll explore the technical architecture—models, tools, and orchestration—that enables agents to learn, plan, and achieve goals on your behalf.
This course equips machine learning practitioners with the essential tools, techniques, and best practices for evaluating both generative and predictive AI models. Model evaluation is a critical discipline for ensuring that ML systems deliver reliable, accurate, and high-performing results in production. Participants will gain a deep understanding of various evaluation metrics, methodologies, and their appropriate application across different model types and tasks. The course will emphasize the unique challenges posed by generative AI models and provide strategies for tackling them effectively. By leveraging Google Cloud's Vertex AI platform, participants will learn how to implement robust evaluation processes for model selection, optimization, and continuous monitoring.
This course is dedicated to equipping you with the knowledge and tools needed to uncover the unique challenges faced by MLOps teams when deploying and managing Generative AI models, and exploring how Vertex AI empowers AI teams to streamline MLOps processes and achieve success in Generative AI projects.