Modupe Ajala
Membro dal giorno 2023
Campionato Argento
37866 punti
Membro dal giorno 2023
This structured course is for developers interested in building intelligent agents using the Agent Development Kit (ADK). It combines hands-on experience, core concepts, and practical application, to provide a comprehensive guide to using ADK. You can also join our community of Google Cloud experts and peers to ask questions, collaborate on answers, and connect with the Googlers making the products you use every day.
In this course, you learn about data engineering on Google Cloud, the roles and responsibilities of data engineers, and how those map to offerings provided by Google Cloud. You also learn about ways to address data engineering challenges.
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
I due componenti chiave di qualsiasi pipeline di dati sono costituiti dai data lake e dai data warehouse. In questo corso evidenzieremo i casi d'uso per ogni tipo di spazio di archiviazione e approfondiremo i dettagli tecnici delle soluzioni di data lake e data warehouse disponibili su Google Cloud. Inoltre, descriveremo il ruolo di un data engineer, illustreremo i vantaggi di una pipeline di dati di successo per le operazioni aziendali ed esamineremo i motivi per cui il data engineering dovrebbe essere eseguito in un ambiente cloud. Questo è il primo corso della serie Data engineering su Google Cloud. Dopo il completamento di questo corso, iscriviti al corso Creazione di pipeline di dati in batch su Google Cloud.
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Machine Learning Engineering professionals use tools for continuous improvement and evaluation of deployed models. They work with (or can be) Data Scientists, who develop models, to enable velocity and rigor in deploying the best performing models.
This course is dedicated to equipping you with the knowledge and tools needed to uncover the unique challenges faced by MLOps teams when deploying and managing Generative AI models, and exploring how Vertex AI empowers AI teams to streamline MLOps processes and achieve success in Generative AI projects.
This course explores a Retrieval Augmented Generation (RAG) solution in BigQuery to mitigate AI hallucinations. It introduces a RAG workflow that encompasses creating embeddings, searching a vector space, and generating improved answers. The course explains the conceptual reasons behind these steps and their practical implementation with BigQuery. By the end of the course, learners will be able to build a RAG pipeline using BigQuery and generative AI models like Gemini and embedding models to address their own AI hallucination use cases.
This course demonstrates how to use AI/ML models for generative AI tasks in BigQuery. Through a practical use case involving customer relationship management, you learn the workflow of solving a business problem with Gemini models. To facilitate comprehension, the course also provides step-by-step guidance through coding solutions using both SQL queries and Python notebooks.
This course explores Gemini in BigQuery, a suite of AI-driven features to assist data-to-AI workflow. These features include data exploration and preparation, code generation and troubleshooting, and workflow discovery and visualization. Through conceptual explanations, a practical use case, and hands-on labs, the course empowers data practitioners to boost their productivity and expedite the development pipeline.
Complete the intermediate Build a Data Warehouse with BigQuery skill badge course to demonstrate skills in the following: joining data to create new tables, troubleshooting joins, appending data with unions, creating date-partitioned tables, and working with JSON, arrays, and structs in BigQuery.
Google Cloud Fundamentals: Core Infrastructure introduce concetti e terminologia importanti per lavorare con Google Cloud. Attraverso video e lab pratici, questo corso presenta e confronta molti dei servizi di computing e archiviazione di Google Cloud, insieme a importanti strumenti di gestione delle risorse e dei criteri.
This Data Analytics course consists of a series of advanced-level labs designed to validate your proficiency in using Google Cloud services. Each lab presents a set of the required tasks that you must complete with minimal assistance. The labs in this course have replaced the previous L300 Data Analytics Challenge Lab. If you have already completed the Challenge Lab as part of your L300 accreditation requirement, it will be carried over and count towards your L300 status. You must score 80% or higher for each lab to complete this course, and fulfill your CEPF L300 Data Analytics requirement. For technical issues with a Challenge Lab, please raise a Buganizer ticket using this CEPF Buganizer template: go/cepfl300labsupport
In this quest, you will get hands-on experience with LookML in Looker. You will learn how to write LookML code to create new dimensions and measures, create derived tables and join them to Explores, filter Explores, and define caching policies in LookML.
Questo corso introduce i modelli di diffusione, una famiglia di modelli di machine learning che recentemente si sono dimostrati promettenti nello spazio di generazione delle immagini. I modelli di diffusione traggono ispirazione dalla fisica, in particolare dalla termodinamica. Negli ultimi anni, i modelli di diffusione sono diventati popolari sia nella ricerca che nella produzione. I modelli di diffusione sono alla base di molti modelli e strumenti di generazione di immagini all'avanguardia su Google Cloud. Questo corso ti introduce alla teoria alla base dei modelli di diffusione e a come addestrarli ed eseguirne il deployment su Vertex AI.
In this course, you learn how to do the kind of data exploration and analysis in Looker that would formerly be done primarily by SQL developers or analysts. Upon completion of this course, you will be able to leverage Looker's modern analytics platform to find and explore relevant content in your organization’s Looker instance, ask questions of your data, create new metrics as needed, and build and share visualizations and dashboards to facilitate data-driven decision making.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
L'integrazione del machine learning nelle pipeline di dati aumenta la capacità di estrarre insight dai dati. Questo corso illustra i modi in cui il machine learning può essere incluso nelle pipeline di dati su Google Cloud. Per una personalizzazione minima o nulla, il corso tratta di AutoML. Per funzionalità di machine learning più personalizzate, il corso introduce Notebooks e BigQuery Machine Learning (BigQuery ML). Inoltre, il corso spiega come mettere in produzione soluzioni di machine learning utilizzando Vertex AI.
Questo corso presenta i prodotti e i servizi per big data e di machine learning di Google Cloud che supportano il ciclo di vita dai dati all'IA. Esplora i processi, le sfide e i vantaggi della creazione di una pipeline di big data e di modelli di machine learning con Vertex AI su Google Cloud.
This course, Building Resilient Streaming Analytics Systems on Google Cloud - Locales, is intended for non-English learners. If you want to take this course in English, please enroll in Building Resilient Streaming Analytics Systems on Google Cloud. Processing streaming data is becoming increasingly popular as streaming enables businesses to get real-time metrics on business operations. This course covers how to build streaming data pipelines on Google Cloud. Pub/Sub is described for handling incoming streaming data. The course also covers how to apply aggregations and transformations to streaming data using Dataflow, and how to store processed records to BigQuery or Cloud Bigtable for analysis. Learners will get hands-on experience building streaming data pipeline components on Google Cloud using QwikLabs.
Le pipeline di dati in genere rientrano in uno dei paradigmi EL (Extract, Load), ELT (Extract, Load, Transform) o ETL (Extract, Transform, Load). Questo corso descrive quale paradigma dovrebbe essere utilizzato e quando per i dati in batch. Inoltre, questo corso tratta diverse tecnologie su Google Cloud per la trasformazione dei dati, tra cui BigQuery, l'esecuzione di Spark su Dataproc, i grafici della pipeline in Cloud Data Fusion e trattamento dati serverless con Dataflow. Gli studenti fanno esperienza pratica nella creazione di componenti della pipeline di dati su Google Cloud utilizzando Qwiklabs.