完成「使用 Dataplex 建構資料網格」技能徽章入門課程,即可證明您具備下列技能:使用 Dataplex 建構資料網格, 以利在 Google Cloud 維護資料安全性,並協助治理和探索資料。您將練習並測試自己的技能,包括在 Dataplex 為資產加上標記、指派 IAM 角色,以及評估資料品質。
本課程將示範如何在 BigQuery 運用 AI/機器學行模型,以執行生成式 AI 任務。透過涉及顧客關係管理的應用實例,您將瞭解運用 Gemini 模型解決業務問題的工作流程。為了便於理解,本課程還提供了採用 SQL 查詢和 Python 筆記本的程式設計解決方案,指導您逐步操作。
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
本課程會說明 Gemini in BigQuery,這是一套由 AI 輔助的功能,可協助「從資料到 AI」的工作流程。這些功能包含資料探索和準備、程式碼生成和疑難排解,以及工作流程探索和視覺化。本課程將透過概念解說、應用實例和實作實驗室,協助資料從業人員提升工作效率,並加速開發 pipeline。
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
在本課程中,您會學到 Google Cloud 上的資料工程、資料工程師的角色與職責,以及這些內容如何對應至 Google Cloud 提供的服務。您也將瞭解處理資料工程難題的許多方法。
In this quest you will use a collection of Google APIs that are all related to language, and speech. You will use the Speech-to-Text API to transcribe an audio file into a text file, the Cloud Translation API to translate from one language to another, the Cloud Translation API to detect what language is being used and translate to a different language, the Natural Language API to classify text and analyze sentiment, and create synthetic speech.
This is the first of two Quests of hands-on labs is derived from the exercises from the book Data Science on Google Cloud Platform, 2nd Edition by Valliappa Lakshmanan, published by O'Reilly Media, Inc. In this first Quest, covering up through chapter 8, you are given the opportunity to practice all aspects of ingestion, preparation, processing, querying, exploring and visualizing data sets using Google Cloud tools and services.
This is the second of two Quests of hands-on labs derived from the exercises from the book Data Science on Google Cloud Platform, 2nd Edition by Valliappa Lakshmanan, published by O'Reilly Media, Inc. In this second Quest, covering chapter 9 through the end of the book, you extend the skills practiced in the first Quest, and run full-fledged machine learning jobs with state-of-the-art tools and real-world data sets, all using Google Cloud tools and services.
完成在 BigQuery 執行預測資料分析技能徽章中階課程, 即可證明您具備下列技能:可匯入 CSV 和 JSON 檔案,在 BigQuery 建立資料集; 可運用 BigQuery 的強大功能與複雜的 SQL 分析概念,包括使用 BigQuery ML 根據足球賽事資料訓練出預期進球模型,評估世界盃進球的精彩程度。
In this course you will learn how to use several BigQuery ML features to improve retail use cases. Predict the demand for bike rentals in NYC with demand forecasting, and see how to use BigQuery ML for a classification task that predicts the likelihood of a website visitor making a purchase.
It's no secret that machine learning is one of the fastest growing fields in tech, and Google Cloud has been instrumental in furthering its development. With a host of APIs, Google Cloud has a tool for just about any machine learning job. In this advanced-level course, you will get hands-on practice with machine learning APIs by taking labs like Detect Labels, Faces, and Landmarks in Images with the Cloud Vision API. Looking for a hands-on challenge lab to demonstrate your skills and validate your knowledge? Enroll in and finish the additional challenge lab at the end of this quest to receive an exclusive Google Cloud digital badge.
大數據、機器學習和人工智慧 (AI) 是時下熱門的 電腦相關話題,但這些領域相當專業,就算想要入門 也難以取得教材或資料。幸好,Google Cloud 提供了此領域的多種服務,而且容易使用。 參加這堂入門課程,您就能踏出第一步, 開始學習運用 BigQuery、Cloud Speech API 以及 Video Intelligence 等工具。
完成「運用 BigQuery ML 建立機器學習模型」技能徽章中階課程,即可證明您具備下列技能: 可使用 BigQuery ML 建立及評估機器學習模型,並根據資料進行預測。
不想花費大把時間,想在幾分鐘內只靠 SQL,就建立好機器學習模型嗎?透過 BigQuery ML,資料分析師可以運用現有的 SQL 工具和技巧,建立、訓練、評估模型, 並使用模型進行預測,降低機器學習的使用門檻。在 本系列的實驗室,您會測試不同類型的模型,瞭解 優良模型應具備的條件。
完成 透過 Vertex AI 建構及部署機器學習解決方案 課程,即可瞭解如何使用 Google Cloud 的 Vertex AI 平台、AutoML 和自訂訓練服務, 訓練、評估、調整、解釋及部署機器學習模型。 這個技能徽章課程適合專業數據資料學家和機器學習 工程師,完成即可取得中階技能徽章。技能 徽章是 Google Cloud 核發的獨家數位徽章, 用於肯定您在 Google Cloud 產品和服務方面的精通程度, 代表您已通過測驗,能在互動式實作環境應用相關知識。完成這個技能徽章課程 和結業評量挑戰實驗室,就能獲得數位徽章, 並與親友分享。
完成「為 Looker 資訊主頁和報表準備資料」技能徽章入門課程, 即可證明您具備下列技能:可篩選、排序和 pivot 資料、合併不同的 Looker 探索結果, 還能使用函式和運算子建構 Looker 資訊主頁和報表,取得資料分析結果和圖表。
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
Want to scale your data analysis efforts without managing database hardware? Learn the best practices for querying and getting insights from your data warehouse with this interactive series of BigQuery labs. BigQuery is Google's fully managed, NoOps, low cost analytics database. With BigQuery you can query terabytes and terabytes of data without having any infrastructure to manage or needing a database administrator. BigQuery uses SQL and can take advantage of the pay-as-you-go model. BigQuery allows you to focus on analyzing data to find meaningful insights.
完成 從 BigQuery 資料取得深入分析結果 技能徽章入門課程,即可證明您具備下列技能: 撰寫 SQL 查詢、查詢公開資料表、將樣本資料載入 BigQuery、使用 BigQuery 的查詢驗證工具 排解常見語法錯誤,以及在 Looker Studio 中 透過連結 BigQuery 資料建立報表。
This advanced-level quest is unique amongst the other catalog offerings. The labs have been curated to give IT professionals hands-on practice with topics and services that appear in the Google Cloud Certified Professional Data Engineer Certification. From Big Query, to Dataprep, to Cloud Composer, this quest is composed of specific labs that will put your Google Cloud data engineering knowledge to the test. Be aware that while practice with these labs will increase your skills and abilities, you will need other preparation, too. The exam is quite challenging and external studying, experience, and/or background in cloud data engineering is recommended. Looking for a hands on challenge lab to demonstrate your skills and validate your knowledge? On completing this quest, enroll in and finish the additional challenge lab at the end of the Engineer Data in the Google Cloud to receive an exclusive Google Cloud digital badge.
完成使用 BigQuery ML 為預測模型進行資料工程技能徽章中階課程, 即可證明自己具備下列知識與技能:運用 Dataprep by Trifacta 建構連至 BigQuery 的資料轉換 pipeline; 使用 Cloud Storage、Dataflow 和 BigQuery 建構「擷取、轉換及載入」(ETL) 工作負載, 以及使用 BigQuery ML 建構機器學習模型。
完成 透過 BigQuery 建構資料倉儲 技能徽章中階課程,即可證明您具備下列技能: 彙整資料以建立新資料表、排解彙整作業問題、利用聯集附加資料、建立依日期分區的資料表, 以及在 BigQuery 使用 JSON、陣列和結構體。
Looking to build or optimize your data warehouse? Learn best practices to Extract, Transform, and Load your data into Google Cloud with BigQuery. In this series of interactive labs you will create and optimize your own data warehouse using a variety of large-scale BigQuery public datasets. BigQuery is Google's fully managed, NoOps, low cost analytics database. With BigQuery you can query terabytes and terabytes of data without having any infrastructure to manage or needing a database administrator. BigQuery uses SQL and can take advantage of the pay-as-you-go model. BigQuery allows you to focus on analyzing data to find meaningful insights. Looking for a hands on challenge lab to demonstrate your skills and validate your knowledge? On completing this quest, enroll in and finish the additional challenge lab at the end of this quest to receive an exclusive Google Cloud digital badge.
完成 在 Google Cloud 為機器學習 API 準備資料 技能徽章入門課程,即可證明您具備下列技能: 使用 Dataprep by Trifacta 清理資料、在 Dataflow 執行資料管道、在 Dataproc 建立叢集和執行 Apache Spark 工作,以及呼叫機器學習 API,包含 Cloud Natural Language API、Google Cloud Speech-to-Text API 和 Video Intelligence API。
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
In this course you will get hands-on in order to work through real-world challenges faced when building streaming data pipelines. The primary focus is on managing continuous, unbounded data with Google Cloud products.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.