This course introduces you to the Transformer architecture and the Bidirectional Encoder Representations from Transformers (BERT) model. You learn about the main components of the Transformer architecture, such as the self-attention mechanism, and how it is used to build the BERT model. You also learn about the different tasks that BERT can be used for, such as text classification, question answering, and natural language inference.This course is estimated to take approximately 45 minutes to complete.
This course will introduce you to the attention mechanism, a powerful technique that allows neural networks to focus on specific parts of an input sequence. You will learn how attention works, and how it can be used to improve the performance of a variety of machine learning tasks, including machine translation, text summarization, and question answering. This course is estimated to take approximately 45 minutes to complete.
This course, Introduction to Large Language Models - Locales, is intended for non-English learners. If you want to take this course in English, please enroll in Introduction to Large Language Models. This is an introductory level micro-learning course that explores what large language models (LLM) are, the use cases where they can be utilized, and how you can use prompt tuning to enhance LLM performance. It also covers Google tools to help you develop your own Gen AI apps.
Celem tego szybkiego szkolenia dla początkujących jest wyjaśnienie, czym jest generatywna AI oraz jakie są jej zastosowania. Szkolenie przedstawia również różnice pomiędzy tą technologią a tradycyjnymi systemami uczącymi się, a także narzędzia Google, które pomogą Ci tworzyć własne aplikacje korzystające z generatywnej AI.