Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying Generative AI Foundations in Python
  • Table Of Contents Toc
Generative AI Foundations in Python

Generative AI Foundations in Python

By : Carlos Rodriguez
4.8 (5)
close
close
Generative AI Foundations in Python

Generative AI Foundations in Python

4.8 (5)
By: Carlos Rodriguez

Overview of this book

The intricacies and breadth of generative AI (GenAI) and large language models can sometimes eclipse their practical application. It is pivotal to understand the foundational concepts needed to implement generative AI. This guide explains the core concepts behind -of-the-art generative models by combining theory and hands-on application. Generative AI Foundations in Python begins by laying a foundational understanding, presenting the fundamentals of generative LLMs and their historical evolution, while also setting the stage for deeper exploration. You’ll also understand how to apply generative LLMs in real-world applications. The book cuts through the complexity and offers actionable guidance on deploying and fine-tuning pre-trained language models with Python. Later, you’ll delve into topics such as task-specific fine-tuning, domain adaptation, prompt engineering, quantitative evaluation, and responsible AI, focusing on how to effectively and responsibly use generative LLMs. By the end of this book, you’ll be well-versed in applying generative AI capabilities to real-world problems, confidently navigating its enormous potential ethically and responsibly.
Table of Contents (13 chapters)
close
close
Lock Free Chapter
1
Part 1: Foundations of Generative AI and the Evolution of Large Language Models
6
Part 2: Practical Applications of Generative AI

Surveying GenAI Types and Modes: An Overview of GANs, Diffusers, and Transformers

In the previous chapter, we established the key distinction between generative and discriminative models. Discriminative models focus on predicting outputs by learning p(outputinput), or the conditional probability of some expected output given an input or set of inputs. In contrast, generative models, such as Generative Pretrained Transformer (GPT), generate text by predicting the next token (a partial word, whole word, or punctuation) using p(next tokenprevious tokens), based on the probabilities of possible continuations given the current context. Tokens are represented as vectors containing embeddings that capture latent features and rich semantic dependencies learned through extensive training.

We briefly surveyed leading generative approaches, including Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), diffusion models, and autoregressive transformers. Each...

Visually different images
CONTINUE READING
83
Tech Concepts
36
Programming languages
73
Tech Tools
Icon Unlimited access to the largest independent learning library in tech of over 8,000 expert-authored tech books and videos.
Icon Innovative learning tools, including AI book assistants, code context explainers, and text-to-speech.
Icon 50+ new titles added per month and exclusive early access to books as they are being written.
Generative AI Foundations in Python
notes
bookmark Notes and Bookmarks search Search in title playlist Add to playlist download Download options font-size Font size

Change the font size

margin-width Margin width

Change margin width

day-mode Day/Sepia/Night Modes

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY

Submit Your Feedback

Modal Close icon
Modal Close icon
Modal Close icon