Book Image

Machine Learning, Data Science and Generative AI with Python [Video]

By : Frank Kane
Book Image

Machine Learning, Data Science and Generative AI with Python [Video]

By: Frank Kane

Overview of this book

This course begins with a Python crash course and then guides you on setting up Microsoft Windows-based PCs, Linux desktops, and Macs. After the setup, we delve into machine learning, AI, and data mining techniques, which include deep learning and neural networks with TensorFlow and Keras; generative models with variational autoencoders and generative adversarial networks; data visualization in Python with Matplotlib and Seaborn; transfer learning, sentiment analysis, image recognition, and classification; regression analysis, K-Means Clustering, Principal Component Analysis, training/testing and cross-validation, Bayesian methods, decision trees, and random forests. Additionally, we will cover multiple regression, multilevel models, support vector machines, reinforcement learning, collaborative filtering, K-Nearest Neighbors, the bias/variance tradeoff, ensemble learning, term frequency/inverse document frequency, experimental design, and A/B testing, feature engineering, hyperparameter tuning, and much more! There's a dedicated section on machine learning with Apache Spark to scale up these techniques to "big data" analyzed on a computing cluster. The course will cover the Transformer architecture, delve into the role of self-attention in AI, explore GPT applications, and practice fine-tuning Transformers for tasks such as movie review analysis. Furthermore, we will look at integrating the OpenAI API for ChatGPT, creating with DALL-E, understanding embeddings, and leveraging audio-to-text to enhance AI with real-world data and moderation.
Table of Contents (15 chapters)
15
You Made It!
Chapter 12
Generative AI: GPT, ChatGPT, Transformers, Self-Attention Based Neural Networks
Content Locked
Section 10
[Activity] Using small and large GPT models within Google CoLab and HuggingFace
This video provides a walkthrough on using GPT-2 with the Hugging Face pipeline for text generation, highlighting the ease of generating multiple variations of text completion. The demonstration includes downloading the model, creating text prompts, and showcasing the varied outputs while emphasizing the model's propensity to fabricate believable but fictitious content.