Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying LLM Design Patterns
  • Table Of Contents Toc
LLM Design Patterns

LLM Design Patterns

By : Ken Huang
3.5 (2)
close
close
LLM Design Patterns

LLM Design Patterns

3.5 (2)
By: Ken Huang

Overview of this book

This practical guide for AI professionals enables you to build on the power of design patterns to develop robust, scalable, and efficient large language models (LLMs). Written by a global AI expert and popular author driving standards and innovation in Generative AI, security, and strategy, this book covers the end-to-end lifecycle of LLM development and introduces reusable architectural and engineering solutions to common challenges in data handling, model training, evaluation, and deployment. You’ll learn to clean, augment, and annotate large-scale datasets, architect modular training pipelines, and optimize models using hyperparameter tuning, pruning, and quantization. The chapters help you explore regularization, checkpointing, fine-tuning, and advanced prompting methods, such as reason-and-act, as well as implement reflection, multi-step reasoning, and tool use for intelligent task completion. The book also highlights Retrieval-Augmented Generation (RAG), graph-based retrieval, interpretability, fairness, and RLHF, culminating in the creation of agentic LLM systems. By the end of this book, you’ll be equipped with the knowledge and tools to build next-generation LLMs that are adaptable, efficient, safe, and aligned with human values. *Email sign-up and proof of purchase required
Table of Contents (39 chapters)
close
close
Lock Free Chapter
1
Part 1: Introduction and Data Preparation
8
Part 2: Training and Optimization of Large Language Models
16
Part 3: Evaluation and Interpretation of Large Language Models
23
Part 4: Advanced Prompt Engineering Techniques
30
Part 5: Retrieval and Knowledge Integration in Large Language Models

Hyperparameter tuning at scale – challenges and solutions

When tuning hyperparameters for LLMs, we face several challenges:

  • Computational cost: Training LLMs is expensive, limiting the number of trials we can run
  • Long training times: Each trial can take days or weeks, making the entire process very time-consuming
  • Large search space: LLMs have many hyperparameters, creating a vast search space
  • Sensitivity to initialization: LLM performance can vary significantly with different random seeds

To address these challenges, we can employ several strategies:

  • Use smaller proxy tasks: Instead of tuning on the full task, use a smaller dataset or fewer training steps to get a quick estimate of performance
  • Leverage pre-trained models: Start from pre-trained weights and focus on tuning fine-tuning hyperparameters
  • Use multi-fidelity optimization: Start with low-fidelity evaluations (e.g., few training steps) and gradually increase fidelity for promising...
CONTINUE READING
83
Tech Concepts
36
Programming languages
73
Tech Tools
Icon Unlimited access to the largest independent learning library in tech of over 8,000 expert-authored tech books and videos.
Icon Innovative learning tools, including AI book assistants, code context explainers, and text-to-speech.
Icon 50+ new titles added per month and exclusive early access to books as they are being written.
LLM Design Patterns
notes
bookmark Notes and Bookmarks search Search in title playlist Add to playlist font-size Font size

Change the font size

margin-width Margin width

Change margin width

day-mode Day/Sepia/Night Modes

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY

Submit Your Feedback

Modal Close icon
Modal Close icon
Modal Close icon