-
Book Overview & Buying
-
Table Of Contents
Mastering NLP From Foundations to Agents - Second Edition
By :
While LLMs are impressive generalists, their true value in production often comes from their ability to function as domain-specific experts. This chapter is about how to transform a foundation model into such a specialist without updating billions of parameters. It is also about alignment: how to ensure the adapted model not only produces fluent text but also behaves in ways users find helpful, safe, and grounded in sound reasoning.
We begin with parameter-efficient fine-tuning (PEFT), the family of methods that unlock efficient adaptation of very large models. You will learn how LoRA inserts low-rank adapters into key linear projections, and how QLoRA extends this idea by storing the frozen backbone in 4-bit precision while training adapters in fp16 or bf16. But adapting parameters is only part of the story. Models must also be aligned with what people actually want. We therefore connect LoRA and QLoRA to the three major preference...
Change the font size
Change margin width
Change background colour