The LLM Stack
Systems Engineering for Generative AI at Scale
Failed to add items
Sorry, we are unable to add the item because your shopping cart is already at capacity.
Add to Cart failed.
Please try again later
Add to Wish List failed.
Please try again later
Remove from wishlist failed.
Please try again later
Adding to library failed
Please try again
Follow podcast failed
Please try again
Unfollow podcast failed
Please try again
$0.00 for first 30 days
Access a growing selection of included Audible Originals, audiobooks, and podcasts.
You will get an email reminder before your trial ends.
Audible Plus auto-renews for $7.95/mo after 30 days. Upgrade or cancel anytime.
Buy for $6.30
-
Narrated by:
-
Virtual Voice
-
By:
-
Ajit Singh
This title uses virtual voice narration
Virtual voice is computer-generated narration for audiobooks.
Philosophy:
The core philosophy of this book is that Large Language Models are complex, distributed computing problems disguised as mative, focusing on performance, reliability, scalability, and cost-efficiency. My goal is to demystify the entire stack, from the GPU kernel level to the application API, providing readers with a holistic and deep understanding of how production-grade Generative AI systems are built.
Key Features:
1. Globally Relevant: The principles and technologies covered are universal, making the book a valuable resource for students and professionals worldwide.
2. End-to-End Coverage: From understanding the Transformer architecture to managing GPU clusters and monitoring production systems, the book covers the entire lifecycle of an LLM application.
3. Deep Dives: Contains dedicated chapters on cutting-edge topics like inference optimization (Quantization, PagedAttention, Speculative Decoding), serving frameworks (vLLM, TGI), and observability for generative systems.
4. Practical Capstone Project: A complete, step-by-step guide to building a Retrieval-Augmented Generation (RAG) system, one of the most popular and powerful LLM application patterns.
5. Vendor-Agnostic Principles: While we use specific open-source tools for examples, the underlying principles taught are applicable across different cloud providers and proprietary platforms.
To Whom This Book Is For:
1. B.Tech/M.Tech Computer Science Students: Students specializing in AI/ML or Systems Engineering will find this a vital textbook that prepares them for the next wave of computing.
2. Aspiring MLOps/LLMOps Engineers: This is your comprehensive guide to breaking into and excelling in one of the most sought-after roles in the tech industry.
3. Software and DevOps Engineers: Professionals looking to transition into the AI/ML space will find a structured path to understanding the unique infrastructure challenges of Generative AI.
4. Data Scientists and AI Researchers: This book will help you understand the engineering realities of deploying your models and research into production.
Welcome to the world of the LLM Stack!
No reviews yet