Transformer Without Magic: How Large Language Models Actually Work Kindle Edition

★★★★★ 4.8 74 reviews

US$90.00
Price when purchased online
Free shipping Free 30-day returns

Sold and shipped by woodspartners.ie
We aim to show you accurate product information. Manufacturers, suppliers and others provide what you see here.
US$90.00
Price when purchased online
Free shipping Free 30-day returns

How do you want your item?
You get 30 days free! Choose a plan at checkout.
Shipping
Arrives May 13
Free
Pickup
Check nearby
Delivery
Not available

Sold and shipped by woodspartners.ie
Free 30-day returns Details

Product details

Management number 220802789 Release Date 2026/05/03 List Price US$90.00 Model Number 220802789
Category

You use ChatGPT, Claude, or Gemini every day — but do you actually understand what's happening under the hood?The transformer is the most consequential software architecture of the decade, powering every major large language model from GPT-4 to LLaMA. Yet most explanations either drown you in linear algebra proofs or hand-wave with vague analogies about "paying attention to words." Neither approach leaves you with real understanding.Transformers Without Magic takes a different path. Starting from first principles and building layer by layer, this book walks you through the complete transformer architecture — from raw text to generated output — with clarity, precision, and zero mysticism. No prerequisites beyond curiosity and a willingness to think carefully.What you'll learn:• How text becomes vectors and why tokenization choices matter• What attention actually computes and why it works so well• How multi-head attention, feed-forward networks, residual connections, and layer normalization fit together• What the residual stream is and why it changes how you think about deep networks• How the output head and sampling strategies turn hidden states into readable text• What the KV cache is and why it's critical for fast inference• How quantization, batching, and the serving stack make LLMs practical at scale• How training works and where the fundamental limitations lieEach chapter builds on the last, giving you a complete mental model of the forward pass — from a single prompt entering the system to a token being generated on the other side. By the end, you won't just know the buzzwords. You'll understand the machinery.Written by Sumeet Kumar, a technologist who has spent years working at the intersection of machine learning and production systems, this book is for engineers, technical leaders, and ambitious learners who refuse to treat AI as a black box.If you want to stop hand-waving and start understanding, scroll up and grab your copy now. Read more

XRay Not Enabled
Language English
File size 5.1 MB
Page Flip Enabled
Word Wise Not Enabled
Print length 287 pages
Accessibility Learn more
Publication date February 8, 2026
Enhanced typesetting Enabled

Correction of product information

If you notice any omissions or errors in the product information on this page, please use the correction request form below.

Correction Request Form

Customer ratings & reviews

4.8 out of 5
★★★★★
74 ratings | 30 reviews
How item rating is calculated
View all reviews
5 stars
87% (64)
4 stars
2% (1)
3 stars
1% (1)
2 stars
0% (0)
1 star
10% (7)
Sort by

There are currently no written reviews for this product.