BACK_TO_FEEDAICRIER_2
Manning launches Evaluation and Alignment book
OPEN_SOURCE ↗
REDDIT · REDDIT// 24d agoPRODUCT LAUNCH

Manning launches Evaluation and Alignment book

Manning’s new MEAP by Hanchung Lee collects seminal papers on how to evaluate and align AI systems, moving from BLEU and ROUGE to BERTScore, COMET, LLM-as-a-judge, RLHF, constitutional AI, and red teaming. The r/MachineLearning launch also includes a 50% discount code, MLLEE450RE.

// ANALYSIS

Hot take: this is less a book about metrics and more a reminder that most real ML failures are specification failures.

  • It treats evaluation as a design choice, not an afterthought, which is exactly how production teams should think about LLMs.
  • The chapter lineup mirrors the field’s evolution: lexical scoring, semantic similarity, judgment-based evaluation, then alignment loops.
  • The MEAP format makes it a living resource rather than a finished textbook, which suits a fast-moving topic like LLM evaluation.
  • The practitioner focus makes it useful for teams trying to align helpfulness, safety, and consistency around a shared evaluation language.
// TAGS
llmsafetyresearchbenchmarkevaluation-and-alignment

DISCOVERED

24d ago

2026-03-18

PUBLISHED

24d ago

2026-03-18

RELEVANCE

7/ 10

AUTHOR

ManningBooks