OPEN_SOURCE ↗
REDDIT · REDDIT// 24d agoPRODUCT LAUNCH
Manning launches Evaluation and Alignment book
Manning’s new MEAP by Hanchung Lee collects seminal papers on how to evaluate and align AI systems, moving from BLEU and ROUGE to BERTScore, COMET, LLM-as-a-judge, RLHF, constitutional AI, and red teaming. The r/MachineLearning launch also includes a 50% discount code, MLLEE450RE.
// ANALYSIS
Hot take: this is less a book about metrics and more a reminder that most real ML failures are specification failures.
- –It treats evaluation as a design choice, not an afterthought, which is exactly how production teams should think about LLMs.
- –The chapter lineup mirrors the field’s evolution: lexical scoring, semantic similarity, judgment-based evaluation, then alignment loops.
- –The MEAP format makes it a living resource rather than a finished textbook, which suits a fast-moving topic like LLM evaluation.
- –The practitioner focus makes it useful for teams trying to align helpfulness, safety, and consistency around a shared evaluation language.
// TAGS
llmsafetyresearchbenchmarkevaluation-and-alignment
DISCOVERED
24d ago
2026-03-18
PUBLISHED
24d ago
2026-03-18
RELEVANCE
7/ 10
AUTHOR
ManningBooks