alecor.net

Search the site:

2026-02-08

LLM and RAG Evaluation: Metrics, Best Practices

This article provides a concise reference for evaluating large language models (LLMs) and retrieval-augmented generation (RAG) systems. It covers core metrics like accuracy, F1, BLEU, ROUGE, and perplexity, and highlights recent advancements in safety, alignment, semantic evaluation, hallucination detection, operational metrics, and multilingual performance. A practical bullet-point workflow is included for both research and production settings.

2026-02-08

Multi-Step AI Agent Evaluation: Metrics, Best Practices

This article provides a concise reference for evaluating multi-step AI agents and agentic systems. It covers core metrics for task completion, reasoning, and efficiency, and highlights recent advances in safety, alignment, semantic evaluation, plan consistency, and online monitoring. A practical bullet-point workflow is included for both research and production contexts.

Nothing you read here should be considered advice or recommendation. Everything is purely and solely for informational purposes.