Glossary
LLM-as-Judge
LLM-as-judge is an evaluation technique where a language model scores or ranks another model's outputs. It enables scalable quality assessment but introduces its own biases and requires calibration against human judgments.
LLM-as-judge is an evaluation technique where a language model scores or ranks another model's outputs. It enables scalable quality assessment but introduces its own biases and requires calibration against human judgments.