← Back to glossary
Glossary

Pre-Training

Reviewed 20 March 2026 Canonical definition

Pre-training is the initial phase of model development where a large language model learns language patterns, knowledge, and reasoning from vast amounts of text data. Pre-training determines a model's base capabilities and embedded biases.