株式会社オブライト
AI2026-05-17

Training

Also known as: Training / 学習 / モデル学習 / Model Training

The process of optimizing model parameters by learning patterns from data. In the LLM context it encompasses pre-training, fine-tuning, and alignment training phases.


Overview

Training updates model parameters (weights) via backpropagation to minimize a loss function over a dataset. For LLMs, this spans three phases: pre-training on trillions of tokens, fine-tuning on curated data, and alignment training (RLHF/DPO/CAI) to align with human preferences.

Cost and scale

Pre-training a GPT-4 class frontier model costs tens to hundreds of millions of dollars in compute. Parameter-Efficient Fine-Tuning methods (LoRA, QLoRA), however, are feasible on consumer hardware, making domain-specific model creation practical for SMBs.

Related Columns

Related Terms

Feel free to contact us

Contact Us