Safety & Alignment
Bias in LLMs
Quick Answer
Systematic prejudices in model outputs reflecting biases in training data or design.
Bias in LLMs reflects biases in training data. Models can amplify stereotypes and perpetuate discrimination. Bias can be demographic (gender, race) or ideological. Detecting bias is challenging—models might be subtly biased. Reducing bias requires careful data curation and alignment training. Bias is an ongoing concern. Bias can cause real harm. Bias mitigation is important for equity.
Last verified: 2026-04-08