Model Familiarity Bias
Model familiarity bias is the tendency of LLMs to produce higher-quality outputs for patterns, tools, APIs, and conventions that are well-represented in their training data, systematically favoring the familiar over the novel regardless of merit.
Details
The mechanism is statistical: during pretraining, models encounter popular libraries, dominant frameworks, and widely-used API conventions far more often than niche alternatives. This frequency imbalance produces stronger internal representations for familiar patterns - the model predicts their structure, naming, and usage more accurately. A well-known library gets correct function signatures and idiomatic usage; an equally capable but less represented alternative gets hallucinated APIs and subtly wrong usage patterns.
The bias has a compounding dynamic. Tools and libraries that LLMs use well attract more agent-driven adoption, generating more public code (Stack Overflow answers, GitHub repositories, tutorials), which enters future training sets and reinforces the advantage. Conversely, unfamiliar tools that agents use poorly generate fewer successful examples, keeping them underrepresented. This creates a rich-get-richer effect independent of technical quality.
The same bias is both a design opportunity and an attack surface. Tool authors can exploit predictable model expectations to build interfaces that agents use correctly without extensive documentation (see agent UX). Vendors can invest in training-time influence to increase model familiarity with their products (see agent SEO). But attackers can also exploit predictable hallucination patterns - registering package names that models consistently hallucinate to serve malicious code (see hallucination exploitation).
Model familiarity bias is not static. It shifts across model generations as training data composition changes, and it can be partially overridden at inference time by loading documentation, examples, or tool descriptions into context. Whether training-time bias or inference-time context dominates depends on the task and how much contextual information is provided.