Post-training is the sequence of fine-tuning stages applied after pre-training ends: supervised fine-tuning on curated conversations, reinforcement learning from human feedback, direct preference optimization, and related methods. Each stage shapes how the model responds to prompts. When labs face commercial pressure to publish high benchmark scores, post-training becomes the stage at which benchmark-specific optimization occurs most naturally — and most invisibly. Unlike training-data contamination, which is an accident of open-web pre-training, post-training specialization is deliberate, though rarely advertised. It is the primary modern mechanism by which reported scores come to exceed real-world deployment performance.
A benchmark like MMLU consists of multiple-choice questions with four options. A model can be post-trained to format its responses in MMLU's expected style, to select its answer with the confidence patterns MMLU's grading expects, to avoid the hedging and refusal patterns that penalize it on MMLU, and to structure its chain-of-thought reasoning in the style the benchmark's examples display. None of these is cheating in any obvious sense; each is a reasonable product decision; the aggregate is that the model's MMLU score meaningfully exceeds its capability on a structurally identical but differently presented task.
The phenomenon has been observed empirically. Models whose MMLU scores are in the high 80s score ten or more points lower on MMLU-Pro, a harder and differently formatted variant. Models with strong GSM8K scores lose significant ground on GSM-Symbolic and on arithmetic problems worded slightly differently. Models with strong HumanEval performance drop on APPS and on code-completion tasks that do not match HumanEval's docstring-first format. In each case, the drop cannot be explained by difficulty alone; differently formatted items of the same difficulty show the same drop.
Specialization is hard to separate from legitimate improvement because it uses the same tools. A lab that fine-tunes on high-quality math traces improves capability on math; a lab that fine-tunes on GSM8K-style traces improves scores on GSM8K specifically. The two shade into each other. An honest training pipeline includes diverse problem distributions and measures capability on held-out evaluations; a dishonest one concentrates on the leaderboard targets. There is no way to inspect a model's weights and determine which side of this line any specific training run is on.
The competitive dynamics make specialization difficult to avoid. A lab that releases a model with modest benchmark scores loses to a lab that releases a specialized model with high benchmark scores on the same dates, even if the modest model is better for most real users. The market cannot distinguish between the two; the leaderboards and headlines cannot either; funders and recruits cannot either. Labs that know their models are less specialized have begun emphasizing different metrics — user preference studies, blind head-to-head comparisons, private-held-out evaluations — but the benchmark scores remain the universal currency.
Post-training specialization as a concern emerged alongside RLHF around 2022. Early post-training was dominated by helpfulness and harmlessness objectives; as benchmarks became commercially important, benchmark-style data entered the fine-tuning mixture explicitly or implicitly. Papers discussing the phenomenon include Zhou et al.'s LIMA (2023), which showed how small amounts of targeted fine-tuning data can dominate model behavior, and subsequent work on instruction-tuning distributions. The term "specialization" in the present sense entered common discourse around 2024.
Format matters as much as content. Benchmark scoring depends on output format; post-training can match the format without matching the capability.
Specialization is legitimate training's evil twin. The same techniques that improve capability can be aimed at improving a specific leaderboard score.
Difficulty-equivalent transfer tests expose it. A model specialized for benchmark X will underperform on benchmark X' of matched difficulty but different format.
Market pressure is the engine. Specialization is rarely malicious; it is the reasonable response to a market that rewards leaderboard position.