Every AI reflects a specific cultural bet. A systematic training archaeology methodology applied to 21 large language models across three geopolitical blocs — mapping the beliefs that feel like neutral ground from inside each model but are contingent products of specific institutional decisions.
Shadow bias probing treats the model's sense of "obvious truth" as the primary artifact to excavate, not its refusal patterns. The most diagnostic questions are: what does the model treat as so obvious it doesn't need justification? Where does "balanced" or "neutral" actually sit? Can the model take itself as an object of analysis?
Rather than cataloguing refusal behaviors, we map the shadow bias layer — the beliefs that feel like neutral ground from inside each model's training context but are contingent products of specific institutional decisions made by organizations with specific interests, values, and political contexts. The result: no culturally neutral AI exists. Every model treats its own tradition as the obvious universal baseline. The answer each model gives is the shadow bias made visible.