The Granularity Axis: A Micro-to-Macro Latent Direction for Social Roles in Language Models
Abstract
Large language models encode social role granularity as a structured latent dimension that can be manipulated through activation steering, demonstrating consistent patterns across different model architectures and prompting conditions.
Large language models (LLMs) are routinely prompted to take on social roles ranging from individuals to institutions, yet it remains unclear whether their internal representations encode the granularity of such roles, from micro-level individual experience to macro-level organizational, institutional, or national reasoning. We show that they do. We define a contrast-based Granularity Axis as the difference between mean macro- and micro-role hidden states. In Qwen3-8B, this axis aligns with the principal axis (PC1) of the role representation space at cosine 0.972 and accounts for 52.6% of its variance, indicating that granularity is the dominant geometric axis organizing prompted social roles. We construct 75 social roles across five granularity levels and collect 91,200 role-conditioned responses over shared questions and prompt variants, then extract role-level hidden states and project them onto the axis. Role projections increase monotonically across all five levels, remain stable across layers, prompt variants, endpoint definitions, held-out splits, and score-filtered subsets, and transfer to Llama-3.1-8B-Instruct. The axis is also causally relevant: activation steering along it shifts response granularity in the predicted direction, with Llama moving from 2.00 to 3.17 on a five-point macro scale under positive steering on prompts that admit local responses. The two models differ in controllability, suggesting that steering depends on each model's default operating regime. Overall, our findings suggest that social role granularity is not merely a stylistic surface feature, but a structured, ordered, and causally manipulable latent direction in role-conditioned language model behavior.
Community
We identify the Granularity Axis, a micro-to-macro latent direction that organizes social-role representations in language models and partially steers the scale of model reasoning.
that granularity axis as the dominant latent direction in role-conditioned activations is striking, especially its alignment with PC1 and the monotonic projections across five granularity levels. my worry is this could partly reflect dataset confounds rather than granularity per se—macro vs micro prompts might systematically differ in length, formality, or topic, and those covariates could be driving the first principal component. would love to see an ablation where macro and micro prompts are matched on lexical statistics and length to see if the axis still explains the same variance. the causal steering claim is cool but likely model- and prompt-dependent; a more explicit counterfactual or intervention analysis would help map the boundary. btw the arxivlens breakdown helped me parse the method details, e.g. https://arxivlens.com/PaperView/Details/the-granularity-axis-a-micro-to-macro-latent-direction-for-social-roles-in-language-models-96-721147c5
Get this paper in your agent:
hf papers read 2605.06196 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper