Papers
arxiv:2605.06196

The Granularity Axis: A Micro-to-Macro Latent Direction for Social Roles in Language Models

Published on May 7
· Submitted by
Chonghan, Qin
on May 8
Authors:
,
,
,
,
,

Abstract

Large language models encode social role granularity as a structured latent dimension that can be manipulated through activation steering, demonstrating consistent patterns across different model architectures and prompting conditions.

AI-generated summary

Large language models (LLMs) are routinely prompted to take on social roles ranging from individuals to institutions, yet it remains unclear whether their internal representations encode the granularity of such roles, from micro-level individual experience to macro-level organizational, institutional, or national reasoning. We show that they do. We define a contrast-based Granularity Axis as the difference between mean macro- and micro-role hidden states. In Qwen3-8B, this axis aligns with the principal axis (PC1) of the role representation space at cosine 0.972 and accounts for 52.6% of its variance, indicating that granularity is the dominant geometric axis organizing prompted social roles. We construct 75 social roles across five granularity levels and collect 91,200 role-conditioned responses over shared questions and prompt variants, then extract role-level hidden states and project them onto the axis. Role projections increase monotonically across all five levels, remain stable across layers, prompt variants, endpoint definitions, held-out splits, and score-filtered subsets, and transfer to Llama-3.1-8B-Instruct. The axis is also causally relevant: activation steering along it shifts response granularity in the predicted direction, with Llama moving from 2.00 to 3.17 on a five-point macro scale under positive steering on prompts that admit local responses. The two models differ in controllability, suggesting that steering depends on each model's default operating regime. Overall, our findings suggest that social role granularity is not merely a stylistic surface feature, but a structured, ordered, and causally manipulable latent direction in role-conditioned language model behavior.

Community

Paper submitter

We identify the Granularity Axis, a micro-to-macro latent direction that organizes social-role representations in language models and partially steers the scale of model reasoning.

that granularity axis as the dominant latent direction in role-conditioned activations is striking, especially its alignment with PC1 and the monotonic projections across five granularity levels. my worry is this could partly reflect dataset confounds rather than granularity per se—macro vs micro prompts might systematically differ in length, formality, or topic, and those covariates could be driving the first principal component. would love to see an ablation where macro and micro prompts are matched on lexical statistics and length to see if the axis still explains the same variance. the causal steering claim is cool but likely model- and prompt-dependent; a more explicit counterfactual or intervention analysis would help map the boundary. btw the arxivlens breakdown helped me parse the method details, e.g. https://arxivlens.com/PaperView/Details/the-granularity-axis-a-micro-to-macro-latent-direction-for-social-roles-in-language-models-96-721147c5

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2605.06196
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2605.06196 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2605.06196 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2605.06196 in a Space README.md to link it from this page.

Collections including this paper 1