Context-Value-Action Architecture for Value-Driven Large Language Model Agents
Abstract
Large language models exhibit behavioral rigidity that worsens with intensified reasoning, prompting the development of a Context-Value-Action architecture that decouples action generation from cognitive reasoning using a Value Verifier trained on human data.
Large Language Models (LLMs) have shown promise in simulating human behavior, yet existing agents often exhibit behavioral rigidity, a flaw frequently masked by the self-referential bias of current "LLM-as-a-judge" evaluations. By evaluating against empirical ground truth, we reveal a counter-intuitive phenomenon: increasing the intensity of prompt-driven reasoning does not enhance fidelity but rather exacerbates value polarization, collapsing population diversity. To address this, we propose the Context-Value-Action (CVA) architecture, grounded in the Stimulus-Organism-Response (S-O-R) model and Schwartz's Theory of Basic Human Values. Unlike methods relying on self-verification, CVA decouples action generation from cognitive reasoning via a novel Value Verifier trained on authentic human data to explicitly model dynamic value activation. Experiments on CVABench, which comprises over 1.1 million real-world interaction traces, demonstrate that CVA significantly outperforms baselines. Our approach effectively mitigates polarization while offering superior behavioral fidelity and interpretability.
Community
The core problem of behavioral rigidity in Large Language Model (LLM)-based agents and the motivation for the proposed Context Value Action (CVA) architecture. While LLMs have shown promise in simulating human behavior across various domains—ranging from virtual avatars and social simulacra to embodied Vision Language Action systems and task-oriented assistants existing agents often fail to capture the complexity, diversity, and stochasticity inherent in real human behavior. A critical flaw identified is behavioral rigidity, where agents produce stereotypical or exaggerated responses due to latent model biases amplified by current design paradigms. This issue is frequently masked by evaluation methods that rely on “LLM-as-a-judge” metrics, which suffer from self-referential bias because the judge model shares similar pre-training data and biases with the agent being evaluated. As a result, such evaluations may incorrectly validate polarized or caricatured behaviors as high-quality simulations.
To address this, grounding agent decision-making in established psychological theories, specifically the Stimulus-Organism-Response (S-O-R) model and Schwartz’s Theory of Basic Human Values. They conceptualize human behavior as a dynamic process of value activation influenced by context, rather than a static output based on fixed personas. For example, an individual may generally value self-direction but prioritize hedonism after a tiring workday, seeking relaxation over productivity. Existing agents fail to model this context-dependent value activation, leading to rigid and unrealistic behaviors. The proposed CVA architecture aims to bridge this gap by explicitly decoupling action generation from cognitive reasoning through a novel Value Verifier trained on authentic human data. This verifier models how values are dynamically activated by context and ensures that generated actions align with these activated values, enhancing both behavioral fidelity and interpretability.
Get this paper in your agent:
hf papers read 2604.05939 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper