Why Every AI Agent Needs a User Model
AI agents are getting tools, memory, and planning. But they are missing the one thing that makes their actions useful: understanding who they are acting for. Every agent needs a user model.
TL;DR
- Major AI agent frameworks provide tools, planning, and memory, but none include user modeling as a built-in primitive [1]. Agents can do anything but do not know who they are doing it for.
- Research shows that injecting user profile context into LLM prompts can improve personalization accuracy by 37% or more [2] compared to raw context. Applying this principle to agent planning reduces irrelevant actions and improves output quality.
- A growing body of research, including work on human-agent alignment [3] and personalized agents from human feedback [4], points to user modeling as the missing layer in agent architectures.
Every AI agent needs a user model because tools, planning, and memory are insufficient without understanding who the agent is serving. A review of eight major agent frameworks [5] (LangChain, LangGraph, AutoGen, CrewAI, AutoGPT, SuperAGI, AgentGPT, LlamaIndex) found robust tool-calling, orchestration, and memory in all of them, and a dedicated user model concept in none. This post covers why capable-but-unaligned agents waste compute, how user context injection changes agent behavior, and the three user model gaps in current agent architectures.
The Problem With Capable but Unaligned Agents
An agent without a user model optimizes for task completion. Give it a goal, and it will pursue that goal with whatever tools are available, in whatever way the planning engine determines is efficient.
But task completion and user value are not the same thing. Researchers at IIIT-Hyderabad demonstrated this gap directly: in their CloudOps experiments, baseline metrics reported perfect task completion while a deeper assessment revealed substantial behavioral failures [6], including scenarios where agents achieved 100% tool sequencing accuracy but only 33% policy adherence. The agent “succeeded” by narrow metrics while failing on dimensions that actually mattered.
Consider an agent tasked with “research competitors and summarize findings.” Without a user model, it will produce a comprehensive, detailed report covering every angle. If the user is a VP of Strategy doing a quarterly review, that is exactly right. If the user is a founder who needs a quick comparison for a pitch deck, that comprehensive report is a waste of the user’s time and the agent’s compute.
The task was completed. The user was not served.
A 2024 Google DeepMind study on human-agent alignment [7] identified six dimensions where agents must align with users: knowledge schema, autonomy, operational training, reputational heuristics, ethics, and human engagement. Current agent evaluation frameworks typically measure only one of these (task completion). The other five require understanding who the user is.
Agent Without User Model
- ×Optimizes for task completion regardless of user context
- ×Same approach for every user with the same query
- ×More tools and planning = more actions (often unnecessary)
- ×Success measured by task completion rate
Agent With User Model
- ✓Optimizes for user value based on self-model context
- ✓Adapts approach based on user expertise and preferences
- ✓Fewer but more relevant actions targeted to user needs
- ✓Success measured by user alignment and value delivery
What User-Aware Agents Do Differently
The evidence for user-aware agents comes from multiple research threads converging on the same conclusion: user context dramatically changes output quality.
A University of Southern California study [8] found that generating structured user profiles from raw context improved LLM personalization accuracy by 37% on preference prediction tasks. Researchers at University College London [9] showed that user profile placement and composition significantly affect LLM personalization, with historical user responses being the most critical factor. And a 2026 paper on personalized agents from human feedback (PAHF) [10] demonstrated that agents with explicit per-user memory “learn substantially faster and consistently outperform both no-memory and single-channel baselines.”
These findings suggest a clear pattern for agent design: injecting user context before action decisions should produce fewer, more relevant actions.
Fewer actions. When an agent understands user preferences, it can skip steps that are not relevant. The baseline agent, with no user context, errs on the side of comprehensiveness, which is just a polite word for waste.
Better action selection. When choosing between tools, a user-aware agent selects based on user preferences (concise vs. detailed, technical vs. conceptual) rather than purely on task efficiency. A tool that produces better output for this user is better than a tool that produces better output on average.
Appropriate depth. Reports and summaries can match the user’s expertise level. The agent does not explain concepts the user already understands, and it does not skip context the user needs. UCL’s research confirms that personalization information is more critical than semantic similarity [11] in determining output quality.
1// Standard agent: plan and execute← No user context2const plan = await agent.plan(task);3const result = await agent.execute(plan);4// Completes the task. May or may not serve the user.56// User-aware agent: understand, then plan, then execute← Self-model injected7const selfModel = await clarity.getSelfModel(userId);89const plan = await agent.plan(task, {10userContext: {11expertise: selfModel.beliefs.expertise_level,12preferences: selfModel.beliefs.communication_style,13goals: selfModel.beliefs.current_objectives,14history: selfModel.recentEpisodes(5)15}16});17// Plan adapts to this user's needs and preferences.1819const result = await agent.execute(plan);20await clarity.observe(userId, { task, result });
The Three User Model Gaps in Agent Frameworks
A review of major agent frameworks reveals three specific gaps where a user model would fundamentally change agent behavior. Codecademy’s 2025 framework comparison [12] lists tool integrations, orchestration layers, memory/state management, and observability as the core architectural components. User modeling does not appear.
Gap 1: Action Prioritization. Current agents prioritize actions based on task decomposition: what needs to happen to complete the task. With a user model, prioritization shifts to value decomposition: what actions deliver the most value to this specific user. Sometimes the most valuable action is skipping a step the user does not need.
Gap 2: Output Calibration. Agents produce outputs (reports, code, summaries, recommendations) calibrated to the task specification. With a user model, outputs are calibrated to the user: their expertise level, communication preferences, and decision-making style. Same information, different delivery, dramatically different value. Recent work on a unified user modeling language [13] proposes standardizing these dimensions (accessibility constraints, interaction preferences, domain knowledge) so that AI systems can consume them consistently.
Gap 3: Proactive Relevance. Current agents are reactive. They do what they are told. A user-aware agent can be proactively relevant: noticing that the task the user requested is similar to one they did last week and adjusting the approach accordingly, or flagging when the task result contradicts something the user previously expressed as a belief. The PAHF framework [14] demonstrates this with its three-step loop of pre-action clarification, preference-grounded action, and post-action feedback.
| Agent Capability | Without User Model | With User Model |
|---|---|---|
| Planning | Decompose task into steps | Decompose task into user-relevant steps |
| Tool selection | Choose most efficient tool | Choose tool that best serves this user |
| Output format | Standard format for all users | Calibrated to user expertise and preferences |
| Action strategy | Comprehensive (err on the side of doing more) | Targeted (skip what this user does not need) |
| Proactive behavior | None | Context-aware suggestions based on user history |
Multi-Agent Systems Need Shared User Models
The problem compounds in multi-agent architectures. When multiple agents collaborate on a task (a researcher agent, an analyst agent, a writer agent) they each operate with their own view of the task. Without a shared user model, the researcher gathers information at the wrong depth, the analyst applies the wrong framework, and the writer uses the wrong tone.
Anthropic’s engineering team encountered a version of this coordination problem when building their multi-agent research system [15]. Without detailed task descriptions, agents “duplicate work, leave gaps, or fail to find necessary information.” Their solution was detailed prompting and explicit coordination rules. A shared user model could serve a similar role: providing every agent in the system with the same understanding of who they are serving, what level of detail is appropriate, and what the user actually cares about.
The user model becomes the alignment layer across the entire multi-agent system. Without it, coordinating agent output quality depends on increasingly complex orchestration logic that still lacks the most important input: who this work is for.
The Agent Framework That Wins
The agent framework that dominates the next era will likely not be the one with the most tools, the most sophisticated planning engine, or the best memory system. It will be the one that makes user modeling a first-class primitive.
A comprehensive survey of personalized LLMs [16] identifies three technical levels where personalization can be applied: prompting (input level), adaptation (model level), and alignment (objective level). Current agent frameworks operate almost entirely at the prompting and adaptation levels for task context, but none apply these techniques systematically for user context. The research community is beginning to fill this gap, but the framework ecosystem has not caught up.
The tools are commoditizing. Planning is commoditizing. Memory is commoditizing. User understanding is inherently individual, making it a durable differentiator.
Trade-offs
Adding user models to agent frameworks is not free.
Latency increases. Querying the user model before every action decision adds latency. For agents that need to act in real time, this overhead matters. Caching and preloading strategies can mitigate but not eliminate this.
Complexity increases. The agent now has an additional input (user context) that influences every decision. This makes the planning space larger and debugging harder. When an agent makes a wrong decision, you need to determine whether the error was in the planning, the tools, or the user model. Anthropic’s engineering team notes that debugging non-deterministic agent behavior [17] is already one of the hardest challenges in multi-agent systems. Adding user context expands this difficulty.
Privacy surfaces expand. User models contain personal information that flows through every agent action. A New America analysis of AI agents and memory [18] warns that agent memory creates “distributed, persistent, and interoperable” data flows spanning multiple services, making traditional privacy frameworks inadequate. Each tool the agent calls potentially receives user context, which requires careful scoping and audit.
User model quality becomes critical. A wrong user model is worse than no user model. If the agent thinks the user is a beginner when they are an expert, every action will be misaligned in a way that feels condescending. The PAHF researchers address this with a dual feedback mechanism [19] (pre-action clarification and post-action feedback) that continuously corrects the model. Confidence thresholds and graceful degradation are essential.
What to Do Next
-
Audit your agent’s action relevance. For your next 20 agent task completions, have the user rate each individual action on a 1-5 relevance scale. Average the score. If it is below 4, your agent is doing unnecessary work, and a user model would tell it what to skip. The Beyond Task Completion framework [20] provides a structured approach for this kind of multi-dimensional agent evaluation.
-
Add user context to the planning prompt. Even without a full self-model, injecting basic user context (expertise level, preferred style, current project) into the agent’s planning prompt improves relevance. Research shows user profile placement at the beginning of the input context has the greatest effect [21]. Start with a simple user profile and measure the difference. Clarity provides the self-model API that agents can query for rich user context.
-
Measure actions per task, not just task completion. The best agent is not the one that completes the most tasks. It is the one that completes tasks with the fewest unnecessary actions. Track the ratio of actions to user-rated value. A lower ratio means a more aligned agent.
Tools make agents capable. User models make agents useful. Give your agents the context they need.
References
- none include user modeling as a built-in primitive
- improve personalization accuracy by 37% or more
- human-agent alignment
- personalized agents from human feedback
- eight major agent frameworks
- baseline metrics reported perfect task completion while a deeper assessment revealed substantial behavioral failures
- 2024 Google DeepMind study on human-agent alignment
- University of Southern California study
- University College London
- 2026 paper on personalized agents from human feedback (PAHF)
- personalization information is more critical than semantic similarity
- Codecademy’s 2025 framework comparison
- unified user modeling language
- PAHF framework
- multi-agent research system
- comprehensive survey of personalized LLMs
- notes that debugging non-deterministic agent behavior
- New America analysis of AI agents and memory
- dual feedback mechanism
- Beyond Task Completion framework
- user profile placement at the beginning of the input context has the greatest effect
Related
Building AI that needs to understand its users?
What did this article change about what you believe?
Select your beliefs
After reading this, which resonate with you?
Stay sharp on AI personalization
Daily insights and research on AI personalization and context management at scale. Read by hundreds of AI builders.
Daily articles on AI-native products. Unsubscribe anytime.
We build in public. Get Robert's weekly newsletter on building better AI products with Clarity, with a focus on hyper-personalization and digital twin technology. Join 1500+ founders and builders at Self Aligned.
Subscribe to Self Aligned →