Skip to main content

Why Every AI Agent Needs a User Model

AI agents are getting tools, memory, and planning. But they are missing the one thing that makes their actions useful: understanding who they are acting for. Every agent needs a user model.

Robert Ta's Self-Model
Robert Ta's Self-Model CEO & Co-Founder 847 beliefs
· · 10 min read

TL;DR

  • Major AI agent frameworks provide tools, planning, and memory, but none include user modeling as a built-in primitive [1]. Agents can do anything but do not know who they are doing it for.
  • Research shows that injecting user profile context into LLM prompts can improve personalization accuracy by 37% or more [2] compared to raw context. Applying this principle to agent planning reduces irrelevant actions and improves output quality.
  • A growing body of research, including work on human-agent alignment [3] and personalized agents from human feedback [4], points to user modeling as the missing layer in agent architectures.

Every AI agent needs a user model because tools, planning, and memory are insufficient without understanding who the agent is serving. A review of eight major agent frameworks [5] (LangChain, LangGraph, AutoGen, CrewAI, AutoGPT, SuperAGI, AgentGPT, LlamaIndex) found robust tool-calling, orchestration, and memory in all of them, and a dedicated user model concept in none. This post covers why capable-but-unaligned agents waste compute, how user context injection changes agent behavior, and the three user model gaps in current agent architectures.

0/8
agent frameworks reviewed with no user model
0%+
personalization accuracy gain from user profiles (research)
0
dimensions of human-agent alignment identified
0
frameworks with built-in user modeling

The Problem With Capable but Unaligned Agents

An agent without a user model optimizes for task completion. Give it a goal, and it will pursue that goal with whatever tools are available, in whatever way the planning engine determines is efficient.

But task completion and user value are not the same thing. Researchers at IIIT-Hyderabad demonstrated this gap directly: in their CloudOps experiments, baseline metrics reported perfect task completion while a deeper assessment revealed substantial behavioral failures [6], including scenarios where agents achieved 100% tool sequencing accuracy but only 33% policy adherence. The agent “succeeded” by narrow metrics while failing on dimensions that actually mattered.

Consider an agent tasked with “research competitors and summarize findings.” Without a user model, it will produce a comprehensive, detailed report covering every angle. If the user is a VP of Strategy doing a quarterly review, that is exactly right. If the user is a founder who needs a quick comparison for a pitch deck, that comprehensive report is a waste of the user’s time and the agent’s compute.

The task was completed. The user was not served.

A 2024 Google DeepMind study on human-agent alignment [7] identified six dimensions where agents must align with users: knowledge schema, autonomy, operational training, reputational heuristics, ethics, and human engagement. Current agent evaluation frameworks typically measure only one of these (task completion). The other five require understanding who the user is.

Agent Without User Model

  • ×Optimizes for task completion regardless of user context
  • ×Same approach for every user with the same query
  • ×More tools and planning = more actions (often unnecessary)
  • ×Success measured by task completion rate

Agent With User Model

  • Optimizes for user value based on self-model context
  • Adapts approach based on user expertise and preferences
  • Fewer but more relevant actions targeted to user needs
  • Success measured by user alignment and value delivery

What User-Aware Agents Do Differently

The evidence for user-aware agents comes from multiple research threads converging on the same conclusion: user context dramatically changes output quality.

A University of Southern California study [8] found that generating structured user profiles from raw context improved LLM personalization accuracy by 37% on preference prediction tasks. Researchers at University College London [9] showed that user profile placement and composition significantly affect LLM personalization, with historical user responses being the most critical factor. And a 2026 paper on personalized agents from human feedback (PAHF) [10] demonstrated that agents with explicit per-user memory “learn substantially faster and consistently outperform both no-memory and single-channel baselines.”

These findings suggest a clear pattern for agent design: injecting user context before action decisions should produce fewer, more relevant actions.

Fewer actions. When an agent understands user preferences, it can skip steps that are not relevant. The baseline agent, with no user context, errs on the side of comprehensiveness, which is just a polite word for waste.

Better action selection. When choosing between tools, a user-aware agent selects based on user preferences (concise vs. detailed, technical vs. conceptual) rather than purely on task efficiency. A tool that produces better output for this user is better than a tool that produces better output on average.

Appropriate depth. Reports and summaries can match the user’s expertise level. The agent does not explain concepts the user already understands, and it does not skip context the user needs. UCL’s research confirms that personalization information is more critical than semantic similarity [11] in determining output quality.

user-aware-agent.ts
1// Standard agent: plan and executeNo user context
2const plan = await agent.plan(task);
3const result = await agent.execute(plan);
4// Completes the task. May or may not serve the user.
5
6// User-aware agent: understand, then plan, then executeSelf-model injected
7const selfModel = await clarity.getSelfModel(userId);
8
9const plan = await agent.plan(task, {
10 userContext: {
11 expertise: selfModel.beliefs.expertise_level,
12 preferences: selfModel.beliefs.communication_style,
13 goals: selfModel.beliefs.current_objectives,
14 history: selfModel.recentEpisodes(5)
15 }
16});
17// Plan adapts to this user's needs and preferences.
18
19const result = await agent.execute(plan);
20await clarity.observe(userId, { task, result });

The Three User Model Gaps in Agent Frameworks

A review of major agent frameworks reveals three specific gaps where a user model would fundamentally change agent behavior. Codecademy’s 2025 framework comparison [12] lists tool integrations, orchestration layers, memory/state management, and observability as the core architectural components. User modeling does not appear.

Gap 1: Action Prioritization. Current agents prioritize actions based on task decomposition: what needs to happen to complete the task. With a user model, prioritization shifts to value decomposition: what actions deliver the most value to this specific user. Sometimes the most valuable action is skipping a step the user does not need.

Gap 2: Output Calibration. Agents produce outputs (reports, code, summaries, recommendations) calibrated to the task specification. With a user model, outputs are calibrated to the user: their expertise level, communication preferences, and decision-making style. Same information, different delivery, dramatically different value. Recent work on a unified user modeling language [13] proposes standardizing these dimensions (accessibility constraints, interaction preferences, domain knowledge) so that AI systems can consume them consistently.

Gap 3: Proactive Relevance. Current agents are reactive. They do what they are told. A user-aware agent can be proactively relevant: noticing that the task the user requested is similar to one they did last week and adjusting the approach accordingly, or flagging when the task result contradicts something the user previously expressed as a belief. The PAHF framework [14] demonstrates this with its three-step loop of pre-action clarification, preference-grounded action, and post-action feedback.

Agent CapabilityWithout User ModelWith User Model
PlanningDecompose task into stepsDecompose task into user-relevant steps
Tool selectionChoose most efficient toolChoose tool that best serves this user
Output formatStandard format for all usersCalibrated to user expertise and preferences
Action strategyComprehensive (err on the side of doing more)Targeted (skip what this user does not need)
Proactive behaviorNoneContext-aware suggestions based on user history

Multi-Agent Systems Need Shared User Models

The problem compounds in multi-agent architectures. When multiple agents collaborate on a task (a researcher agent, an analyst agent, a writer agent) they each operate with their own view of the task. Without a shared user model, the researcher gathers information at the wrong depth, the analyst applies the wrong framework, and the writer uses the wrong tone.

Anthropic’s engineering team encountered a version of this coordination problem when building their multi-agent research system [15]. Without detailed task descriptions, agents “duplicate work, leave gaps, or fail to find necessary information.” Their solution was detailed prompting and explicit coordination rules. A shared user model could serve a similar role: providing every agent in the system with the same understanding of who they are serving, what level of detail is appropriate, and what the user actually cares about.

The user model becomes the alignment layer across the entire multi-agent system. Without it, coordinating agent output quality depends on increasingly complex orchestration logic that still lacks the most important input: who this work is for.

The Agent Framework That Wins

The agent framework that dominates the next era will likely not be the one with the most tools, the most sophisticated planning engine, or the best memory system. It will be the one that makes user modeling a first-class primitive.

A comprehensive survey of personalized LLMs [16] identifies three technical levels where personalization can be applied: prompting (input level), adaptation (model level), and alignment (objective level). Current agent frameworks operate almost entirely at the prompting and adaptation levels for task context, but none apply these techniques systematically for user context. The research community is beginning to fill this gap, but the framework ecosystem has not caught up.

The tools are commoditizing. Planning is commoditizing. Memory is commoditizing. User understanding is inherently individual, making it a durable differentiator.

0%+
personalization accuracy gain from structured profiles
0
levels of LLM personalization (prompting, adaptation, alignment)
0
dimensions of human-agent alignment (Google DeepMind)

Trade-offs

Adding user models to agent frameworks is not free.

Latency increases. Querying the user model before every action decision adds latency. For agents that need to act in real time, this overhead matters. Caching and preloading strategies can mitigate but not eliminate this.

Complexity increases. The agent now has an additional input (user context) that influences every decision. This makes the planning space larger and debugging harder. When an agent makes a wrong decision, you need to determine whether the error was in the planning, the tools, or the user model. Anthropic’s engineering team notes that debugging non-deterministic agent behavior [17] is already one of the hardest challenges in multi-agent systems. Adding user context expands this difficulty.

Privacy surfaces expand. User models contain personal information that flows through every agent action. A New America analysis of AI agents and memory [18] warns that agent memory creates “distributed, persistent, and interoperable” data flows spanning multiple services, making traditional privacy frameworks inadequate. Each tool the agent calls potentially receives user context, which requires careful scoping and audit.

User model quality becomes critical. A wrong user model is worse than no user model. If the agent thinks the user is a beginner when they are an expert, every action will be misaligned in a way that feels condescending. The PAHF researchers address this with a dual feedback mechanism [19] (pre-action clarification and post-action feedback) that continuously corrects the model. Confidence thresholds and graceful degradation are essential.

What to Do Next

  1. Audit your agent’s action relevance. For your next 20 agent task completions, have the user rate each individual action on a 1-5 relevance scale. Average the score. If it is below 4, your agent is doing unnecessary work, and a user model would tell it what to skip. The Beyond Task Completion framework [20] provides a structured approach for this kind of multi-dimensional agent evaluation.

  2. Add user context to the planning prompt. Even without a full self-model, injecting basic user context (expertise level, preferred style, current project) into the agent’s planning prompt improves relevance. Research shows user profile placement at the beginning of the input context has the greatest effect [21]. Start with a simple user profile and measure the difference. Clarity provides the self-model API that agents can query for rich user context.

  3. Measure actions per task, not just task completion. The best agent is not the one that completes the most tasks. It is the one that completes tasks with the fewest unnecessary actions. Track the ratio of actions to user-rated value. A lower ratio means a more aligned agent.


Tools make agents capable. User models make agents useful. Give your agents the context they need.

References

  1. none include user modeling as a built-in primitive
  2. improve personalization accuracy by 37% or more
  3. human-agent alignment
  4. personalized agents from human feedback
  5. eight major agent frameworks
  6. baseline metrics reported perfect task completion while a deeper assessment revealed substantial behavioral failures
  7. 2024 Google DeepMind study on human-agent alignment
  8. University of Southern California study
  9. University College London
  10. 2026 paper on personalized agents from human feedback (PAHF)
  11. personalization information is more critical than semantic similarity
  12. Codecademy’s 2025 framework comparison
  13. unified user modeling language
  14. PAHF framework
  15. multi-agent research system
  16. comprehensive survey of personalized LLMs
  17. notes that debugging non-deterministic agent behavior
  18. New America analysis of AI agents and memory
  19. dual feedback mechanism
  20. Beyond Task Completion framework
  21. user profile placement at the beginning of the input context has the greatest effect

Building AI that needs to understand its users?

Talk to us →
The Clarity Mirror

What did this article change about what you believe?

Select your beliefs

After reading this, which resonate with you?

Stay sharp on AI personalization

Daily insights and research on AI personalization and context management at scale. Read by hundreds of AI builders.

Daily articles on AI-native products. Unsubscribe anytime.

Robert Ta

We build in public. Get Robert's weekly newsletter on building better AI products with Clarity, with a focus on hyper-personalization and digital twin technology. Join 1500+ founders and builders at Self Aligned.

Subscribe to Self Aligned →