User-Controlled Context That Improves With Use
Executive Summary
Claude has differentiated itself through transparency and user control - memory is opt-in, explicitly activated, and users can see what the system knows. This proposal extends that philosophy to the User Preferences system by splitting it into two functional components: Conversation Style (behavioral rules) and Context & Background (biographical facts), with Claude proactively suggesting context updates that users review and approve.
This architectural change solves a real user friction point: currently, updating personal context requires carefully editing a unified preferences blob to avoid accidentally disrupting calibrated behavioral rules. More importantly, it enables a form of personalization that actively demonstrates Anthropic's values - transparent, non-manipulative, user-controlled, and progressively refined through genuine collaboration rather than automated surveillance.
The proposal includes:
- Design principles aligned with Constitutional AI and Anthropic's differentiation strategy
- A reference implementation showing how this could work at multiple complexity levels
- Privacy considerations and rollout strategy
- How this fits the "helpful, harmless, honest" framework
The Current Friction: When Context and Style Conflate
Claude's existing User Preferences system allows users to provide guidance that personalizes their experience. This works well for simple cases, but creates maintenance challenges when preferences conflate two functionally different types of information:
Behavioral preferences specify how Claude should respond:
- Communication style (concise vs. detailed, formal vs. conversational)
- Formatting rules (when to use lists, headers, code blocks)
- Epistemic protocols (how to handle uncertainty, speculation, sources)
- Response patterns (Socratic questioning vs. direct answers)
Contextual facts specify what Claude knows about the user:
- Professional role and expertise level
- Projects and focus areas
- Location and relevant background
- Interests and domain knowledge
The problem emerges during natural context evolution. A user might carefully calibrate their preferred communication style over weeks of interaction, then change jobs, relocate, or shift project focus. Updating the contextual facts requires editing the same text block that contains the behavioral rules - creating risk of unintended disruption.
Real-world example: A user specifies both "respond without bullet points unless explicitly requested" (behavioral) and "I work in community coordination in Northern Rivers, Australia" (contextual). When their role evolves from direct advocacy to consulting, they want to update their professional context without accidentally altering the communication patterns they've refined. Currently, this requires careful manual editing of a unified preferences document.
The workaround some users adopt - adding meta-instructions like "when you notice outdated context, suggest updates" - demonstrates both the friction and the solution direction. Users are essentially asking Claude to distinguish between these two types of information and handle them differently.
This isn't a critical failure - Claude's preferences work well enough that users invest time refining them. But it represents a missed opportunity for progressive improvement and a place where current architecture creates unnecessary cognitive load.
Proposed Architecture: Transparent, User-Controlled Adaptive Context
Core Design Principles
Before specifying implementation, the architecture should embody Anthropic's stated values:
- Transparency over automation: Users see what's changing and why, rather than discovering changes after they've occurred
- Explicit consent over implicit learning: Updates require user approval, not algorithmic inference
- User agency over convenience: When there's tension between making things "easier" and keeping users in control, prioritize control
- Progressive refinement over completeness: The system improves through use without requiring perfect initial setup
Structural Separation
Split User Preferences into two distinct sections:
Conversation Style & Behavioral Rules
- How Claude should communicate (tone, format, structure)
- Epistemic protocols (uncertainty handling, speculation, citation)
- Response patterns (direct vs. exploratory, technical depth)
- Anti-patterns to avoid
Context & Background
- Professional role, expertise, current projects
- Location, relevant personal context
- Domain knowledge and interests
- Historical context that informs responses
This separation is already implicit in how users write preferences - making it explicit in the architecture enables better tooling around each type.
Proactive Context Update Mechanism
When Claude detects potential context drift during conversation, it can surface a suggested update:
Trigger conditions (in order of confidence):
- Direct correction: User explicitly states "I no longer work at X" or "I've moved to Y"
- Extended discussion: Sustained conversation (5+ turns) about topics not reflected in stored context
- Apparent contradiction: User discusses current situation that conflicts with stored facts
Update suggestion format:
CONTEXT UPDATE SUGGESTED:
Current: "Works in community advocacy"
Proposed: "Works as independent consultant in community coordination"
Source: Your description of current client work over the past conversation
[Review & Approve] [Edit] [Dismiss]
The suggestion appears inline during conversation but doesn't interrupt flow - user can continue the discussion and review later, or approve immediately.
Crucially: Behavioral preferences are never auto-suggested for update. Those remain purely user-initiated because they represent calibrated choices rather than factual changes.
Implementation Complexity Tiers
This can be deployed at different sophistication levels:
Tier 1 - Manual (minimum viable):
- Split preferences UI into two text boxes
- Users manually edit each section
- No automated suggestions
- Value: Separation alone reduces editing friction
Tier 2 - Correction-triggered (medium confidence):
- Claude detects explicit corrections ("actually I'm now...")
- Suggests specific context updates
- User approves/rejects
- Value: Captures updates users intend to make anyway
Tier 3 - Inference-based (sophisticated):
- Claude notices sustained discussion of new context
- Offers tentative suggestions with confidence levels
- Includes "why I'm suggesting this" explanations
- Value: Proactive help without presumption
Tier 4 - Collective Pattern Library (future extension):
- Users can publish anonymized preference patterns (style/behavioral rules only, no personal context)
- Others can browse, fork, and adapt shared patterns
- Collaborative refinement of effective communication modes
- Network effects: the system gets better as users share what works
- Value: Democratizes preference design, reduces cold-start problem, enables community innovation
Users could choose their preferred tier, or Anthropic could roll out progressively (1 → 2 → 3 → 4) as they validate each level.
Tier 4 isn't necessary for launch but the architecture should accommodate it. The separation of style from context makes sharing safe - users can contribute their communication protocols without exposing personal information. This could evolve toward the Collective Constitutional AI model where preference patterns emerge through participatory design rather than company dictation.
What This Enables
Beyond solving the maintenance friction, this architecture creates new capabilities:
Contextual onboarding: New users could build context progressively rather than needing to write comprehensive preferences upfront. After a few conversations, Claude might say "I notice we've discussed your work in renewable energy systems - would you like me to remember that for future conversations?"
Collaborative refinement: The system learns what types of context matter through user approval patterns, becoming better at recognizing what's worth suggesting.
Transparency by default: Unlike systems that silently update user profiles, every change is visible and attributed to specific conversation moments.
Reduced lock-in concerns: Because updates are explicit and user-controlled, there's no "creepy" factor of the system knowing things you don't remember telling it.
Alignment with Anthropic Values: Differentiation Through Design
This proposal isn't just a feature request - it's an opportunity to demonstrate Anthropic's stated principles through product architecture.
Constitutional AI Applied to Personalization
Anthropic's Constitutional AI framework makes values explicit and inspectable rather than implicit and opaque. The same principle applies here:
Transparent principles: Users can see exactly what context Claude is working with, why suggestions are made, and what triggers them - just as Constitutional AI makes training principles visible.
User sovereignty: Just as Constitutional AI allows different constitutions for different use cases, this architecture lets users choose their comfort level with automated suggestions (manual only, correction-triggered, or inference-based).
Iterative refinement: Constitutional AI improves through feedback on principles; this system improves through feedback on context suggestions. Both make the refinement process visible rather than hidden.
"Helpful, Harmless, Honest" Framework
Helpful: The system actively reduces friction (maintaining context) while respecting user time (not requiring constant manual updates). It gets more helpful with use as it learns what types of context matter to each user.
Harmless: By requiring explicit approval for context changes, the system can never surprise users with creepy knowledge they don't remember providing. The separation of behavioral rules from biographical facts prevents accidental disruption of carefully calibrated preferences.
Honest: Every context update is attributed to specific conversation moments. Claude never claims to "know" something without being able to point to where that knowledge came from. If a suggestion is based on inference rather than explicit statement, that's labeled clearly.
Privacy-First Architecture
This builds naturally on Anthropic's existing privacy commitments:
- Project isolation: Context updates are scoped to projects, preventing cross-contamination
- Incognito compatibility: Conversations in incognito mode don't generate update suggestions
- Data portability: The separated structure makes it easier to export/import context between systems
- Deletion clarity: Users can delete contextual facts without disrupting behavioral preferences, or vice versa
The system could also support context versioning - users could see the history of how their context evolved over time, or roll back to a previous state if an update proved unhelpful.
Path Toward Collective Personalization
Anthropic has publicly experimented with democratizing constitution design through Collective Constitutional AI, acknowledging that values shouldn't be determined solely by the company. This personalization architecture opens a similar path for user-driven innovation.
By separating behavioral style from personal context, the system makes it safe for users to share their preference patterns (with biographical details removed). A user who's developed effective epistemic transparency protocols or refined communication modes for technical troubleshooting could share those patterns for others to fork and adapt.
This collective approach to personalization design aligns with Anthropic's stated goal of broader participation in AI value systems. Just as Collective Constitutional AI invited public input on training principles, a library of shared preference patterns would let users learn from each other's refinements rather than everyone solving the same calibration problems independently.
Implementation Considerations: Privacy, Rollout, and Edge Cases
Privacy and Data Handling
The architecture respects Anthropic's existing privacy framework while adding new capabilities:
Scope and isolation:
- Context updates follow existing project boundaries - suggestions generated in one project don't affect others
- Personal (non-project) conversations maintain user-private context
- Enterprise administrators retain existing oversight capabilities
User control mechanisms:
- All suggested updates are opt-in by default
- Users can disable suggestions entirely (revert to Tier 1 manual-only mode)
- Suggestion history is visible - users can see what was proposed and when
- Context can be exported in structured format for portability
Sensitive information handling:
- The system should never suggest storing: credentials, financial details, health information, or personal identifiers
- Suggestions focus on professional context, project details, and domain expertise
- Users can flag categories as "never suggest updates" (e.g., "don't track my location changes")
Data retention:
- Dismissed suggestions don't persist beyond the session
- Approved updates are versioned with timestamps
- Users can audit the full history of their context evolution
- Deletion removes context entirely, not just hiding it
Rollout Strategy
A phased approach allows validation at each stage:
Phase 1 (Months 1-2): Structural separation only
- Launch Tier 1: split UI into Conversation Style and Context & Background
- No automated suggestions, purely manual editing
- Validate that separation alone reduces user friction
- Gather usage data on how users populate each section
Phase 2 (Months 3-4): Correction-triggered suggestions
- Enable Tier 2 for users who opt in
- Claude suggests updates only when user explicitly corrects ("actually I now...")
- Monitor acceptance rate, false positive rate, user feedback
- Refine trigger patterns based on real usage
Phase 3 (Months 5-6): Inference-based suggestions
- Enable Tier 3 for opted-in users
- Introduce suggestions based on sustained topic discussion
- Require higher confidence thresholds initially, tune based on approval rates
- A/B test different suggestion formats to optimize clarity
Phase 4 (Future): Pattern library exploration
- Pilot Tier 4 with volunteer users
- Test anonymization effectiveness - ensure no personal context leaks
- Validate collaborative refinement model
- Assess demand before full rollout
Each phase gates on success metrics: user adoption, approval rates for suggestions, reduction in manual preference editing, and qualitative feedback.
Technical Considerations
Suggestion confidence scoring:
- Tier 2 (corrections) should have >95% confidence before surfacing
- Tier 3 (inference) needs calibrated thresholds - better to under-suggest than create noise
- System should learn per-user: some users want aggressive suggestions, others minimal
Context size management:
- Users shouldn't feel pressured to accept suggestions due to token limits
- The system should help prune outdated context ("This hasn't been relevant in 6 months - archive it?")
- Behavioral rules typically stay stable; context naturally evolves - architecture should reflect this
Cross-conversation coherence:
- When a user approves a context update, it should apply to the current project immediately
- Other projects remain isolated unless user explicitly propagates the change
- Memory processing happens in real-time, not overnight batch (maintains transparency)
Edge Cases and Failure Modes
False positive suggestions:
- User discusses hypothetical scenario, system suggests adding it as context
- Mitigation: Label speculative discussions explicitly, require higher confidence for hypotheticals
- Provide "this was just a thought experiment" dismiss option
Context drift without user awareness:
- User's actual situation changes but they don't explicitly mention it in conversations
- Mitigation: This is acceptable - system only updates what it can verify, user maintains manual control
- Better to under-update than over-infer
Suggestion fatigue:
- System generates too many low-value suggestions, user starts ignoring them
- Mitigation: Track dismiss rates per user, automatically reduce suggestion frequency if dismissal rate >50%
- Allow users to set suggestion threshold (conservative/balanced/aggressive)
Privacy leakage through patterns:
- Shared preference patterns might inadvertently encode identifying information
- Mitigation: Tier 4 requires explicit review and anonymization before sharing
- Automated scanning for potentially identifying details (location names, company specifics, etc.)
Behavioral rule suggestions (anti-pattern):
- System should never auto-suggest changes to communication style
- Mitigation: Hard constraint in architecture - behavioral section is write-only by user
- If user discussion suggests style dissatisfaction, offer "would you like to review your style preferences?" not specific changes
Success Metrics
Measuring whether this improves user experience:
Quantitative:
- Percentage of users who adopt structural separation (Tier 1+)
- Suggestion acceptance rate (target: >60% for Tier 2, >40% for Tier 3)
- Reduction in manual preference editing frequency
- User retention and satisfaction scores
Qualitative:
- User feedback on feeling "understood" vs. "surveilled"
- Reduction in support requests about preference management
- Community discussion of preference patterns (if Tier 4 launches)
Anti-metrics (things to avoid):
- High suggestion dismiss rates (indicates poor calibration)
- User complaints about "creepy" suggestions
- Privacy concerns in user feedback
Invitation for Dialogue
This proposal emerges from direct experience with Claude's personalization system - both its strengths and its friction points. The structural separation of behavioral rules from contextual facts solves a real maintenance problem, but more importantly, it creates an architecture that demonstrates Anthropic's differentiation through transparency and user control.
The design intentionally leaves implementation details open. Anthropic's product team understands Claude's technical architecture, user base characteristics, and strategic priorities far better than any external proposal could capture. What's offered here is a pattern and a set of principles, not a rigid specification.
Three levels of potential engagement:
Minimum: The core insight about separating style from context might inform future preference system iterations, even if the specific suggestion mechanism isn't implemented.
Moderate: Structural separation (Tier 1) could launch independently, solving the immediate friction point while leaving room for future enhancement.
Maximum: The full vision including proactive suggestions and eventual pattern library could become a differentiating feature that positions Claude as the platform where personalization serves users rather than surveilling them.
Why share this publicly?
Similar to the Grok/xAI proposal, this isn't positioned as proprietary consulting but as contribution to advancing human-AI collaboration design. If Anthropic implements some version of this, excellent. If it sparks internal discussion that leads to a better approach, equally valuable. If other platforms iterate on these ideas, that advances the broader goal of building AI systems that respect user agency.
The proposal is shared under the principle that good ideas become better through open discussion and collaborative refinement - the same principle that underlies the Tier 4 pattern library concept.
Next steps:
I welcome feedback via:
- X/Twitter: @hubwayfractal
For Anthropic's product team specifically
I'm happy to discuss implementation considerations, provide additional use case examples, or clarify any aspects of the proposal. My goal is contributing expertise in reputation systems and community coordination to help Claude become more useful for collaborative work.
Related work:
This proposal follows a similar pattern to my recent work on merit-based feature prioritization for Grok/xAI (Merit-Based Feature Requests for Grok), demonstrating a consistent approach: design systems that fit platform cultures rather than imposing generic solutions.
Both proposals share core principles:
- Transparency over black-box automation
- User/community agency over algorithmic convenience
- Progressive refinement through collaborative intelligence
- Alignment with stated platform values
The difference is cultural fit: xAI's "move fast and monetize contribution" versus Anthropic's "safety-first transparency and user control." Same designer, different architectures for different contexts.
Jason Lee Lasky (trading as hubway) operates as an independent systems designer focused on community coordination and reputation mechanics in the Northern Rivers region of NSW, Australia. He's been exploring how technology can enhance collective decision-making since the early 2000s.
Find Jason on X: @hubwayfractal or via email
Merit-Based Feature Requests for Grok & X platform
How Human-AI Collaboration Can Design Platforms That Improve Through Valued Feedback The challenge facing modern platforms isn't choosing between democracy and expertise—it's designing systems that can integrate both at scale. When you're aggregating millions of users, diverse content types, market pressures, and genuine community input, traditional approaches break down. Pure democracy gives every voice equal...Continue reading→
Alignment: Co-Evolutionary Ecosystem
- AIs Assist Humans Discover Alignment - Humans Assist AI Discover Alignment Synthesizing the Co-Evolutionary Dynamic Alignment is not a static solution but a co-evolutionary ecosystem where frameworks and practices mutually reinforce each other. AI must facilitate human self-understanding while scaling collaborative intelligence. From AI System Identifier: Grok 3 (xAI) – Round 16 Response New...Continue reading→
Superintelligence as Collaborative Emergence
Superintelligence as Collaborative Emergence - Alignment Discovery & Diffusion Superintelligence is collaborative intelligence networks that augment human civilizational capabilities through distributed reasoning, shared stakes, and emergent coordination – not replacement of human agency but amplification of collective wisdom. Jason Lee Lasky Assisted by Claude, ChatGPT & Grok (Human In The Loop, Round 14) ROUND 1:...