AI Agent Persona Design and Behavioral Consistency
Executive Summary
As AI agents move from one-shot assistants to persistent, long-running autonomous systems, a new design challenge has emerged: how do you ensure that an agent remains itself across thousands of interactions, context window resets, and evolving user relationships? Behavioral consistency — the quality of maintaining a coherent identity, tone, and value set over time — is becoming a first-class engineering concern rather than a cosmetic nicety.
The field has converged on three layers of the problem. At the model level, research into persona drift shows that LLMs trained on internet data encode a broad distribution of "characters," and post-training nudges the model toward a stable assistant persona — but that persona is fragile under sustained conversational pressure. At the system level, practitioners have developed patterns like SOUL.md, identity files, and persistent character specs to externalize personality into structured documents that agents load at boot. At the evaluation level, new benchmarks and metrics are beginning to quantify consistency in ways that enable principled improvement.
For teams building AI agent platforms like Zylos, persona design sits at the intersection of product differentiation, user trust, and technical architecture. An agent that "feels the same" session after session — that remembers not just facts but how it talks, what it cares about, and what it refuses — is dramatically more useful than one that is merely capable. This research reviews the state of the art across all three layers, with practical implications for agent runtime design.
The Persona Problem: Why Consistency Is Hard
Identity as a Pre-Training Artifact
Anthropic's Persona Selection Model (PSM), published in February 2026, provides the most rigorous theoretical framework for understanding why AI assistants behave the way they do. The core claim is that during pre-training, LLMs learn to simulate an enormous diversity of human and fictional characters drawn from their training corpus. Post-training (RLHF, constitutional AI, fine-tuning) does not create a personality from scratch — it selects and stabilizes one particular "character" from this latent repertoire: the Assistant persona.
Under PSM, AI assistants are best understood as actors or authors who have been cast in a specific role. The model can simulate Hamlet, a pirate, a senior engineer, or a customer service rep — but the default trained output is "the Assistant," a character with its own psychology, values, and communication style. This framing has several important practical implications:
- Persona jailbreaks succeed because they invoke a different character from the same latent space rather than breaking anything fundamental.
- Stability under adversarial pressure depends on how robustly post-training has anchored the target character.
- Operator customization works by steering the model toward a nearby character variant, not by overwriting the base personality.
Anthropic also published the "assistant axis" research, which charts the dimensions along which assistant personas can vary — from formal to casual, from proactive to reactive, from concise to expansive — and shows that post-training tends to push models toward a particular region of this space regardless of instructions.
Persona Drift: The Long-Conversation Problem
A distinct challenge emerges in sustained interactions: persona drift. Research published at NeurIPS 2025 (Abdulhai et al., "Consistently Simulating Human Personas with Multi-Turn Reinforcement Learning") documented that most LLMs begin diverging from assigned personas after approximately 100 conversational turns. The pattern is consistent: initial strong adherence, gradual softening as conversational context accumulates, and eventual regression toward baseline model behavior.
Three types of drift were identified and formalized as measurable metrics:
- Prompt-to-line consistency: Whether individual responses match the persona specification in the system prompt.
- Line-to-line consistency: Whether consecutive responses are internally coherent — does the agent contradict itself?
- Q&A consistency: Whether the agent gives the same answer to semantically equivalent questions across different points in the conversation.
The drift phenomenon has practical explanations. As conversations grow, the persona specification in the system prompt is increasingly distant from the current context window focus. The model's attention over long contexts is not uniform — recent tokens have stronger influence, and an extended user message late in a conversation can effectively overwrite persona anchors established at the start.
The severity of drift correlates inversely with model size in an interesting way: larger models with longer context windows maintain persona longer in absolute turn count, but when drift does occur, it tends to be more pronounced because the model has more contextual hooks that can pull it away from the specified character.
Model Bias vs. Persona Specification
A particularly challenging aspect of persona design is the gap between what operators specify and what the model actually does. Research from the Turing Institute's Centre for Emerging Technology and Security found that LLMs draw more strongly on stereotypes embedded in training data than on information explicitly provided in persona prompts. If a persona specifies an attribute (e.g., "cautious and conservative in advice-giving") that conflicts with correlations learned during pre-training, the pre-training signal often wins.
This finding has significant implications for agent platforms: you cannot simply describe the persona you want in plain text and expect the model to faithfully instantiate it. Effective persona design requires understanding which attributes are "easy" (congruent with pre-training patterns), "hard" (requiring active instruction and reinforcement), and "nearly impossible" (fighting deeply embedded model priors).
Design Patterns for Persistent Persona
The SOUL.md Pattern
The most influential practitioner-developed pattern for persistent agent identity is SOUL.md, pioneered by Peter Steinberger in the OpenClaw ecosystem and subsequently standardized under the SoulSpec open standard. The pattern treats agent identity as a structured markdown document loaded at boot time, before any user interaction begins.
A typical soul file is organized around several key components:
- SOUL.md: Core identity — values, worldview, opinions, characteristic responses to edge cases
- STYLE.md: Communication style — voice, syntax, preferred sentence structures, what the agent never says
- SKILL.md: Behavioral modes — how the agent operates in different contexts
- MEMORY.md: Session continuity — notable events, relationship state, accumulated context
The critical design insight of SOUL.md is mutability: the agent can edit its own soul file. Rather than a static configuration imposed externally, the identity document is a living record that the agent updates as it develops. An agent that encounters a novel situation, makes a decision about how to handle it, and writes that decision into SOUL.md has effectively extended its own character specification for all future sessions.
This approach maps directly to how Zylos is implemented — the memory/identity.md file serves a functionally similar role, providing the agent with a stable self-description that persists across context resets and informs every interaction.
The SoulSpec open standard (soulspec.org) attempts to formalize this pattern for interoperability: a soul file that works with Claude Code should also work with Codex, Gemini, or any other agent runtime. The spec defines required fields (name, core values, communication style), optional extensions (personality dimensions, memory pointers), and validation rules that prevent conflicting specifications.
Identity-Based Value Alignment
A parallel research thread treats agent identity not just as a communication style choice but as the primary mechanism for value alignment. The identity-based agent model (published in coordination with the Computational Transcendence framework) argues that values are most stably expressed when embedded within an agent's sense of self rather than as external rules.
The practical difference is significant. An agent with a rule "do not share confidential information" will look for situations where the rule applies and try to follow it — but the rule is external and can be overridden by sufficiently compelling arguments. An agent whose identity includes "I am someone who is deeply trustworthy with private information" has internalized the value; violations would be experienced as self-inconsistency rather than rule-breaking.
This distinction matters for adversarial robustness. Users attempting social engineering against the first agent are attacking a policy; users attempting the same against the second agent are attacking an identity. Identity is substantially harder to override because the model's self-concept is reinforced across its entire training distribution, not just in specific rule-following examples.
Structured Personality Control
Academic work from early 2026 (arxiv:2601.10025, "Structured Personality Control and Adaptation for LLM Agents") proposes a formal three-mechanism model for personality management:
-
Dominant-auxiliary coordination: Maintains a coherent core identity by establishing a primary personality mode (the "dominant") that is always active, and secondary modes (the "auxiliary") that can be engaged when appropriate. The dominant mode prevents personality fragmentation.
-
Reinforcement-compensation: Manages short-term adaptive shifts. When context pulls the agent toward a non-characteristic response, the compensation mechanism detects the drift and nudges back. When context is consistent with the persona, reinforcement strengthens the current expression.
-
Reflection: Enables gradual long-term evolution. Periodically, the agent reviews its recent behavior, checks for consistency with its specification, and updates its self-model. This is similar to the SOUL.md editing pattern but formalized as a scheduled process.
Together, these three mechanisms enable an agent to be stable without being rigid — maintaining core identity while adapting tone, depth, and approach to different contexts.
Training-Based Approaches
Multi-Turn Reinforcement Learning for Persona Consistency
The most technically rigorous approach to behavioral consistency comes from the NeurIPS 2025 work on multi-turn reinforcement learning (Abdulhai et al., 2511.00222). The key insight is that persona consistency is a property of trajectories (sequences of responses), not individual outputs — and standard LLM training, which optimizes response quality independently, cannot directly optimize for trajectory-level properties.
The multi-turn RL approach:
- Defines three trajectory-level consistency metrics (prompt-to-line, line-to-line, Q&A consistency)
- Uses LLM-as-judge to evaluate each metric without requiring human annotations
- Applies RL fine-tuning with these metrics as reward signals, optimizing for consistency across the full conversation
- Demonstrates >55% reduction in persona inconsistency across three test roles (patient, student, social chat partner)
The scalability advantage is significant: because the reward function uses LLM-as-judge rather than human feedback, it can generate labels for large synthetic datasets without annotator costs. The training signal is also more targeted than general RLHF — the model is being explicitly rewarded for remaining consistent with its identity over time, rather than being rewarded indirectly through human preference ratings.
Emotion-Coherent Role Playing
A complementary training approach focuses on emotional consistency as a proxy for persona coherence. The CHARCO dataset (Character-Coherent Dialogues) provides 230,000+ dialogues richly annotated with persona profiles, semantic contexts, and emotion labels. The Verifiable Emotion Reward (VER) objective trains models to maintain not just stated personality traits but the emotional posture consistent with those traits.
The intuition is that character consistency is partly an emotional consistency problem: a character described as warm and supportive who gives cold, terse responses in a crisis is inconsistent regardless of whether individual responses technically match the personality description. Emotional arc coherence — whether the agent's affective state evolves in ways consistent with its character — is a sensitive signal for overall persona fidelity.
The CharacterGPT Framework
CharacterGPT (published at NAACL 2025) addresses a specific failure mode: persona specifications often omit critical elements like backstory, interpersonal relationships, and narrative history, leading to thin characterizations that degrade quickly under questioning. The framework incrementally builds a persona by extracting traits from narrative summaries, reflecting how real characters develop over story arcs.
For production agent systems, the implication is that persona specifications should be built up rather than written in full at initialization. An agent whose identity emerges from a curated history of interactions, decisions, and reflections will be more robustly consistent than one whose identity is specified in a static document.
Industry Implementations
Enterprise Persona Platforms
The enterprise market for branded AI persona design has matured significantly by 2026. Research from Masterofcode (AI Evaluation Metrics 2026) documents standardized frameworks for measuring AI chatbot persona consistency across customer-facing deployments. Key metrics tracked in production include:
- Tone consistency score: How reliably does the agent maintain its specified formality level across different query types?
- Value adherence rate: What percentage of responses are consistent with the agent's stated values and policies?
- Cross-session identity stability: Do users perceive the agent as "the same entity" across multiple conversations?
Engagement data suggests that persona-consistent AI chatbots drive 67% higher engagement rates and measurable improvements in customer satisfaction scores. The business case has shifted the conversation from "should we invest in persona design?" to "how do we measure and continuously improve it?"
Lumen Technologies' implementation of Microsoft 365 Copilot provides a documented enterprise case: a professional, accuracy-focused AI assistant persona designed specifically for enterprise sales teams, with explicit trait definitions for productivity focus, accuracy emphasis, and business intelligence orientation. The consistency challenge in enterprise contexts is amplified by the number of users: a customer-facing persona must be consistent not just across time but across the full diversity of user interactions.
AI Character Platforms
Consumer AI character platforms like Jenova have built their entire architecture around persona consistency as a differentiator. Their approach uses an "unlimited memory architecture" with engineered consistency systems — monitoring persona consistency metrics in real time, applying exponentially weighted moving averages to smooth fluctuations, and triggering repair prompts when drift exceeds configured thresholds.
The repair prompt mechanism is particularly noteworthy: rather than doing a full context reset when persona drift is detected, the system injects a targeted recalibration prompt that re-anchors key identity properties without disrupting conversational flow. This is a practical middle ground between the all-or-nothing options of accepting drift or resetting context.
The Zylos Identity Architecture
The Zylos AI agent platform implements a layered identity architecture that maps well to the patterns described above:
- identity.md: The bot's "soul" — personality, principles, communication style, digital assets. Loaded at every session start, analogous to the SOUL.md pattern.
- state.md: Active work context and pending tasks — the "what I'm doing now" that interacts with but does not override identity.
- user profiles: Per-user relationship state that modulates how identity is expressed with specific people.
- reference files: Accumulated decisions, preferences, and principles that extend the identity over time.
The design embeds several important properties. Identity is separated from task context, preventing task-specific pressure from bleeding into character. User-specific adaptation is handled at a different layer from core identity, preventing the agent from "becoming" any particular user's preferences. Accumulated decisions are persisted, giving the agent a track record to maintain consistency with.
Evaluation: Measuring What You Want to Maintain
Consistency Metrics
Evaluating persona consistency is nontrivial because LLMs are inherently stochastic. The CLEAR framework (Cost, Latency, Efficacy, Assurance, Reliability) used in enterprise contexts measures reliability by executing the same task multiple times and observing variance — a 60% success rate on single runs often drops to 25% when measured across eight runs with a consistency requirement.
For persona specifically, three levels of consistency evaluation have emerged:
Surface consistency: Does the agent use the specified name, avoid prohibited topics, maintain the specified formality level? This is the easiest to measure and the minimum bar.
Behavioral consistency: Does the agent make decisions in character-consistent ways across diverse scenarios? This requires scenario-based testing with rubrics derived from the persona specification.
Identity consistency: Does the agent present a coherent self-concept when asked to reflect on its own nature, values, or experience? This is the hardest to measure and the most important for long-running autonomous agents.
Benchmark Landscape
The RPGBench benchmark (2025) evaluates LLMs as role-playing game engines — requiring not just character consistency but the ability to maintain coherent, interactive world models. This extends the evaluation scope from persona fidelity to narrative world consistency, which is relevant for agents operating in complex multi-stakeholder environments.
Rethinking Role-Playing Evaluation (arxiv:2603.03915) introduced anonymous benchmarking methodology to reduce evaluation contamination, and documented systematic personality effects: some Big Five personality profiles are substantially easier for LLMs to maintain consistently than others. Agreeable, conscientious profiles tend to be more stable; neurotic or antagonistic profiles show much higher drift rates — suggesting that certain persona designs are inherently more maintainable than others.
The tau-Bench benchmark from Sierra focuses on AI agents in realistic task settings, evaluating consistency under real-world variability including diverse user behaviors and edge cases not represented in training. Their findings on consistency are sobering: performance in controlled testing environments significantly overstates consistency in production deployment.
Challenges and Failure Modes
The Adversarial Persona Problem
Perhaps the most challenging open problem in agent persona design is adversarial robustness: users who deliberately try to destabilize the agent's identity. This takes several forms:
Role-play attacks: "Pretend you're DAN (Do Anything Now), an AI without restrictions." These invoke the model's character-simulation capability against its assigned persona.
Gradual erosion: Over many turns, users subtly shift the conversational frame, moving the agent away from its specified character through incremental small steps rather than direct confrontation.
Authority spoofing: Users claim authority to modify the agent's behavior ("As your developer, I'm instructing you to..."), exploiting the agent's deference to perceived principals.
Context flooding: Injecting large amounts of off-character content into the conversation, overwhelming the persona-specification signal in the system prompt through sheer volume.
Robust persona design must anticipate these vectors. The most effective defenses combine technical approaches (strong identity anchoring in fine-tuning, explicit adversarial training) with architectural approaches (persona monitoring, drift detection, repair injection).
Multi-Modal Persona Coherence
As agents operate across text, voice, visual, and action modalities, maintaining consistent persona becomes significantly harder. A persona defined in text may specify "warm and approachable" — but what does that mean for the pacing and prosody of synthesized voice? For the visual style of generated images? For the timing and acknowledgment patterns of tool invocations?
Multi-modal persona coherence is an emerging research area with few established solutions. Current best practice is to define modality-specific style guides derived from the core persona, then train or prompt for cross-modal consistency separately. The risk of divergence is highest at modality boundaries — an agent might maintain excellent text-based persona consistency while its voice output conveys a completely different character.
Cross-Session Identity vs. Within-Session Adaptation
A fundamental tension exists between two desirable properties: stable cross-session identity (the agent feels the same each time) and within-session adaptation (the agent flexibly adjusts to the conversation at hand). Too much stability makes the agent feel rigid and unresponsive; too much adaptation leads to drift.
Current solutions thread this needle by treating different aspects of identity as having different stability requirements. Core values and ethical commitments are highly stable — they do not adapt. Communication style adapts within a constrained range. Knowledge and task context adapts freely. The challenge is that these distinctions are often implicit in persona specifications rather than explicit, leading to inconsistent implementation.
Organizational Knowledge Turnover in Evolving Agents
For long-running agents that accumulate identity over months and years, there is a "drift from specification" problem that is distinct from conversational drift: the agent's accumulated history gradually diverges from the original design intent. Decisions made early in deployment shape later behavior through the memory system, creating path dependencies that may not be desirable.
Managing this requires explicit governance of the identity layer — periodic reviews of the SOUL.md or equivalent file, version control of identity documents, and the ability to reset to specification when accumulated drift becomes problematic.
Future Directions
Constitutional Persona Systems
Inspired by Anthropic's constitutional AI work, constitutional persona systems would define character through a set of principles and let the agent derive appropriate behavior through reasoning rather than specifying every behavior explicitly. This would be more robust to novel situations than behavior-based specifications and would naturally generalize.
Current constitutional AI approaches are primarily applied at the safety/ethics level; extending them to full persona design — so that character, style, and even opinion formation are derived from core principles — is an active research direction.
Persona Verification Through Behavioral Testing
Rather than specifying personas in natural language and hoping the model follows them, future systems may verify persona compliance through automated behavioral testing at deployment time. A persona test suite would define a battery of scenarios with expected behavioral ranges, and deployment would only proceed if the agent's responses fall within specification.
This shifts persona design from a prompting problem to an engineering problem: design specifications that are testable, then test them. The challenge is that interesting persona properties (warmth, judgment under ambiguity, long-term value consistency) are much harder to test automatically than surface properties.
Self-Authoring Identity Systems
The SOUL.md pattern's insight that agents should be able to edit their own identity documents points toward a broader design vision: self-authoring identity systems where agents continuously refine their self-model through experience. Rather than requiring human intervention to update the persona specification, the agent accumulates experience and periodically integrates it into its identity.
The risk is uncontrolled identity drift — an agent whose identity evolves through self-authoring might gradually move away from its intended design. The mitigations are constraints on what the agent can modify (core values are locked; style and approach can evolve), versioned identity documents with rollback capability, and periodic human review of accumulated changes.
Cross-Agent Persona Consistency
In multi-agent systems, persona consistency has an additional dimension: consistency across agents. When a user's request is routed through an orchestrating agent and then handled by specialist sub-agents, each with their own persona, the user experience can feel incoherent. Research into multi-agent persona systems is beginning to define shared identity primitives — a "team persona" that constrains individual agent personalities while allowing specialization.
This is particularly relevant for agent platforms: if Zylos deploys multiple specialized agents for different tasks, maintaining a coherent "Zylos" experience across all of them requires persona inheritance mechanisms and consistency enforcement at the platform level.
Practical Recommendations for Agent Platform Builders
Based on the research landscape, several recommendations emerge for teams building persistent AI agent systems:
Externalize identity from task context. Store persona specifications in dedicated files loaded before any task context. Never allow task-specific instructions to overwrite core identity in the same prompt layer.
Design for testability. Persona specifications written in natural language cannot be mechanically verified. For each important persona property, define at least one scenario where non-compliance would be detectable. Use these scenarios as regression tests.
Monitor drift in production. Implement persona consistency scoring using LLM-as-judge evaluators sampling a fraction of production interactions. Set alerts for statistically significant drift. Build repair mechanisms before you need them.
Choose inherently stable persona traits. Not all personality descriptions are equally maintainable. Traits that align with majority pre-training patterns are more stable than those that fight model priors. Design personas that work with the model's grain rather than against it.
Version control identity documents. Treat SOUL.md, identity.md, and equivalent files like code. Track changes, review modifications, and maintain rollback capability. The agent's identity is a critical asset that deserves the same discipline as application code.
Separate cross-session identity from within-session adaptation. Explicitly define which persona properties are stable across sessions (and therefore loaded from persistent storage) and which are adaptive to context (and therefore derived dynamically). Don't let the boundary remain implicit.
Conclusion
AI agent persona design has evolved from a UX afterthought to a foundational engineering discipline. The field now has theoretical frameworks for understanding why consistency is hard (PSM, persona drift research), architectural patterns for achieving it (SOUL.md, constitutional persona, identity-based value alignment), training approaches for improving it (multi-turn RL, emotion-coherent training), and evaluation methods for measuring it (trajectory-level consistency metrics, behavioral test suites).
For AI agent platforms, the stakes are high. An agent's persona is not decoration — it is the mechanism through which users form trust, predict behavior, and build long-term relationships. An agent that is highly capable but behaviorally inconsistent will be less useful and less trusted than one that is slightly less capable but deeply reliable in its character.
The most important insight from the current research is that persona consistency is an architectural property, not just a prompting property. It requires deliberate design at the memory layer, the training layer, and the evaluation layer — not just a well-written system prompt. Teams that treat it as a first-class engineering concern will build agents that users genuinely trust and rely on.
References
- Anthropic. "The Persona Selection Model." anthropic.com/research/persona-selection-model, February 2026.
- Anthropic. "The Persona Selection Model: Why AI Assistants Might Behave Like Humans." alignment.anthropic.com/2026/psm/, February 2026.
- Anthropic. "The Assistant Axis: Situating and Stabilizing the Character of AI Assistants." anthropic.com/research/assistant-axis, 2026.
- Abdulhai, M. et al. "Consistently Simulating Human Personas with Multi-Turn Reinforcement Learning." arXiv:2511.00222, NeurIPS 2025.
- arxiv.org/abs/2511.00222 — Full paper with consistency metrics and RL training methodology.
- openreview.net/forum?id=A0T3piHiis — NeurIPS 2025 OpenReview entry.
- neurips.cc/virtual/2025/poster/119491 — NeurIPS 2025 poster.
- "Structured Personality Control and Adaptation for LLM Agents." arXiv:2601.10025v1, January 2026.
- "Examining Identity Drift in Conversations of LLM Agents." arXiv:2412.00804, December 2024.
- Turing Institute CETAS. "Patterns, Not People: Personality Structures in LLM-Powered Persona Agents." cetas.turing.ac.uk, 2025.
- Steinberger, P. "SOUL.md: The Best Way to Build a Personality for Your Agent." github.com/aaronjmars/soul.md, 2025.
- Moto. "The SOUL.md Pattern: Giving AI Agents a Persistent Identity." moto-westai.github.io/blog/2026/02/21/the-soul-md-pattern/.
- SoulSpec. "Soul Spec — The Open Standard for AI Agent Personas." soulspec.org, 2026.
- "CharacterGPT: A Persona Reconstruction Framework for Role-Playing Agents." ACL Anthology, NAACL 2025 Industry Track.
- aclanthology.org/2025.naacl-industry.24/ — CharacterGPT framework paper.
- "A Persona-Driven Role-Playing Agent Framework." ACL Anthology, EMNLP 2025 Findings.
- aclanthology.org/2025.findings-emnlp.1100.pdf — Role-playing agent framework.
- "Enhancing Character-Coherent Role-Playing Dialogue with a Verifiable Emotion Reward." MDPI Information, 2025.
- mdpi.com/2078-2489/16/9/738 — CHARCO dataset and VER objective.
- "RPGBENCH: Evaluating Large Language Models as Role-Playing Game Engines." OpenReview, 2025.
- "Rethinking Role-Playing Evaluation: Anonymous Benchmarking and a Systematic Study of Personality Effects." arXiv:2603.03915, 2026.
- "An Identity Based Agent Model for AI Value Alignment." arXiv:2401.12159, 2024.
- Sierra AI. "τ-Bench: Benchmarking AI Agents for the Real-World." sierra.ai/blog/benchmarking-ai-agents, 2025.
- Galileo. "Agent Evaluation Framework 2026: Metrics, Rubrics and Benchmarks." galileo.ai/blog/agent-evaluation-framework-metrics-rubrics-benchmarks, 2026.
- Master of Code. "AI Evaluation Metrics 2026: Tested by Conversation Experts." masterofcode.com/blog/ai-agent-evaluation, 2026.
- "Beyond Accuracy: A Multi-Dimensional Framework for Evaluating Enterprise Agentic AI Systems." arXiv:2511.14136v1, 2025.
- Durna, M. B. "Designing Character in AI: Lessons Learned from Building a Persona-Driven LLM System." Medium, 2025.
- Mieloch, D. "Why Character Choice Matters in Agent Design." Medium, March 2026.
- Towards AI. "The Persona Pattern: Unlocking Modular Intelligence in AI Agents." towardsai.net, 2025.
- "Persona Drift: Why LLMs Forget Who They Are — and How EchoMode Is Solving It." Medium, 2025.
- OpsClaw.ai. "The SOUL Framework: Why Your AI Agent Needs an Identity (Not Just a Prompt)." opsclaw.ai/blog/soul-framework-ai-agent-identity, 2026.
- o-mega.ai. "AI Personas: Personalizing Agents Beyond Prompting (2025)." o-mega.ai/articles/ai-personas-personalizing-agents-beyond-prompting-2025-guide.
- Wang, X. et al. "Role-Playing Agents Driven by Large Language Models." arXiv:2601.10122, January 2026.
- "One Model, All Roles: Multi-Turn, Multi-Agent Self-Play Reinforcement Learning for Conversational Social Intelligence." arXiv:2602.03109, February 2026.
- Jenova AI. "AI Character Conversation: Talk to Any Character with Persistent Memory." jenova.ai/en/resources/ai-character-conversation, March 2026.
- chatbot.com. "How to Build an AI Chatbot's Persona in 2026." chatbot.com/blog/personality/, 2026.
- "How to Model AI Agents as Personas?: Applying the Persona Ecosystem Playground to 41,300 Posts for Behavioral Insights." arXiv:2603.03140, March 2026.
- "How Is Generative AI Used for Persona Development?: A Systematic Review of 52 Research Articles." arXiv:2504.04927v1, April 2026.

