Pathic MR is a mixed reality empathy-training system that helps people practice emotionally difficult conversations through embodied perspective-taking, role reversal, and AI-guided dialogue.
Pathic transforms abstract empathy concepts into embodied experiences. Users do not just learn about perspective-taking; they step into another person’s position, navigate conflict from multiple viewpoints, and reflect on how emotion, tone, and assumptions shape the outcome.
This is not just a prototype of an interface. It is a prototype of behavior change.
This project represents the convergence of spatial computing, conversational AI, learning experience design, and human-centered research to address a persistent problem: people know empathy matters, but most current training does not create the emotional realism or practice conditions needed for behavior change.
"Almost all stakeholders identified the critical importance of stepping into the other's shoes."
- Research Finding
The corporate soft-skills training market exceeds $1.8B annually in the US, with empathy and conflict resolution consistently ranking among the top L&D priorities. Remote and hybrid work has made interpersonal tension harder to detect and easier to ignore - accelerating demand for scalable, asynchronous conflict training that doesn't require a live facilitator.
Meanwhile, enterprise MR hardware costs have dropped 60%+ since 2021. Meta Quest 3 business deployments are now viable for L&D budgets that previously couldn't touch VR. The window for an embodied, AI-driven empathy training product has materially opened in the last 18 months.
This was a collaborative project, and that collaboration is explicit here. I worked with Sejal Shail on the research foundation and learning direction of Pathic. Sejal’s contributions around workplace conflict, emotional awareness, reflective practice, and progress-oriented learning are integrated into this page as part of the project story rather than hidden in the background.
Sejal’s research framing strengthened the project in four important ways: clearer articulation of workplace conflict as the initial high-value use case, greater emphasis on emotional awareness and psychological safety, stronger reflection and progress-tracking concepts, and clearer participant segmentation across managers, HR, employees, and organizational support roles.
I translated those insights into product structure, XR interaction flows, scenario design, and the spatial conversation model. This page explicitly shows that collaboration rather than flattening it into a single-author story.
I wanted this project to read clearly as learning experience design, not only XR prototyping. The product structure is grounded in verifiable learning principles: psychologically safe practice, active participation, reflection, feedback, transfer, and repetition.
Pathic creates a low-risk rehearsal space. This supports deliberate practice: learners can try difficult conversations, make mistakes, and recover without damaging real relationships.
Role-switching converts empathy from abstract instruction into lived experience. This supports stronger emotional encoding than static lecture or scripted role-play alone.
Reflection is built into the system because learning does not come from immersion alone. Debrief prompts help learners notice emotion, intent, and communication patterns so the experience can transfer beyond the headset.
Voice-first interaction, scenario specificity, and guided prompts reduce interpretive overhead so learners can focus on emotional signals and decision-making instead of interface mechanics.
Pathic MR is a mixed reality empathy training platform built on Meta Quest 3 that uses spatially anchored AI characters, emotionally responsive dialogue, and role-shifting mechanics to create genuine perspective-taking experiences.
AI characters powered by Unity and XR Interaction Toolkit, positioned in meaningful spatial relationships with the user.
Convai-powered responses with custom emotion-scaffolding prompts for empathetically intelligent dialogue.
Experience conversations as yourself or as the other person, creating emotional understanding that changes how users respond in real conversations, not just how they think about them.
Empathy scenarios grounded in real conflict patterns across workplace, academic, and relationship contexts.
Real-time narrative transitions based on where users look, enabling natural interaction flow without constant controller reliance.
Hands-free, embodied interaction enabling natural conversation without breaking immersion.
Workplace conflict is often avoided rather than addressed. Misunderstandings, emotional tension, and unspoken frustrations quietly build up across teams. These unresolved conflicts reduce collaboration, lower morale, and damage trust.
People often struggle to see another person’s perspective when emotions escalate — exactly the moment when perspective-taking matters most.
The gap between knowing you should empathize and actually experiencing another person’s point of view creates a persistent barrier to conflict resolution, learning, and growth.
"A lot of conflicts just come from people misunderstanding each other’s tone or intent. Emotions get lost in translation."
- Interviewee 1
"Training works only when it feels real, not theoretical."
- Stakeholder Research
Create an immersive way to train emotional perspective-taking in workplace conflicts so people can build empathy, confidence, and better conflict habits through interactive mixed reality experiences.
The research foundation combined literature review, stakeholder interviews, scenario analysis, and iterative prototype thinking to understand both the emotional problem space and the design requirements for MR-based empathy training.
Pathic must support emotional awareness, not only problem-solving.
The product should encourage independent conflict-resolution skills rather than overreliance on external help.
Reflection and follow-up matter if learning is supposed to persist beyond the headset.
Stakeholder research revealed multiple user archetypes, each facing different conflict dynamics and learning needs.
Struggles with academic pressure and power dynamics in advising situations. Needs empathetic communication that goes beyond policy and time pressure.
Goal: Build empathy and communication skills to guide students more holistically.
Navigates workplace tension but lacks safe ways to process emotion and rehearse responses before real conversations happen.
Goal: Gain confidence and improve communication through guided immersive practice.
Feels overwhelmed by complex theory and wants a clearer, more hands-on way to build conflict-resolution skill over time.
Goal: Track growth through guided MR sessions and simple progress feedback.
The hardest design problem in Pathic wasn’t spatial — it was conversational. Building an avatar that could sustain emotionally coherent dialogue across an unpredictable role-reversal scenario required a three-layer architecture: a branching narrative structure, a grounded knowledge retrieval system, and a carefully engineered prompt objective that kept the avatar in character without breaking the user’s psychological safety.
Tool: Convai Narrative Design Editor
Every scenario in Pathic was architected as a branching conversation tree inside Convai’s Narrative Design editor. Each node represents a discrete conversational moment: an Objective (what the avatar says or does to open a section), followed by Decision branches that route the conversation based on user intent.
In the academic advising scenario, the opening node greets the user as a museum host and forks into two branches: the user accepts a tour (routes to Section 2) or refuses (routes to Section 3 - a recovery path designed to re-engage without pressure). This structure ensures the avatar always has a contextually appropriate response regardless of what the user says, preventing the immersion-breaking dead ends that kill role-play simulations.
The Convai Narrative Design editor: the Welcome Section objective branches into two user decisions - accept or refuse - each routing to a different section of the conversation graph. No user path leads to a dead end.
Each section opens with a prompt that sets the avatar’s emotional register and communicative intent. Not a rigid script — a behavioral anchor the LLM improvises within.
Decision nodes capture the most likely user responses at each fork. The NLU layer interprets utterances against these labels - routing the tree even when the user’s exact words don’t match.
A refusal branch doesn’t terminate - it routes to a re-engagement section. Every dead end was redesigned as a psychologically safe re-entry point.
Architecture: Retrieval-Augmented Generation via Convai Knowledge Bank
The core risk in AI-driven role-play is hallucination: the avatar fabricating details that break scenario consistency or undermine psychological safety. To solve this, we used Convai’s Knowledge Bank as a retrieval-augmented generation layer - giving the avatar a curated factual foundation to draw from before generating any response.
Each scenario’s Knowledge Bank was populated with scenario-specific context: persona background, institutional policies, the student’s situation, and the emotional subtext of the conflict. When a user asks “Why wasn’t I told about this deadline?” - the system retrieves relevant context before generating a response, rather than confabulating from general training data. This is the same RAG architecture used in enterprise AI products, applied here to target psychological consistency rather than just factual accuracy.
Persona Context
Character background, role, communication style, and emotional state entering the conversation.
Scenario Facts
Institutional policies, deadlines, prior interactions, and factual context that ground the conflict in reality.
Emotional Subtext
What the character feels but hasn’t said — unspoken frustration, fear, or need beneath the surface conflict.
Guardrails
What the avatar must never say — topics that break psychological safety or terminate the scenario prematurely.
Behavioral intent lives in the objective prompt. Factual and emotional grounding lives in the Knowledge Bank. Keeping these separate means the LLM can improvise naturally within its behavioral anchor while retrieving accurate, scenario-specific context on demand - the same separation of concerns that makes production RAG systems reliable.
At runtime the three layers combine into a seamless experience. The narrative tree determines which section the avatar is in. The Knowledge Bank grounds what it knows. The LLM generates the actual words - improvising within the objective’s behavioral anchor.
Spatial environment and avatar establish emotional register. The user knows what kind of conversation they’re entering before they speak a word.
The user speaks naturally. NLU interprets intent against the current decision branches and routes the narrative tree accordingly.
Avatar retrieves from Knowledge Bank, generates within the objective anchor, delivers with synchronized expression and pacing. No hallucination.
At key branch points, a coaching prompt surfaces - a short question that makes the user notice what just happened emotionally before continuing.
Too prescriptive and the avatar sounds scripted, breaking immersion. Too open and it hallucinates outside the scenario. The answer: behavioral anchors that define emotional posture and intent without scripting words.
Early builds let users reach conversational dead ends where the avatar exited character. Every dead end was redesigned as a re-engagement section - a psychologically safe way back into the scenario.
Behavioral intent → objective prompt. Factual and emotional grounding → Knowledge Bank. Keeping these separate lets the LLM improvise naturally while still retrieving scenario-specific context on demand.
Competitive analysis revealed that most platforms address one dimension of the problem - guidance, wellness support, or scripted practice - but none combine embodied role-switching, adaptive AI dialogue, and structured reflection in a single facilitator-free system.
Talespin and Mursion come closest on immersion, but neither supports true perspective-switching and both require human facilitators - adding cost and scheduling friction that limits scalable deployment. Pathic's combination of embodied role reversal + adaptive AI dialogue + structured reflection + autonomous sessions is not replicated anywhere.
Pathic is not just an immersive scene. It is a learning loop designed to move the user from orientation to practice to reflection to follow-up.
The debrief and follow-up structure is what pushes Pathic toward learning experience design. It makes the system less like a one-time XR demo and more like a repeatable skill-building product.
Pathic MR delivers a complete empathy training experience: fully interactive AI-driven MR conversations that respond dynamically to user input and emotional state. As a learning experience, it combines embodied perspective-taking, emotional mirroring, adaptive simulation, and reflection to help users build conflict-resolution skill through practice rather than theory alone.
Experience the same conflict from multiple viewpoints, creating genuine understanding rather than only intellectual agreement.
Avatars respond with appropriate nuance to user behavior throughout extended training sessions, reinforcing emotional awareness in context.
Adjusts difficulty and intensity based on user responses and training objectives, supporting more relevant and transferable practice.
The research prototype proved the concept and the learning loop. Turning Pathic into a real product requires a clear north star, a prioritized next build, and an honest look at the hardest unsolved risk.
Behavior transfer rate - percentage of users who report applying a conflict insight from a Pathic session to a real conversation within 7 days.
Why not session completion or satisfaction scores? Because Pathic’s value proposition is behavior change, not entertainment. If people finish sessions but don’t change how they handle conflict, the product hasn’t worked. Transfer is the only metric that proves the learning system is real.
A structured post-session debrief with manager visibility. Not because it’s the most technically interesting feature - because it’s the unlock for institutional B2B sales.
HR and L&D buyers need to demonstrate ROI. A debrief that generates a structured insight report - shareable with a manager or coach - turns Pathic from a personal tool into an organizational training asset. That’s what justifies a per-seat enterprise license.
AI emotional consistency over extended sessions. Convai delivers responsive dialogue, but emotional tone drift - where the AI character’s affect gradually desynchronizes from the scenario’s emotional arc - is a real risk that we haven’t stress-tested at depth.
A single tonally inconsistent response at a moment of high emotional stakes can break the psychological safety that the entire system is built on. Longitudinal emotional consistency testing across 20+ minute sessions is the critical pre-launch validation.
Partner with 2–3 university HR or L&D departments for a controlled pilot. Measure transfer rates, session completion, and facilitator time saved. Generate the case study that unlocks enterprise sales conversations.
Expand beyond academic advising into workplace feedback, performance reviews, and team conflict. Each new scenario type opens a new buyer segment: corporate L&D, healthcare communication training, legal mediation prep.
Port core scenarios to mobile AR to remove the hardware barrier for individual users and smaller organizations. This expands addressable market from enterprises with headset budgets to any organization with smartphones.
Pathic MR taught me that designing for emotion is fundamentally different from designing for utility. Technical systems succeed when they work reliably. Emotional systems succeed when they feel authentic, and authenticity emerges from restraint, timing, reflection, and psychological safety.
Working with conversational AI revealed how much human communication depends on non-verbal cues that language models do not fully capture. The challenge was not simply generating plausible text. It was aligning generated dialogue, avatar behavior, pacing, spatial positioning, and reflection design into a coherent learning experience.
This project also clarified my identity as a learning experience designer working through emerging technology. The real product value is not the headset or the AI alone. It is the learning system: safe practice, embodied perspective-taking, guided reflection, and stronger transfer into real-world behavior.
Empathy can be trained, but training requires embodiment. Telling people to understand others does not work. Creating safe spaces where they can inhabit another person’s perspective, reflect, and try again creates learning that sticks.