AI Coaching Conversational UX Agentic Systems Behavioral Design

The coach
that shows up
between
sessions

Paloma is an AI coaching companion designed to close the gap between live sessions — helping clients reflect, regulate, and sustain momentum when their coach isn't in the room.

1.4×
Weekly engagement
5
Memory layers
0
Trauma stored
RAG
Orchestration model
PALOMA Session 6 · Jan 14 What felt different this week compared to the last time you felt this way? Paloma · 9:14 AM I actually paused before reacting. That's new for me. You That pause — that's the shift your coach and you have been working toward. On a scale of 1–10, how intentional did it feel? 1 3 5 7 9 10 Win logged · Session prep Say something…
Project Context
Role
Lead AI Product Designer
System architecture, UX, cognitive design, safety model
Platform
Mobile-first · Coach Portal
Client companion app + coach-facing dashboard
Model
RAG · QLoRA-tuned LLM
Orchestration layer with solution-focused coaching alignment
Domain
Executive Coaching · Behavioral Design
Human-in-the-loop AI between monthly coaching sessions
The Problem
What the industry assumed

Coaching is what happens in the session.

The traditional model treats the 60-minute session as the product. Between meetings, clients are on their own — expected to journal, remember insights, hold commitments, and stay regulated without any infrastructure for doing so.

The result: value leakage. Coaches spend the first 20 minutes of each session re-establishing ground they already covered. Clients arrive either emotionally dysregulated or having forgotten what they committed to. The work doesn't compound.

What was actually happening

Change takes root between sessions.

The moments that matter most — the pause before reacting, the decision made under pressure, the small win nobody noticed — happen in daily life, not in the coaching room.

The continuity gap isn't a coaching problem. It's a systems design problem. What was missing wasn't more sessions — it was an intelligent system that could hold the thread, prompt reflection, and surface the right memory at the right moment.

"Paloma doesn't try to be a coach. It's the infrastructure that makes coaching work."

Design principle · Human-first architecture
Behavior Layer — How Paloma Thinks

A coaching agent cannot behave like a general chatbot. Every design decision — what Paloma says, stores, remembers, and refuses — was shaped by a set of explicit behavioral guardrails.

01 · Dialogic Orientation
Present, not prescriptive

Every utterance is oriented toward resourcefulness and co-construction. Paloma reflects what the client already knows. It never advises, diagnoses, or directs — because the moment it does, it's no longer a coach's partner. It's a liability.

02 · Forward Focus
Toward the desired state

Rooted in Solution-Focused Coaching, Paloma uses scaling questions, Miracle Question sequences, and progress anchoring. "How would you know you're at a 7 instead of a 5?" is infinitely more generative than "What went wrong?"

03 · Emotional Safety
She never stores the hard stuff

If emotional activation is detected, Paloma shifts into grounding, pacing, and somatic micro-prompts — and then redirects to the live coach. Trauma, crisis, and clinical territory are hard boundaries. Trust is the product.

04 · Regulated Presence
Grounding as architecture

The Regulation Profile stores pre-approved, context-free techniques for in-the-moment activation. When a client is overwhelmed, Paloma doesn't ask them to reflect — it helps them regulate first. Then it asks.

05 · Permission + Agency
Clients control what gets shared

Paloma always asks before storing or sharing insights with a coach. Clients determine what's visible. This isn't just an ethical requirement — it's the design decision that makes people willing to be honest with the system.

06 · Human-First
The coach is the anchor

Paloma is not a replacement. The live coach is the relationship — Paloma is the extension. Every session, Paloma generates a one-page prep summary for the client and coach: what shifted, what's working, what they want next.

Memory Architecture — 5 Layers

Over-remembering destroys trust. Under-remembering makes the system useless. Paloma's memory system was designed to feel personally attuned without feeling surveilled — a precise calibration between depth and restraint.

System Architecture
Entry
Client Conversation
Behavior Layer · DOQ Guardrails
Layer 1
Working Memory
Layer 2
Episodic Summary
Layer 5
Regulation Profile
Layer 3
Semantic Identity
Layer 4
Action & Progress
Safety Layer · Retrieval · RAG
Output
Client UX
Output
Coach Dashboard
L1
Working Memory
Short-term conversational context only. Not persisted. Resets each session — intentionally.
Not stored
L2
Episodic Summary
User-approved session takeaways, meaning extraction, and insight shifts. Nothing stored without explicit permission.
User-approved
L3
Semantic Identity
Strengths, preferences, values, and desired identity statements. Slow-changing, human-reviewed. This is who the client is becoming — not who they were.
Human-reviewed
L4
Action & Progress
Micro-commitments, small wins, and longitudinal signals of behavior change. The layer that makes coaching feel like it compounds over time.
Longitudinal
L5
Regulation Profile
Pre-approved, context-free grounding techniques for in-the-moment activation. Reduces dependency on live coach availability during emotional spikes.
Always available
Outcomes
1.4×
Weekly engagement

Clients using Paloma between sessions engaged with their coaching work significantly more often than those using session notes alone.

0
Trauma stored in memory

The safety model held across all edge case testing — emotional content, crisis mentions, and clinical redirects all handled without storing harmful data.

5
Memory layers, zero data creep

The episodic retrieval system maintained deep personalization without triggering the "she knows too much" response — the design benchmark for trust.

−20m
Session re-establishment time

Coaches reported spending far less time re-grounding clients at session start. Paloma's session prep summary gave both parties a shared starting point.

100%
Coach-approved sharing model

Clients controlled every insight shared with their coach. Voluntary data sharing from clients to coaches became a trust signal, not a compliance mechanism.

QLoRA
Fine-tuning for coaching tone

Light model fine-tuning gave precise control over the solution-focused coaching philosophy — preventing the generic chatbot drift that breaks the therapeutic frame.

What This Work Demonstrates
How do you design an AI that earns trust instead of assuming it?

You give users explicit control over everything the system remembers and shares. Trust is a design output, not a brand promise. Every permission prompt, every consent moment, every "Paloma won't store this" message is load-bearing architecture.

What does "human-in-the-loop" actually mean in practice?

It means designing the AI to amplify the human relationship, not substitute it. Paloma doesn't schedule sessions, evaluate progress, or hold the therapeutic relationship — the coach does. Paloma holds the thread between the moments that matter.

Why does agentic AI need behavioral guardrails at the prompt level?

Because general-purpose models drift. Left unguided, an LLM will advise when it should reflect, praise when it should question, and empathize when it should redirect. Coaching alignment has to be baked into the cognitive architecture — not bolted on after.