Srishti
Kush

Designing at the intersection of AI, automation, and human experience. Currently shaping the Internet of Agents at Outshift by Cisco.

Srishti Kush

Lead Product Designer crafting the future of human-agent collaboration at Outshift by Cisco.

I work at the intersection of AI, automation, and infrastructure, shaping new interaction models, defining agentic design principles, and guiding research on how we will collaborate with intelligent systems.

Previously five years at GEP designing enterprise tools, including leading UX for Quantum Studio across 20+ teams, and initiating the Figma design system.

MS in UX Design and Research from NYU. Designing for humans and, increasingly, agents ever since.

Currently

Lead Designer
Outshift by Cisco

Previously

Lead Designer, GEP
Designer, Northwell Health
Designer, SHEROES

Education

MS UX Design & Research
NYU, 2017

Overview

Over several months, we designed and built ActionEngine an agentic system capable of operating across the browser and terminal. Our goal was to create a collaborative system that learns from users, adapts to their workflows, and remains transparent at every step.

The Problem

Modern engineering workflows are fragmented across tools terminals, SaaS platforms, internal systems. These workflows are highly repetitive but nearly impossible to automate with traditional tools because they require contextual judgment and vary from session to session. Engineers spend significant time on toil that shouldn't demand their full attention, yet fully autonomous automation introduces unacceptable risk.

Our Approach

We designed ActionEngine as an extension of developer intent not a replacement for it.

Observe

Watches user behavior in real-time across browser and terminal without interrupting the flow.

Learn

Identifies repeatable patterns and builds reusable workflows from what the engineer actually does.

Automate

With explicit human review and approval, executes those workflows in future matching scenarios.


System Design

Planning-First Architecture

The agent creates a plan before taking any action. This plan is visible and editable by the user enabling trust and meaningful collaboration before a single step is executed.

Multi-Agent System

Three agents work in concert:

  • Planning Agent manages tasks and sequences
  • Thinking Agent explains reasoning at each step
  • Executor Agent performs actions in the environment

Context Engineering

We balance structured DOM data and Markdown to give the agent usable, accurate context about the current state of the user's environment avoiding both information overload and critical blind spots.

Secure Execution

ActionEngine runs in a Docker sandbox with an isolated browser and terminal, ensuring agent actions are contained and reversible.


Key Interaction Patterns

Learning Through Observation

The agent learns in real-time, watching what the engineer does within the UI and building a model of intent before surfacing any suggestions.

Asking for Help

Rather than guessing when it lacks context, the agent pauses and asks for clarification. This keeps the human genuinely in the loop and prevents compounding errors from silent assumptions.

Workflow Review

Before any workflow is executed, users review and edit the proposed steps. Plans are the primary interface not opaque automation that happens behind the scenes.


HAX Principles Applied

ActionEngine was designed in alignment with the HAX framework, grounding every interaction in the same principles that govern human-agent collaboration across Outshift.

Clarity

Plans and reasoning are always visible before the agent acts.

Control

Users define what the agent can do and approve each workflow before execution.

Recovery

Sandboxed execution means mistakes are contained. Workflows can be stopped or revised at any point.

Collaboration

The agent contributes actively learning, suggesting, asking while the human leads.

Traceability

Every agent action is logged and attributable, so engineers can audit what happened and why.

Evaluation

We evaluated ActionEngine using Mind2Web and G-Eval benchmarks to measure correctness and usability ensuring the system performed reliably across realistic, varied engineering scenarios.

Key Insights

  • Trust must be designed it doesn't emerge from capability alone
  • Automation should be collaborative not something that happens to the user
  • Context is critical the quality of agent decisions depends entirely on the quality of context it receives
  • Plans are the interface making the agent's intent visible and editable is what makes human-in-the-loop real

Outcome

ActionEngine transforms fragmented, repetitive tools into collaborative teammates systems that learn alongside engineers, handle the toil, and stay transparent throughout. The future of AI in engineering workflows isn't just intelligence. It's alignment with human intent.

Project in active development. Happy to share more on the agent model, UI design, or implementation patterns.

Impact

75% Faster app creation
80% Fewer support tickets
1–2 Clicks to resume (was 5–6)

The Problem

GEP's 550+ clients needed to configure and update 20+ enterprise applications but even after a low-code studio launched, app deployment still took up to 5 months. Clients remained fully dependent on GEP's engineering and customer management teams for any change, and support requests barely moved (only 6% reduction).

Long cycles

Updating applications could take 2 weeks to several months depending on scope.

Engineering dependency

Any change no matter how small required a development team sprint allocation.

No consistency

Without standard templates, every app felt different. Clients wanted flexibility but got fragmentation.

Research

Conducted remote task-based usability testing with application developers and customer support reps (sessions lasting 1–1.5 hours). Four key pain points emerged:

  • Too many clicks 5–6 clicks just to pick up where they left off between sessions
  • No in-context documentation first-time users had no guidance within a complex tool
  • Iconography without context left nav and header icons were not intuitive without labels
  • Legacy jargon technical architecture terms appeared in the UI, confusing non-developer users

Design Process

Iteration 1 Persona-based home screen

Returning users saw a "pick up where you left off" widget to resume in 1 click. New users saw guided building options and tutorial paths. Added toggleable tooltips and replaced icon-only navigation with text + icon labels. All 6 returning users in testing validated the widget.

Iteration 2 Reducing cognitive load

Decluttered home screen: banner replaced full-image background, application cards simplified to name + status only. Working area expanded by modifying navigation layout. Tabs introduced in the header for cross-module navigation without full-page transitions.

Removing jargon

Worked with backend teams to handle legacy architecture terms under the hood surfacing them only in advanced settings, keeping the primary UI clean for all user types.

Outcome

Clients could make and preview changes to their applications in real time with zero dependency on engineering. App creation time down 75%. Support tickets down 80%. Click depth to resume work reduced from 5–6 to 1–2.

Impact

50% Faster setup time
20+ Users in discovery workshop
15+ Users in validation interviews

The Problem

Every new GEP client onboarding required setting up an organizational structure, uploading master data, defining rules and workflows, and configuring features across 20+ applications. The process was complex, unsequenced, and opaque. Client support and admin teams constantly needed hand-holding from GEP's product and engineering teams.

"What is the process to setup? There are so many options on the screen, where do I start?"

Research

Discovery Workshop

A 3-hour remote workshop with 20+ expert users surfaced recurring patterns: no defined sequence, invisible prerequisites and interdependencies, and constant hand-holding while experienced users were comfortable with the existing screens.

Validation Interviews

Online focus groups and individual interviews with 15+ experienced internal team members confirmed the real setup sequence: always starting with organizational configuration, then per-application setup, with different go-live dates per app. This validated organizing Guided Setup by application rather than by function.

Design Process

Information Hierarchy

Explored two organizational models by application vs. by functionality. User testing showed strong preference for application-based organization, aligning with how clients actually go live.

Journey-Based Entry

When logging into a new domain for the first time, users choose their path: Guided Setup (for new/novice users) or the familiar DIY home screen (for experienced users). This decluttered the home screen without removing functionality.

Guided Setup Features

Progress tracking with per-application drill-down checklists. "Next recommended steps" surfaced contextually after each completed item. Prerequisites and interdependencies made visible inline with help content.

What Worked

  • Progress tracking and drill-down checklists per application were highly valued by all users
  • "Next recommended steps" after completing items was appreciated across new and existing users
  • Separation of new vs. returning user journeys significantly reduced home screen clutter

Outcome

Setup time reduced by up to 50%, with significantly less dependency on GEP support and engineering teams. Designs were validated across a mixed group of 12 new and experienced users before handoff to development.

Outshift by Cisco

ActionEngine

Designing a human-in-the-loop agentic system that automates engineering workflows, not by replacing developer intent, but by learning from it.

Role

Lead Product Designer

Team

Design · Engineering · AI Research

Environment

Browser · Terminal

Framework

HAX Principles

ActionEngine interface: browser, terminal and chat panel working in parallel

ActionEngine in action: browser and terminal running in parallel, with the agent surfacing each step in the chat workflow panel

An agent that works with engineers, not around them

ActionEngine is an agentic system that operates across browser and terminal environments, bringing together multiple agents for planning, contextual understanding, and performance evaluation. It collaborates with engineers by observing and learning from their actions, then creates reusable workflows to automate those tasks.

The core design challenge was not the architecture; it was the collaboration model. How does an agent integrate into an engineer's existing workflow without disrupting it? How does it earn trust incrementally, stay visible, and remain correctable at every step?

Rather than treating the agent as a standalone tool, we approached it as an extension of developer intent: something that acts in alignment with user goals, not one that requires explicit instruction at every turn.


The agent watches and learns

Training starts with the engineer doing, not describing. As they work, the agent observes each action in real time and surfaces it inline within the conversational UI as a discrete, labelled step. There is no separate recording panel or background process. Every step appears as it is captured, giving the engineer a continuous view of what the agent is building and the ability to flag or correct anything before it becomes part of the workflow.

Agent observing and recording steps during user training

Each action is captured and surfaced inline as the engineer demonstrates the task

Learning in the open

Keeping observation visible in the conversational flow meant engineers never had to wonder what the agent was capturing. What you see is what gets automated.

Correction as you go

Because each step appears immediately, engineers can intervene while context is fresh rather than catching errors in a final review when the workflow is already fully formed.

No mode switching

Training happens within the same conversational interface used for everything else. There is no separate recording mode to enter or exit, reducing friction and cognitive overhead.


The agent asks when it does not know

When the agent hits a point it cannot resolve from observation alone, it does not guess and continue. It stops, surfaces the ambiguity as a direct question, and waits for the engineer to respond. This was one of the most deliberate design decisions in ActionEngine. Silent assumptions compound quietly. An agent that admits uncertainty turns a potential error into a collaborative moment, and keeps the engineer meaningfully in the loop rather than just nominally so.

Agent pausing and asking for clarification when stuck

The agent pauses mid-training and surfaces a specific question before continuing

Uncertainty made legible

The question is specific, not generic. Engineers see exactly where the agent got stuck and why, which makes it fast to resolve and builds confidence that the agent is reasoning, not just recording.

Error prevention, not error recovery

Catching ambiguity mid-training means incorrect assumptions never reach the final workflow. Fixing a question is far cheaper than debugging a workflow that ran on bad inference.

Trust through honesty

An agent that admits what it does not know is more trustworthy than one that fills gaps silently. Asking for help is not a weakness in the design. It is the design.


Review before anything runs

Once training is complete, the agent does not execute immediately. It presents the full workflow as a structured, editable draft and asks for approval. Every step is visible and can be modified, reordered, or removed. Only once the engineer signs off does automation begin. This is where intent is confirmed, not assumed. The workflow is not handed off; it is co-signed.

User reviewing and modifying a workflow draft before execution

The agent presents a full structured draft; the engineer edits a step before approving execution

A shared object, not an output

Framing the draft as something to be edited together, rather than approved or rejected, shifted how engineers engaged with it. They treated it as a starting point, not a final answer.

Full visibility before execution

The structured format makes it easy to read the workflow as a sequence of real actions, not abstract steps. Engineers know exactly what is about to run, and why each step is there.

Explicit approval as a feature

Requiring sign-off before execution is not a friction point; it is a trust-building mechanism. Engineers who approved the draft were far more confident when the workflow ran.


Automation that stays transparent and aligned with human intent

With ActionEngine, the goal was to design an agentic system that automates engineering workflows while keeping humans genuinely in control, not as a formality, but as a structural property of how the system works.

The agent works collaboratively: staying visible while learning, surfacing uncertainty rather than guessing, and treating every workflow as a shared artefact before handing it off to execution. The result is automation that engineers can trust, not because it is infallible, but because it is legible, correctable, and designed to ask for help when it needs it.

The future of agentic design is not just capability: it is alignment with intent. Agents that surface their reasoning, admit uncertainty, and hand control back to users will earn trust in ways that autonomous systems never can.


Grounded in HAX: how the principles showed up in practice

ActionEngine was built in parallel with the HAX framework, and every major design decision maps back to one of its five principles. The three flows above are where that grounding became concrete.

PrincipleHow it showed up in ActionEngine
ClarityEvery step the agent records is surfaced inline, with enough context for engineers to understand what was captured and why.
ControlEngineers can edit, reorder, or remove any step in the draft before a workflow ever runs. Approval is explicit, not assumed.
RecoveryAmbiguities caught mid-training can be resolved immediately. Nothing incorrect propagates forward into the final workflow.
CollaborationTraining is a shared act. The agent builds from what the engineer does, asks when it is uncertain, and presents a draft both parties shape before execution.
TraceabilityThe full workflow is reviewable as a structured sequence at every stage, from live observation through final approval.

Next project

GEP Quantum Studio →

GEP · Lead Designer · 2022–2024

Redesigning Quantum Studio

Redesigning an existing low-code tool to meet business and user goals: empowering admins and partners to configure applications without engineering, and cutting deployment time from 4–5 months to 1 week.

Role

Lead Designer

Timeline

2022–2024

Platform

Supply Chain & Procurement
Enterprise Web App

Company

GEP · B2B
Internal & Client Teams

Quantum Studio redesigned application overview

The numbers

4–5 moDeploy time, before
1 wkDeploy time, after
70%+Reduction in Jira tickets (target)

A low-code tool for supply chain applications

Quantum Studio is GEP's low-code platform used to create supply chain and procurement applications from scratch or using templates. It features data modeling, a UI Designer, and automation capabilities. The platform is used by three personas: developers who build from scratch, admins who customize applications for specific clients, and partners involved in onboarding and implementation.

The three main goals of Quantum Studio are to ship applications faster, empower admins and partners to configure without developer help, and increase developer efficiency.


Version 1 didn't move the needle

After launching the first version of the tool, there was no significant reduction in the changes clients needed to make. Admin and partner teams were constantly reaching out to internal product and development teams to make changes or update their applications. This meant developer bandwidth was consumed by redundant tasks, and deployment time remained at 4–5 months. Our goal was to bring it down to 2 weeks.

Engineering dependency

Every change, no matter how small, required a development team sprint allocation.

Long deploy cycles

Application deployment took 4–5 months. The business needed this cut to 1 week.

Lack of personalization

No customization based on persona type. Advanced features overwhelmed non-technical admin users.

Lack of standardization

Information was not grouped in standard categories, creating inconsistency across the 36 config panels.


Clear goals, measurable targets

Going into this project the objective was clear. We needed to reduce deployment time from 4–5 months to 1 week, and bring down the number of Jira support tickets by 70%+ by the end of the year. To get there, we first needed to understand how all three personas actually created applications.


Two very different users, one shared platform

The platform served two primary personas with fundamentally different needs. Developers are technical users who create applications from scratch, develop reusable plugins, and need advanced development features. Admins are non-technical users who customize applications for specific clients, are involved in onboarding and implementation, and need to preview and deploy in real time. Designing the same features for users with varying levels of technical expertise required careful consideration at every step.


Understanding where the user flow broke down

To reach our goal, we needed to understand how all three personas created applications. We mapped out the user flow for application creation and tested each flow to identify what was not working. We analyzed customer pain points, exits, and needs, then mapped all problem areas against product features using affinity mapping.

Prioritization and affinity mapping session

User testing: prioritization and affinity mapping


Config Panels: 36 panels, two big problems

As the platform grew, two major problems surfaced in the configuration panels, which spanned 36 panels across the application. First, lack of personalization: there was no customization based on persona type, meaning the same advanced features seen by developers were also shown to admin users, creating cognitive overload. Second, lack of standardization: information was not grouped into standard categories, making navigation inconsistent and unpredictable.

Config Panels existing design: overlay hid information behind, users couldn't see changes simultaneously

Problem 1: the config overlay hid the page, so users couldn't see changes in real time

36 config panels: both problems annotated: overlay hid information and back-and-forth to preview changes

Problem 2: to preview any change, users had to go back and forth with no undo


New design principles for Config Panels

The insights from research gave us clear goals for improving the config panel experience. The new design was built around three principles: in-context editing instead of a slider with overlay, decluttering by removing unnecessary fields and customizing panels per persona, and grouping all fields into relevant standard categories.

Config Panels final design: condition builder using natural language constructs and AI-generated logic

Final Design: condition builder with natural language constructs + AI-generated logic

Redesigning the Expression Builder

Expression builders can be challenging to use due to their complexity in handling various data structures, logic operations, and user interfaces, often requiring users to possess deep programming knowledge. Our users were struggling with ours. The insight: condition-based expression builders are easier to understand because they allow users to construct logic by setting conditions and actions in a more intuitive manner, resembling natural language constructs. We redesigned to showcase this flow.

Config Panels Redesign Variation 1 - Expression Builder redesigned to showcase flow

Variation 1: Expression Builder redesigned around condition-based logic flow


What this project taught us

The GEP technical architecture heavily influenced how the low-code studio was designed. Designing identical features for user personas with varying levels of technical expertise requires careful consideration and planning.

Impact of legacy architecture

The technical architecture shaped design constraints in ways that weren't immediately obvious. Identifying those constraints early prevented late-stage rework.

Working with engineers early

Bringing engineers in from the start helped the team design within real constraints rather than discovering blockers after handoff.

Research-led roadmap

Research guided prioritization across problem areas, making the approach scalable as the platform continued to grow.

Next project

GEP Guided Setup →

GEP

GEP Guided Setup

Redesigning the onboarding experience for enterprise clients so they could configure applications independently without hand-holding from product or engineering teams.

Role

Lead Designer

Research

Workshop · Interviews
Focus Groups

Users

20+ in discovery
15+ in validation

Outcome

Setup time ↓ 50%

GEP Guided Setup: enabling new users to set up applications

The numbers

50%Faster setup time
20+Users in discovery workshop
15+Users in validation interviews

Clients got lost every time they onboarded

Every new GEP client onboarding required setting up an organizational structure, uploading master data, defining rules and workflows, and configuring features across 20+ applications. The process was complex, unsequenced, and opaque. Client support and admin teams constantly needed hand-holding from GEP's product and engineering teams.

"What is the process to setup? There are so many options on the screen, where do I start?"


Discovery workshop + validation interviews

Discovery Workshop

A 3-hour remote workshop with 20+ expert users surfaced four recurring patterns: no defined sequence for setup, invisible prerequisites and interdependencies, constant hand-holding required, and no way for new users to know where to start.

Research: heatmaps and usage data showing where users struggled

Heatmaps and behavioural data: 51% of users spent 43s on key tracked events; 40% struggled to find Setup Manager

Validation Interviews

Online focus groups and individual interviews with 15+ experienced internal team members confirmed the real setup sequence: always starting with organizational configuration, then per-application setup, with different go-live dates per app.

This validated our decision to organize Guided Setup by application rather than by functionality matching how clients actually go live.

Solution: Application Wise Setup matrix, organising setup by application not by functionality

Application Wise Setup: the validated structure, organised the way clients actually go live


From cluttered home screen to journey-based entry

Journey-Based Entry

When logging into a new domain for the first time, users choose their path: Guided Setup for new and novice users, or the familiar home screen for experienced users. This decluttered the home screen without removing any functionality.

New users got lost: the overwhelming Setup Manager home screen

"What is the process to setup? There are so many options on the screen, where do I start?"

Wireframe Variations

We tested two wireframe variations for the Guided Setup widget. Users preferred the version that gave an overview of progress upfront with a checklist aligning with their expectation of application-by-application go-lives.

Visual Design Concept 1: Guided Setup widget with progress tracking and checklist

Visual Design Concept 1: Guided Setup widget with progress tracking and per-application checklist

Visual Design Iterations

Early visual designs combined the Guided Setup widget with full home screen elements which overwhelmed users. We separated the experiences entirely, giving Guided Setup its own dedicated space.

Guided Setup Features

The final design includes progress tracking with per-application drill-down checklists, "Next recommended steps" surfaced contextually after each completed item, and prerequisites and interdependencies made visible inline alongside help content.

Final prototype: Welcome to GEP Guided Setup with application-by-application progress

Final prototype: application-by-application progress, help centre integration, and recommended path

Validated across new and experienced users

  • Progress tracking and drill-down checklists per application were highly valued by all 12 test participants
  • "Next recommended steps" after completing each item was appreciated across new and experienced users
  • Separation of new vs. returning user journeys significantly reduced home screen clutter
  • Users wanted to hide applications once setup was complete flagged for next iteration

Setup time cut by 50%

Setup time reduced by up to 50%, with significantly less dependency on GEP support and engineering teams. Designs were validated across a mixed group of 12 new and experienced users before handoff to development.

Next project

Agentic Security →

Outshift by Cisco

Agentic Security

Designing transparent multi-agent systems with the Visibility Funnel, making distributed AI reasoning understandable and trustworthy for security analysts.

Role

Lead Product Designer

Domain

Security AI · Multi-Agent

Framework

HAX Principles · Visibility Funnel

Output

Design Framework · UX Patterns

Research in progress — this project is exploratory work that will eventually be integrated into Cisco security teams. Designs are research-stage artifacts, not shipped product.

Agentic Security dashboard overview

Dashboard: system health, agent compliance, and recent fraudulent interactions

Security AI is no longer just about detecting threats it's about making distributed reasoning understandable

Modern security systems are becoming increasingly autonomous and distributed. Multiple specialized agents observe activity, interpret signals, and build on each other's outputs. The challenge is no longer just detection accuracy it's making that distributed reasoning visible, traceable, and trustworthy to the analysts who depend on it.

This work explores how UX design can bring clarity, control, and transparency to complex multi-agent security environments applying the HAX framework through a structured model called the Visibility Funnel.


Analysts can see what agents decided. They can't see why.

When multiple agents collaborate on a security finding, each contributes a fragment observing, interpreting, flagging. But their reasoning is often fragmented, insights can conflict or overlap, and there's no coherent picture of how a conclusion was formed.

Incident list with fraud scores and confidence

Incidents view: analysts see fraud scores and confidence values, but no reasoning behind them

The critical gap: Lack of visibility into agent reasoning reduces trust, limits control, and makes it nearly impossible to investigate or correct errors in a multi-agent system.

When something goes wrong, analysts can't tell which agent introduced the error, whether it propagated, or how to recover without losing context. Trust erodes and with it, the value of the system.


From observing actions → to understanding intent

Traditional security tools focus on what agents do their outputs and actions. Agentic Security introduces a different layer: analyzing reasoning instead of just outcomes.

Detective Agent incident report showing reasoning chain

Incident Report: the Detective Agent surfaces its reasoning, verdict, and compliance failures in a structured view

Reasoning-Level Threat Detection

Detect when agent intent is altered, reasoning paths are corrupted, or decisions appear valid on the surface but are fundamentally flawed.

Emerging Threat Protection

Designed to catch memory poisoning, false information propagation, inter-agent trust exploitation, and intent divergence threats that traditional tools miss entirely.

Intercepting Threat Propagation

Stop corrupted reasoning at transfer points before it cascades across agents, preventing a single error from compounding into a system-wide failure.

Traceable Reasoning

Every insight carries intent-level evidence, root-cause explanation, and full traceability so analysts can understand and verify the path from signal to conclusion.


The Visibility Funnel: transforming distributed activity into coherent narrative

The Visibility Funnel is a four-layer model that structures how multi-agent activity gets surfaced to analysts. Each layer builds on the one below from raw knowledge to actionable insight.

The Visibility Funnel: four layers from Knowledge/Memory to Surfaced UI

The Visibility Funnel: four layers from raw knowledge to observable UI, each building on the last

1 · Knowledge Layer

The foundation

Entities and relationships, historical patterns, policies, and agent memory the raw material that agents operate on.

2 · Reasoning Layer

Where agents think

Agents propose hypotheses, validate relationships, challenge assumptions, and flag inconsistencies. This layer is typically invisible the Visibility Funnel surfaces it.

3 · Cognitive Artifacts

Structured evidence

Evidence packets, anomaly clusters, policy validations, contradiction flags, and hypotheses structured outputs that preserve reasoning in a reviewable form.

4 · Surfaced UI

What analysts see

Timelines, entity-level insights, investigative paths, uncertainty indicators, and drill-down capabilities designed for fast comprehension and confident action.


No single agent has the full picture value emerges from alignment

The system is built around a multi-agent ecosystem where each agent contributes a specialized perspective. Designing for this meant understanding not just what each agent does, but how their outputs relate to and depend on each other.

Knowledge base graph showing agent relationships

Knowledge Base graph: agent relationships, active incidents, and connected entities visualised together

Drift Detection Agent

Identifies gradual shifts in agent behavior that may signal compromise or degradation over time.

Fraud Agent

Analyzes transactional and behavioral patterns for intent-level anomalies and coordinated manipulation.

Observability Agent

Monitors system-wide activity streams, surfacing signals that other agents can act on.

Detective Agent

Synthesizes outputs from all other agents to form the final investigative narrative and recommend next steps.



Proof Trees: letting analysts follow the thread of reasoning

In complex multi-agent systems, a single finding is rarely the product of one agent. It emerges from a chain, with one agent observing, another interpreting, a third validating. Without a way to navigate that chain, analysts are left with a conclusion and no path back to the reasoning that produced it.

Proof Trees are a UX pattern designed to solve exactly this. They expose the full decision chain, from the top-level verdict all the way down through each contributing agent, as a structured, navigable tree. Each node shows what the agent concluded, what evidence it used, and how confident it was. Analysts can expand any node to interrogate that step, drill into the evidence, or view the inter-agent discourse that shaped the decision.

Proof Tree showing agent reasoning chain from Detective Agent through AgentAuditor to EvoGuard

Proof Tree: Detective Agent flags a fraud case. Expanding the tree reveals AgentAuditor's classification rationale and EvoGuard's fingerprint match, each step linked to its evidence.

The key design principle: the tree is not a log. It is a navigation structure. Every path through it answers a question an analyst would actually ask: "Which agent decided this?", "What evidence triggered that conclusion?", "Where did the reasoning diverge?" The Proof Tree makes those questions answerable without requiring analysts to switch context or reconstruct the chain manually.

Proof Trees turn multi-agent reasoning from a black box into a map, one that analysts can follow, interrogate, and trust.

This pattern applies wherever agents build on each other's outputs. In security it means tracing a fraud verdict back through fingerprint analysis and intent classification. More broadly it is the foundational affordance for any system where explainability must travel with the decision, not live on a separate page.

Trust by design the same principles, applied to security

Each HAX principle maps directly to a specific design challenge in multi-agent security systems.

PrincipleHow Agentic Security applies it
ControlAnalysts can pause, redirect, or override any agent at any point. The system respects human judgment without creating friction for routine decisions.
ClarityEvery finding surfaces what each agent detected and why not just the conclusion, but the reasoning chain that produced it.
RecoveryWhen an agent makes an error, analysts can correct it without losing the surrounding context or forcing the system to start over.
CollaborationA continuous human-agent feedback loop ensures analyst corrections improve the system's future reasoning not just the current case.
TraceabilityEvery insight is linked to its source agent, evidence, and reasoning path. Full audit trails are available without requiring extra steps from the analyst.
HAX principles in the analyst UI

HAX in practice: Clarity (Detective Agent's reasoning surfaced), Traceability (source attribution), Control (compliance flags with actionable states)


Agent reasoning that travels not just lives on one screen

One of the core design challenges was ensuring that the context, evidence, and reasoning behind a finding doesn't get lost as analysts move between views. Portable Transparency means agents carry consistent UI patterns across surfaces.

Agent Discourse panel showing inter-agent communication

Agent Discourse: the conversation between AgentAuditor and Detective Agent, surfaced inline within the analyst view

Activity Overview

A scannable summary of what each agent has been doing visible across surfaces without requiring a dedicated deep-dive view.

Context Map

Relationships between entities, agents, and findings presented consistently whether viewed in a timeline, a detail panel, or a notification.

Impact Trace

A linked chain from detection signal to current finding so analysts always know how a conclusion was reached, regardless of where they encounter it.


What this work revealed about designing for multi-agent trust

  • Visibility is the product in multi-agent systems, making reasoning visible is not a feature. It is the core value proposition.
  • Structure reduces anxiety analysts working with complex agent ecosystems don't just need information. They need it organized in a way that mirrors how they think through a problem.
  • Errors compound silently without visible reasoning, a single agent error can propagate through a system before anyone notices. Transparency is the only defense.
  • Trust is earned incrementally the system needs to consistently behave as explained, surface what it got wrong, and give analysts a path to correct it without penalty.

Transparency as the foundation of trust

The Visibility Funnel framework provides a reusable model for any domain where multi-agent reasoning needs to be surfaced to human decision-makers. Applied to security, it transforms distributed agent activity from a black box into a structured, navigable, and correctable investigation one where analysts remain genuinely in control.

The future of AI-powered security isn't just more accurate detection. It's detection that analysts can understand, verify, and trust.

Next project

GEP Mobile App →

GEP · Lead Designer

GEP Mobile App: Requisition Creation

A complete redesign of the GEP mobile app adding requisition creation, modernising the experience, and enabling a new class of on-the-go procurement workflows.

Role

Lead Designer

Team

PM · Dev Lead · User Researcher

Responsibilities

UX · IA · Visual · Interaction · Stakeholder

Platform

iOS · App Store

GEP mobile app redesigned key screens
12%Increase in mobile requisitions in 2 months
19%Reduction in requisition approval time
4New enterprise clients won via sales demo app

The app was outdated, incomplete, and blocking users from working on the go

The existing GEP mobile app had two compounding problems. First, it wasn't persona-specific it treated all users the same regardless of role, which meant the most frequent and high-value use cases weren't prioritised. Second, it had significant feature and visual gaps compared to the web experience, making it feel like a lesser alternative rather than a capable companion.

The most critical missing capability: users couldn't create a requisition from mobile at all. Anyone who needed to initiate a purchase while away from their desk had to wait until they were back slowing approval chains and creating frustration for clients who had specifically asked for this capability.

"Users who were on the go had to wait to get to their desktops to create a requisition a feature particularly important for clients actively using the app."

Existing GEP mobile app: dated UI, no requisition creation, role-unaware navigation

The existing app: no creation capability, role-unaware navigation, outdated visual language


Work on the go, without compromise

Work on the go

Give all personas including managers access to the information they need to take decisions instantly, without being tied to a desktop.

Provide flexibility

Enable users to create, edit, view, and approve procurement documents from mobile matching the flexibility that competitors were already delivering.

Reduce time

Remove the lag that desktop-only submission, approval, and tracking introduced. Process requests in real time, wherever users are.


Mapping primary personas to understand the real usage landscape

Based on client needs and current usage data, we mapped the primary personas. This helped identify distinct flow needs across user types from approvers who primarily take action on existing documents, to requesters who need to initiate procurement workflows from scratch.

User personas: Field Worker, Manager, Executive, and their mobile needs and priorities

Three primary personas: Field Worker (create on the go), Manager (quick approve/reject), Executive (dashboard metrics)


Competitive analysis revealed four features our app was missing

We analysed direct competitors to understand the landscape and identify gaps. Four patterns stood out as near-universal across other procurement mobile apps none of which were present in the existing GEP app.

Competitive analysis: SAP, Oracle, Coupa, Procede, Proactis

Competitive landscape: SAP, Oracle, Coupa and others all offered capabilities missing from GEP mobile

Pending actions on home

Most apps prioritised surfacing tasks that needed attention directly on the home screen not buried in navigation.

Upfront approve / reject

Users could take the most critical actions approve or reject without needing to enter document details first.

Bulk actions

Approvers could act on multiple documents at once from the landing screen, saving significant time for high-volume users.

Alerts and notifications

Proactive alerts surfaced actions requiring immediate attention rather than expecting users to check in manually.


Low-fidelity first establishing information architecture before visual direction

We started with wireframes to nail down the information architecture and test the structure before investing in high-fidelity design. The goal was to validate navigation patterns, action placement, and creation flow with real feedback.

Wireframes: home screen with pending actions, create requisition entry, and form structure

Wireframes: home screen, creation entry, and form structure validated before visual direction

Focus group with the internal solution design team surfaced three clear issues

We ran a focus group with GEP's internal solution design team the team responsible for working directly with clients and understanding their mobile needs. They gave us direct, specific feedback across three areas.

Navigation

Settings appeared too prominently despite infrequent use. Not all users have access to Dashboard and Create these needed to be role-gated rather than globally visible.

Actions

Adding quick actions on the home screen was the right direction but users wanted to take action without entering detail screens at all. Upfront approve / reject needed to be a first-class pattern.

Creation flow

The three-option creation approach was well received. The team recommended leveraging GEP AI capabilities to make the process smoother and faster reducing manual entry where possible.


Key screens

With structure validated, we moved into high-fidelity design applying the updated GEP visual language and building out the full key-screen set for development handoff.

Key screens: home with pending actions, catalog, and requisition details with approve/reject

Key screens: home with pending tasks and role-aware navigation, catalog, and requisition detail with inline approval

Three paths to create with AI-assisted entry

We designed three creation options to serve different user contexts and preferences, with AI assistance woven into the flow to reduce manual input and speed up the process.

Create requisition: three paths: Requisition Form, Minerva AI Chatbot, and guided creation flow

Three creation paths: standard form, Minerva AI chatbot, and guided flow with progress tracking and mandatory sections


Shipped to the App Store and immediately put to work

We launched a fully working version of the app on the iOS App Store. The sales team used it immediately to pitch to new clients the working app became a sales tool in its own right, enabling demonstrations of real procurement workflows on a real device.

4New enterprise clients added via sales enablement
19%Reduction in time to approve requisitions post-launch
12%Increase in requisitions created on mobile in 2 months

Back to portfolio

← All Work

Outshift by Cisco

HAX Human-Agent Experience

From the Internet of Agents to Portable Explainability: how we designed, built, and shipped a unified framework for human-agent collaboration.

Role

Lead Product Designer

Team

Design, Design Ops,
Engineering, AI Research

Output

Framework · 5 Principles
React SDK · Open Standard

Status

Shipped · Adopted across
Cisco product teams

HAX Human-Agent Experience framework

Every team at Cisco was solving the same problems from scratch

We were building AI-powered features across multiple products. And every team kept asking the same questions: How should an agent explain what it's doing? How much control should the user have? What happens when something goes wrong?

There were no shared answers. Agent behavior was inconsistent across products. Transparency was limited. Tooling was fragmented. And none of it scaled as agent ecosystems grew.

We kept seeing the same pattern: a team ships an agent-driven feature, users don't trust it, and the team bolts on explanations and controls after the fact. It was clear we needed a shared foundation before the problem got worse.

The opportunity: standardize how agents communicate and commit across surfaces before the ecosystem fragmented further.

A fundamental change in how humans work

We weren't designing for chatbots anymore. Users weren't just interacting with an AI to get answers they were conversing with independent agents that act on complex, multi-step workflows as part of a business process. That shift changed everything about what good design meant.

Simple Automation vs. Autonomous Agents Working With Humans

Four categories of friction kept surfacing

Gaps in Contextual Understanding

Sharing more than data: Agents shared raw data instead of complex states, limiting meaningful collaboration.

Missed human nuances: AI failed to grasp team dynamics, intuition, and shared understanding.

Probabilistic Operations

Inconsistent outcomes: Variable, unpredictable results disrupted workflows.

Misinterpretation of intent: Agents made decisions that deviated from team priorities.

Challenges in Mental Models

Hard to keep up: Evolving AI systems made it difficult to maintain accurate mental models.

Divergent operational logic: Agent behavior didn't align with how users expected things to work.

Trust Over Time

Initial overconfidence: Early errors caused steep trust decline after high initial expectations.

Trust through reliability: Long-term trust only builds through consistency and transparency.

From screens to states

The most fundamental reframe: we weren't designing pages and flows anymore. We were designing persistent decision states that evolve with the agent and the user.

Screens: flight booking form: Searching, Selecting, Confirming, Restarting

Traditional UX: discrete screens that reset with each action

States: travel planning agent: Thinking, Gathering, Proposing, Awaiting, Executing, Recovering

Agentic UX: persistent states that carry context, evidence, and explanation forward

Traditional UX designs screens discrete layouts that reset with every user action. Agentic UX requires designing states: continuous, context-carrying moments where the agent is thinking, gathering, proposing, awaiting, executing, or recovering.

"When conversation, context, and UI are all fused… UX becomes the bridge between what's understood and what's visible."


How we built the knowledge base

HAX started as a research effort to understand what makes human-agent collaboration effective, predictable, and trustworthy. We ran experiments across agentic product prototypes, observed human decision patterns, and interviewed teams building with LLMs and automation frameworks across five interdisciplinary research areas.

Research areas behind HAX: five interdisciplinary areas
  • Foundational Principles Evidence-based design principles for intuitive human-agent interfaces
  • Cognitive Frameworks How humans and AI agents process information and make decisions collaboratively
  • Ethical AI Translating the ethical dimensions of agentic systems into practical design considerations
  • Societal Impact Broader implications of AI-human collaboration on work and society
  • Security & Privacy Safeguarding collaborative systems while preserving user trust

What we learned: Consistent patterns emerged around explainability, control, recovery, and collaboration shaping a shared model for how agents should communicate and act.

How we narrowed to five principles

The principles didn't come from a whiteboard session. They emerged from failure patterns we kept seeing across the research places where agent experiences broke down in similar ways, regardless of domain or product.

When things went wrong, they went wrong in one of a small number of predictable ways:

  • Users couldn't tell what the agent was doing or why it made a choice
  • Users didn't feel like they could change the agent's behavior, or didn't know how
  • When the agent made a mistake, there was no clear path to fix it
  • The agent acted independently with no sense of shared work
  • There was no way to look back at what had happened or understand why

Those failure modes mapped directly to what became the five principles. We stress-tested the set against agent scenarios across Cisco products. After multiple rounds of review with engineers, AI researchers, and product teams, the set held.

Five research-based design principles

PrincipleDesign Rule
ControlDesign clear limits for agent behavior. Define boundaries and provide intuitive controls so users can adjust agent behavior easily.
ClarityMake agent reasoning easy to understand. Surface decision-making processes in user-friendly formats no black boxes.
RecoverySupport safe reversal of actions. Plan for edge cases with clear, actionable recovery steps that maintain continuity when AI encounters challenges.
CollaborationDesign agents to work with people, not around them. Create intuitive handoffs that preserve context, ensuring AI amplifies human potential.
TraceabilityExpose the source of information and confidence. Make it easy to review past AI actions so users can assess where the AI excelled or fell short.

A unified framework not just a principles doc

HAX connects principles to the tools, components, and behavioral checks developers use every day. It has four interconnected parts:

HAX SDK and Component Library: SDK, Components, Add-Ons, Custom Repos

The HAX SDK and Component Library: UI + semantics + light governance, out of the box

Principles

Design for collaboration

Five research-based rules that define trustworthy agent behavior across any domain.

SDK

Build with consistency

Turns principles into schemas, components, and behavioral checks. UI + semantics + light governance, out of the box.

Custom Repositories

Reusable explainability

Organization-owned catalogs for approved agents, components, prompts, and behaviors.

Portable Explainability

Consistency everywhere

A behavior layer that travels with the agent same reasoning and evidence across every surface.

From developer code to shared behavioral components

The HAX SDK is a schema-driven, React-based component library that helps developers build agents with consistent behavior, explainability, and user control. Today, developers hand-build each agent's logic and explanation from scratch. With HAX, approved behaviors are encoded as shared, reusable components with schema, policy, and explainability built in.

HAX SDK component library

SDK Components

Source AttributionTimelineDetails DashboardData VisualizerMindmapForm BuilderCode EditorInline RationaleCapability ManifestContextual ExplanationFindingsDiagnostic ReportThinking ProcessWorkshop CardCo-Editing InterfaceRich Text Editor

Principles in practice: the Network Agent

To demonstrate HAX applied end to end, we built the Network Agent, a Cisco AI agent that monitors network health, proposes remediations, and collaborates with engineers in real time. Each screen maps directly to a HAX principle, making the framework tangible rather than theoretical.

Clarity: make reasoning visible

The agent surfaces suggested actions with clear rationale, impact labels, and explicit Accept / Modify / Reject controls. Users always know what the agent found and why it's recommending a specific action.

HAX Clarity: Network Agent surfacing suggested actions with Accept/Modify/Reject controls

Control: define limits for agent behavior

The Agent Control Panel gives users explicit, granular control over what the agent is allowed to do automatically and what requires manual approval. Permissions are time-bounded and auditable.

HAX Control: Agent Control Panel with Permission Adjustment and time-bounded rules

Collaboration: work with people, not around them

When the user asks the agent to block a router, the agent flags the risk and offers alternatives instead of executing blindly. The human picks a direction; the agent executes with the parameters the human specifies.

HAX Collaboration: human-agent conversation surfacing alternatives with quantified impact

Recovery: support safe reversal of actions

When the agent's rerouting fails, it presents a structured error report: what happened, a timestamped failure log, and clear recovery actions (Inspect, Monitor, Rollback). Users can recover without losing context.

HAX Recovery: Error Details panel with failure log and rollback actions

Traceability: expose the history of agent actions

The agent surfaces a full history of similar actions it has taken, with dates and measured outcomes. Users can review past decisions, assess reliability, and optimize future behavior from a position of evidence.

HAX Traceability: Similar actions taken by agent with historical outcomes and View Details

That's the promise of HAX: behavioral collaboration that is predictable, reusable, and consistent by default, across every team and surface.


Shipped as an open standard

HAX shipped as an open standard and is being considered for adoption across multiple product teams at Cisco. It gave teams a shared language for designing AI-driven interactions and a practical toolkit for implementing them consistently.

The next phase is about how explainability itself can travel across systems maintaining trust and transparency through portable structures for reasoning, evidence, and confidence. The question isn't just how to build one good agent experience. It's how to make every agent experience trustworthy by default.

Visit outshift.design/hax →

Back to portfolio

← All Work

Beyond the day job

Side Projects

Things I build, write, and play with outside of work, design thinking in a different register.


Writing

Design Quorum

A newsletter exploring the messy, human side of design, where craft meets collaboration, and process meets politics. Essays on design leadership, decision-making, and what it actually takes to do good work inside complex organizations.

Read on Substack →

Image coming soon


Game Design

Rebus

A word puzzle game built around visual logic, with images standing in for syllables and sounds, rewarding lateral thinking over rote recall. Designed with the same attention to clarity and delight I bring to product work, just with fewer stakeholders.

Play Rebus →

Image coming soon