Geunhee Lee
works
about
Financial AI Agent UX Research
Financial AI Agent UX Research
2026
what I did

UX Research Project Lead

with

HX Research

more
Trusting AI with Wealth: Designing for Control and Collaboration

This project explored how users perceive a financial AI agent's intervention in managing their wealth. By understanding the psychological and behavioral boundaries of trust, we designed a research framework to determine what factors lead users to accept or reject automated financial tasks, translating these insights into actionable UX directions.

Disclaimer: AI-generated mockups are used to protect the confidentiality of this pre-launch project.
Process

Interest Area and Context

In the sensitive domain of finance, the concept of AI automation carries unique risks. Our initial exploration revealed that users overwhelmingly prefer a manual invocation model—they want to actively request help from the AI. "Automatic intervention" by the AI was quickly perceived as a "loss of control" and triggered strong user resistance.

While users were generally open to AI for simple information retrieval or narrowing down options, they demanded much stronger assurance and strict manual control for actions involving actual money transfers or long-term financial decisions.

User Research & Prototyping

To understand these boundaries, we conducted an online survey (N=277) and Focus Group Discussions (N=18) using interactive prototypes of AI financial features. We evaluated user responses based on two main axes: the level of task risk and the degree of user control.

Problem Statement

We found that for users, AI automation in finance is not just a matter of functional convenience; it is deeply tied to their sense of control and accountability. The challenge was: How do we design an AI agent that reduces the burden of complex financial tasks without making the user feel like they have lost control over their own money?

Design Value and Direction

Users clearly perceived AI not as a final decision-maker, but as an 'assistive tool for preliminary judgment.'

Based on our findings, users preferred a collaborative approach where AI organizes complex information first, allowing them to verify the rationale before execution. We defined the following UX principles:

  • Rationale-First Explanations: Show the data and conditions behind the AI's answer before showing the final result.
  • Separation of Suggestion and Execution: Always provide a step where the user can review, modify, or hold the AI's suggestion.
  • AI as a Tool, Experts as a Safety Net: Provide a seamless transition to a human expert when making high-risk decisions, positioning experts as a necessary "safety net."

Strategic Roadmap

We proposed a phased integration strategy to gradually build user trust:

  1. Phase 1: Safe Helper — Focuses on building foundational trust by assisting with low-risk tasks and demonstrating analytical value.
  2. Phase 2: Smart Advisor — Introduces personalized financial advice with explicit user review and confirmation steps.
  3. Phase 3: Proactive Partner — Envisions the AI proactively detecting risks and offering personalized suggestions, carefully balancing the tension between "resistance to automation" and "expectations of utility."
Robo Investment UX
Retirement Goal UX