Samsung Bixby: A system level AI assistant

Introduction

Bixby is Samsung’s system-level AI assistant, with over 4 million monthly active users. This project aimed to transform Bixby into a more practical and capable AI assistant by introducing a suite of AI-powered tools and real-time conversational (RTC) capabilities.

My role

Design lead: User survey, workshop, competitors research, design strategy, design ideation

the team

2 Design assistants, 2 Product manager, 20+ Developers

Timeline

Jul 2024 ~ Dec 2025

Old Bixby AI QA flow

What was the problem

When I took over Bixby , it was limited to basic Q&A interactions, returning generic text results. Many users felt it was not useful for solving real tasks, highlighting the need to evolve Bixby into a more practical AI assistant.

User survey

With interviews with 20 AI-experienced users, a key insight emerged: Users expect AI to help complete real work and life tasks, not simply providing generic responses.

Understand users’ needs

From the interviews, we identified 3 primary user needs:

Personalization

Expect AI to deliver results tailored to their specific needs.

Efficiency

Want to quickly learn and adopt new AI tools to improve productivity.

Effortless Interaction

Prefer hands-free conversations with AI, without constantly tapping the screen.

The goal was to transform Bixby from a traditional voice assistant into a more capable AI assistant that supports real user tasks. Working within complex stakeholder constraints, I introduced practical AI features and established unified interaction guidelines to ensure a consistent and scalable AI experience.

Create new AI agents as utility toolkit

Design target

Key insight of user research

Our design target

Expect AI to deliver results tailored to their specific needs.

Introduce a set of scenario-based AI agents that provide personalized solutions for real user tasks.

Understand users’ needs

Based on insights of user interview, we prioritized 5 generative AI agents that address 80% of user scenarios.

Entertainment

Productivity

AI podcasts

PPT generate

Translate

“I want to explore and learn a topic deeply in a fun way.”

“I’d rather listen to the answer instead of reading.”

“I need fast translation for documents and emails.”

“I wish the content to be more organized, and easy to share.”

Music generate

Video generate

“Putting an AI videos on my social media is super cool.”

“I need to create my own background music for my video.”

Bussiness constraints

Bixby is a system-level product involving multiple stakeholders, feature details and decisions were heavily influenced by business and technical constraints.

Korean Design Center

Samsung Design China

Bussiness Center

Development Center

Defined the core Bixby AI interaction

Led the experience design of new features

Focused on partnership and monetization

handled implementation with a limited engineering team

The experience of new features had to remain consistent with existing core features.

UX often conflicted with CP’s requirements, causing frequent feature changes.

Limited engineering resources constrained design implementation.

My place

Design strategy

Given the business and technical constraints, delivering a perfect solution in one iteration was unrealistic, I proposed a phased design strategy:

Phase 1: Deliver core functionality in the simplest form, ensure a timely launch aligned with business and engineering needs.

Phase 2: Improve the UX based on feasibility testing and competitive analysis.

At the early stage, feature details were frequently changing due to business partnerships. To ensure fast delivery, I used Bixby’s existing conversational interaction for all AI agents. Within 3 months, we successfully launched the core functionality of 5 AI agents.

Solution

Stage 1: Run all AI agents following the current conversational interaction.

Main interaction of Music Generation agent

Key usage data

Key user feedback

Validating the solution

After launch, we tracked 2 weeks of usage data and conducted user testing.

Users were generally satisfied with the AI agents, but many reported frustration with the interaction flow, which felt complex and error-prone, leading us to focus on improving the interaction experience.

4.2/5.0

Result Satisfaction

Most users felt the AI-generated results met their needs.

1% <

Feature Adoption

Very few users tried the new AI agent features.

Low Error Tolerance

“Once I choose a parameter, it’s difficult to change it. Every step feels tense.”

Key insights from competitors

Agent name

index

Agent

Agent

Agent

Agent

Try this task with Agent 1

Try this task with Agent 2

Try this task with Agent 3

index

Competitors research

I believed the deep entry and multi-turn interaction flow were the main reasons behind the poor experience.

To explore more efficient interaction patterns, I analyzed six leading AI products and found out:

Embedding parameters directly in the input field significantly improves efficiency.

Proactively recommending relevant tools increases feature discoverability.

Doubao

GPT

Grok

Gemini

Yuanbao

Information architecture optimization

I redesigned the Agent flow in two ways:

Integrated key parameters directly into the input field, allowing users to configure them naturally.

Introduced proactive recommendations, helping users find features without manually typing in.

Current IA

Optimized flow

Select agent manually

Choose recom-mended agent

Enter prompt

Set parameter 1

Set parameter 2

Set parameter n

View available parameters

Select parameter to modify

Adjust parameter

Configuration complete

Configuration complete

Submit request

Submit request

Interaction Redesign

Increase visibility of AI agents by adding entries in home page.

Integrate parameter settings into the input field, making customization more flexible.

Before

After

1

1

2

2

Proactively recommend AI agents with clear task suggestions to help users quickly understand what each agent can do.

Optional parameter configuration allows users to personalize results while maintaining control of the process.

Add an AI agent entry above the input field using a capsule layout, enabling quick access to them.

Follow-up actions after generation are displayed as capsules above the conversation area, keeping the interaction consistent.

AI podcasts

PPT generate

Translate

Music generate

Video generate

Implementation

In the first release, we launched 5 AI agents with a unified interaction structure, which later became the design guideline for future agents.

Trade off

Although the redesigned received positive feedback in UT, implementation faced strong resistance from Korean Design Center.

After further discussions, I realized the key limitation:

Interactive components were hard to be embedded directly inside the input field in the current system architecture.

Solution

To work within these constraints, I redesigned the interaction by moving parameter configuration to a secondary page.

Although this introduced an extra step, it avoided the current limitations and allowed greater flexibility for complex agents.

User test

We tracked 3 months of usage data and collected survey voices from 101 users.

“AI podcasts are amazing — AI is evolving so fast!”

“It can generate a full PPT directly!”

“Bixby finally understands what I want.”

Some verbatim feedback

90.1%

Users preferred the new version over the previous design.

62.4%

Users used AI agents at least

twice per week.

52.6%

Users felt the new features were practical and fun to use.

User data

Intelligent recommendations

Design target

Key insight of user research

Our design target

新增AI实用工具体系

针对具体场景提供个性化的AI实用工具, 针对性地满足用户实际需求

Improve efficiency to learn new AI tools.

Want to quickly learn and adopt new AI tools to improve productivity.

Key insight of user research

Our design target

Users’ feedback

One month after launch, we collected user feedback and found that the 2 biggest efficiency pain points were:

Key insights

Entry point too deep

Only one entry point currently exists, requiring users to open the Bixby home first. In real usage, users rarely do this and expect to access the function directly from the current interface.

Users mentioned more use scenarios, such as problem solving and travel planning.

Many users are not aware that these AI tools are available.

Feature types need further expansion

New feature awareness

Mentions

Details

1

2

AI tools were difficult to discover, as the entry points were too deep in the product.

Users were unaware of new features and did not know how to use them.

Understand the problem

Current flow to use an AI tool

Step 1: Active Bixby and swipe up the chatbox to enter it’s main page

Step 2: Find a proper AI tool manually

low discoverability: To access AI agents, users had to enter the Bixby main page first, which interrupted their current task.

high decision cost: Users had to manually browse and evaluate multiple agents to find the right one, making the process time-consuming and inefficient.

Solution

1

2

Detect on-screen content and automatically recommend relevant AI agents.

Users can configure and execute AI tools directly on the current screen, without switching pages or interrupting their task.

The current flow required jumping between pages, which created a fragmented experience. To address this, I leveraged Screen Understanding technology to design a no-navigation flow.

We also introduced a Discovery Center that showcases AI capabilities and usage scenarios, allowing users to learn how each AI agent works via real examples.

User test

After launch, the usage of AI agents were significantly increased, and remained stable in the following months.

1D

1M

3M

1Y

32.7%

Real-Time Call (RTC) mode

Design target

Key insight of user research

Our design target

新增AI实用工具体系

针对具体场景提供个性化的AI实用工具, 针对性地满足用户实际需求

Introduce a better interaction, enabling users to have fluent and continuous conversations with AI.

Prefer hands-free conversations with AI, without constantly tapping the screen.

Old interaction

chatting with Bixby

Understand the problem

The current Bixby is limited to simple Q&A exchanges within the chat interface.

Users must repeatedly type inputs and read responses, which makes longer interactions inefficient and tiring.

Understand users’ needs

Cooking guidance

Mock practice

Research assistance

Emotional support

Travel planning

Daily Q&A

File processing

Creative inspiration

In user analysis, I realized multi-turn AI conversations are needed for personal and complex tasks. Then I identified several key scenarios most users needed:

43%

users require multi-turn conversations to complete personalized and complex tasks.

64%

users felt the current interaction was inefficient during multi-turn conversations.

The key scenarios users need AI

The core issue was that the chat-based interaction was not suitable for longer, task-oriented conversations with AI.

Solution

We introduced Real-Time Call simulating a phone call with AI. The interaction shifts from manual typing to continuous voice conversation.

1

3

2

Scenario-based conversation modes

Covers most of the key scenarios, providing more targeted assistance.

File-based conversations

Users can upload files and discuss the content with AI during the call.

Camera-based understanding

Bixby can analyze images captured by the camera and provide contextual responses.

On-screen context awareness

Bixby can observe the current screen and talk about the content in real time.

4

Validation

62%

Users used the RTC feature at least twice a week

83%

Users believed the feature was useful

>40

Users felt confused or anxious about status

Within one month of launch, we evaluated the RTC feature through usage data, surveys, and feedback from the Samsung community.

The results showed strong interest in the RTC mode, but also revealed an important issue: While we focused heavily on adding features, we overlooked the core conversational experience.

Understand the problem

Call AI

Standby

Listening→Speaking

Error

Abrupt connection transition

When the call connected, the interface changed directly from a blank screen to the AI orb, making the transition disconnected.

Unclear listening and speaking states

The animation during the listening–speaking cycle lacked clear transitions.

Overly negative error feedback

The full black error screen felt too harsh and created discomfort.

Users expect a clear visual transition when the AI call connects.

Users expect visual feedback showing the AI is receiving their speech, with animations that respond to speaking pace.

Users expect quick responses, or a clear loading state when processing takes longer.

Users expect a state distinct from listening, with softer animations to stay focused.

Users expect clear feedback about what went wrong, visually distinct from normal conversation.

Calling AI

Listening

Standyby

Thinking

Speaking

Error

User survey

To improve the conversational experience, we interviewed 18 users to understand their expectations at each stage of the AI conversation.

Competitor research

I conducted a competitive analysis of leading AI products’ RTC experiences, evaluating each interaction stage to define our product’s design principles.

Calling AI

Listening

Thinking

Speaking

Error

Presence

Focus

Perceived Intelligence

Focus

Clarity

Response Speed

Input Visualization

Anxiety Relief

Flexibility

Warmth

Provide a clear connection state so users know the AI is ready.

Show clear feedback that the AI is receiving the user’s speech.

Keep the interaction simple while clearly indicating the AI is processing.

Visually distinguish speaking from thinking and allow users to interrupt anytime.

Provide clear error feedback without overly negative visuals.

Solution

Based on research insights, I defined design principles for each interaction stage.

Calling AI

Listening

Thinking

Speaking

Error

User test

We tracked 3 months of usage data and collected survey feedback from 101 users.

Reflection

The best solution isn’t the “best experience”, but the most suitable one.

The key is to understand the product’s strengths and constraints and find the solution that fits best.

Users often can’t clearly tell their needs.

Instead of relying only on feedback directly, we need to identify patterns through user behavior and research.

Product design goes beyond design itself.

A successful product balances business goals, user experience, and technical feasibility.

Leveraged system-level capabilities to create a more seamless interaction experience.

Analysis through user interviews, surveys, data analysis, and feasibility testing.

Balanced business collaboration, user experience, and technical implementation.

What I did in this project

Want to talk deeper?

Contact me anytime if you're interested in one of my works

Want to talk deeper?

Contact me anytime if you're interested in one of my works

Create a free website with Framer, the website builder loved by startups, designers and agencies.