Scripted

Turning real conversations into structured language learning

UX/UI Designer · Team of 4 · 10 weeks · 2025

Scripted — home screen, recording interface, conversation summary, and lesson screen
Home, recording, transcript analysis, and lesson screens.

Overview

Scripted is a mobile app designed to support intermediate language learners in developing conversational fluency through real-time conversations and personalised feedback.

My role

  • Led design of the recording experience and interaction flow (solo, group and AI conversations)
  • Designed key interfaces including home, recording, summary-first feedback, and AI conversation screens
  • Redefined the achievement system based on usability testing (independent extension)
  • Contributed to navigation structure, user flows, and iterative decision-making within a collaborative team

The Problem

Intermediate language learners often plateau: they understand grammar and vocabulary but struggle to apply it in real conversations.

Existing tools fall into two extremes:

  • Lesson-based apps provide structure but lack real conversational practice
  • Community platforms enable conversation but offer little meaningful feedback

As a result, learners lack both consistent practice opportunities and actionable insight into their spoken performance.

Product Strategy

Scripted is an app designed as an accessory to conversational practice for intermediate language learners.

The strategy focused on:

  • Supporting real conversations between learners (friends, family, partners)
  • Making practice more accessible through optional AI, solo modes and HelloTalk integration
  • Providing real-time conversation assistance without disrupting flow
  • Creating a system that encourages continued engagement over time

Concept Evolution

We explored four distinct directions to propose a solution to our problem domain, each targeting a different method of language learning.

Recording

Recording app concept screen

Media

Media immersion concept screen

Vocabulary

Vocabulary trainer concept screen

Community

Community platform concept screen
  • Recording app: Capturing real conversations and generating transcripts
  • Media immersion: Learning through curated content such as films, music, games and short-form media
  • Vocabulary trainer: Focusing on saving and revisiting difficult words for targeted practice
  • Community platform: Enabling social interaction through messaging and daily shared content

Key Insight

Initial testing revealed that while all concepts addressed aspects of learning, none fully supported real conversational fluency.

The recording concept was the strongest, but limited by reliance on conversation partners.

Final Direction

We evolved the solution by combining:

  • Real conversations recorded to produce a transcript (later extended to include AI and solo practice modes)
  • HelloTalk integration rather then recreating community features
  • Short form lessons based on the transcript insights

This expanded accessibility while maintaining focus on real conversational interaction.

Core Experience

Recording a Conversation

Scripted is built around real-world conversation practice. The core experience was designed to support users during live interaction without interrupting conversational flow.

The challenge was balancing:

  • Providing timely support
  • Without increasing cognitive load or screen dependency

Flow

Start

Users either:

  • Record a real conversation
  • Or practice with an AI chatbot

During recording, the system automatically detects speakers (solo, pair, or group) and assigns visual indicators in the transcript.

Colour differentiation between speakers was used as a primary visual cue; future iterations would further improve accessibility by strengthening contrast and adding non-colour indicators.

Record

The interface is intentionally minimal:

  • Waveform
  • Timer
  • Primary controls

Users are not required to interact with the screen, allowing full focus on conversation.

Contextual Support (Triggered)

Support appears only when explicitly prompted:

“How do you say…?”

This triggers a lightweight suggestion overlay that:

  • Provides vocabulary support
  • Does not interrupt the flow of conversation
  • Avoids persistent UI clutter

This interaction was refined through testing to reduce intrusiveness.

Summary-first Feedback

After recording, users land on a structured summary:

  • Conversation context
  • Key metadata (duration, participants)
  • Quick indicators (e.g. weaknesses, vocabulary)

Users can then choose to explore detailed transcript analysis.

During testing, users expected the weaknesses and glossary words to be clickable. This revealed an opportunity to evolve the feature into an interactive summary in future iterations.

Home recording screen

Home

Recording conversation screen

Recording

'How do you say...?' support screen

How do you say...?

Summary screen

Summary

How do you say before iteration

Before

How do you say after iteration

After

Earlier versions displayed the prompt as a central popup, this was repositioned to a less intrusive area following usability feedback.

Key Design Decisions

  • Designed for hands-free interaction during live conversation
  • Made support user-triggered, not system-imposed
  • Shifted from intrusive overlays to subtle contextual prompts
  • Prioritized summary-first feedback over full transcript exposure

Outcome

The recording flow achieved 100% task completion in usability testing, indicating strong alignment with user expectations and mental models.

More importantly, it enabled users to maintain conversational flow while still receiving structured feedback.

Usability Testing & Iteration

We conducted usability testing on the high-fidelity prototype to evaluate how well the experience supported real conversational practice and post-conversation feedback.

Participants: 8 intermediate language learners

Method: Think-aloud protocol, task-based testing

Tasks: Record a conversation, review feedback, complete a lesson

Key Findings

  1. Feedback felt overwhelming

    Users found transcript review dense and difficult to engage with.

    “I would just scroll down, look at a couple of the mistakes, and exit out.”

  2. Progress and achievement are essential

    Users were motivated by visible progress, but the experience lacked positive reinforcement.

    “I like the weakness levels and being able to see [the bar] go from red to yellow to green as I practice.”

  3. Onboarding is needed

    Some icons and features were unclear without guidance.

    “In the onboarding you could explain what the different features and icons are.”

Home screen

Home screen recording tab

Group conversation

Group conversation recording screen

Conversation summary

Group conversation summary screen

Transcript analysis

Transcript analysis screen

Review tab

Review tab weaknesses screen

Weakness Details

Weakness details screen

Complete a lesson

Complete a lesson screen

Testing Conclusion

Key considerations

  • High use of LLMs: Requires strong guardrails to ensure accuracy and trust
  • Privacy & transparency: Critical due to recorded conversation data
  • Engagement: The experience needs stronger reward and motivation systems

What worked well

  • Strong concept: Addressing a clear gap in intermediate language learning
  • Personalised value: Individual lessons based on real conversations
  • Balanced, seamless user flows: Easy to navigate and supports conversation without disruption

Opportunity

The product is functional and intuitive, but lacks a sense of reward and progression.

This insight led me to explore how the product could better support motivation and long-term engagement.

Enhancing Engagement Through Achievement Design

To address the lack of motivation identified in testing, I redesigned the feedback system to focus on progress, reinforcement, and habit formation.

Solution — Achievement System

I redesigned the feedback layer to introduce a system of:

  • Achievement cards
  • Milestones
  • Goal-based progress tracking

Key Features

1. Achievement Cards

Highlight small wins after each conversation

Example: “You spoke for 20 minutes that's your longest conversation yet”

Reinforces progress rather than mistakes

2. Goal Setting

Users define goals to achieve during conversations

Progress is visible and trackable

Encourages intentional practice

3. Milestones

Tracks consistency and long-term progress

Example: “5 conversations completed this week”

Builds habit formation

4. Progress Framing

Shifts feedback tone from corrective → encouraging

Makes improvement visible on the interface

Design Rationale

This approach applies principles of behavioural design:

  • Positive reinforcement increases repeat behaviour
  • Visible progress supports motivation
  • Small wins reduce perceived effort over time

The system reframes language learning as continuous progress, not error correction.

Outcome (Conceptual)

While not implemented in the final prototype, this extension demonstrates how the product could evolve to support long-term engagement.

It directly responds to usability findings and introduces a scalable motivational layer across the product.

10 conversations milestone screen

10 conversations

Goal setting screen

Goal setting

Home screen with goal overview

Home screen goal

Conversation with goal tracking screen

Conversation with goal

The achievement system was designed as a continuous loop across key moments in the product:

  • After conversation → reward
  • Before next session → goal setting
  • During use → reinforcement
  • Over time → progress tracking

Reflection

  • Accessibility constraints shape product direction — Solo and AI modes emerged from the need to support users without consistent partners.
  • Usability does not equal engagement — The achievement redesign addressed motivation, not just clarity.
  • Summary-first feedback reduces overwhelm — Prioritizing the summary over the full transcript improved clarity and scannability.
  • Low-interruption interaction design — Designing the recording flow required minimizing on-screen interaction, enabling users to stay focused on conversation without disruption.