Volkswagen – Voice-First In-Car Productivity for Commuters

How we explored conversational UX to transform dead commute time into productive work sessions through hands-free, eyes-on-road voice interaction design.

Role and team


Sprint Master (Microsoft), Product Owner (Volkswagen), UX Designer (Me), Developer (Microsoft), Research & Test Users (Volkswagen & Microsoft)

Methods


Context of use

User Survey

Personas

User Scenarios

Task Analyze

Wireflow

Conversational Design

Prototyping

Timeline


10 days

Tool Kit


Google Forms

OAxure

Sketch

Pen & Paper

GoPro Camera


The Challenge: Unlocking Productive Commute Time

This design sprint brought together Volkswagen and Microsoft to explore a provocative question: How can we enable car commuters to be productive while driving, without compromising safety?

We applied Product Thinking methodology to solve this specific problem for a defined user group – car commuters who lose productive hours daily to driving, unlike train passengers who can work during transit.

The Opportunity: With voice assistants like Alexa and Cortana gaining adoption, and autonomous driving on the horizon, there was a window to prototype interim solutions that could gather data, prove concepts, and prepare for a voice-first future in vehicles.


The Car Commuter Problem

Drivers are legally and safety-wise obligated to keep their eyes on the road and hands on the wheel. Visual and cognitive attention must remain on traffic. This creates dead time for knowledge workers who could otherwise handle emails, plan meetings, or capture ideas.

Existing in-car systems were built for entertainment and navigation – not productivity. No solution existed for hands-free, eyes-free task management and communication.


Train Commuters Have It Better (and Worse)

Train passengers can freely use screens, keyboards, video calls – achieving full productivity. However, they lack privacy: entire compartments can overhear phone calls and meetings, creating social discomfort.

In contrast, cars offer privacy and isolation. Voice interfaces in cars could enable productivity without the privacy downside of trains – IF we could design the interaction correctly.

Design Insight: Voice interfaces are MORE appropriate in cars than trains due to privacy. The challenge is designing voice-first interactions that are safe, efficient, and feel natural while driving – not simply porting screen-based UX to voice.

Design Process

We structured the 10-day sprint around the Product Thinking methodology: understanding the problem deeply before jumping to solutions, validating vision and strategy, then rapidly building and testing.

1 Discover

User Problem: Facilitated discussions about what the problem actually is, how it feels to users, problem magnitude, and whether solving it creates real value. Conducted context of use analysis and user surveys to understand commute patterns.

Target Audience: Identified segments within car commuters: daily long-distance commuters, occasional drivers, business travelers, parents on school runs – each with different productivity needs.

2 Define

Vision: Positioned this as an interim solution toward autonomous driving, generating valuable voice UX data. Focused on leveraging existing technology (smartphones + Bluetooth) rather than requiring new hardware.

Strategy: Decided to build cross-platform (not VW-exclusive) to maximize data collection. Partnered with Microsoft Cognitive Services for voice recognition. Created feature map structured in epics, features, and functions.

3 Build and Test

Sketched UX for four user scenarios based on main problem statements. Conducted task analysis of ideated flows. Created wireframes and dialog scripts. Built GUI in Axure and simulated VUI with pre-recorded Cortana phrases. Tested in real driving conditions with GoPro documentation.

Strategic Vision: The Road to Autonomy

🎯 Goal: Voice UX Research

Primary goal wasn’t immediate profit but knowledge acquisition: How do people interact with voice interfaces while driving? What productivity tasks work hands-free? What safety concerns emerge? This data informs next-gen product development.

🚗 Vision: Why This Matters

The ultimate solution is autonomous driving – full productivity with zero driving responsibility. But that’s years away. This app serves as a bridge product, validating voice-first productivity concepts while collecting real-world usage data to inform future in-car systems.

🤝 Strategy: Microsoft Partnership

Built on Microsoft Cognitive Services (Cortana), creating strategic partnership potential. Rather than limiting to VW customers, we designed a smartphone app compatible with all vehicles via Bluetooth – maximizing data collection and market validation.


Context of Use

Understanding the possible contexts of use for users while commuting by car was foundational to our design decisions.

Strategic UX Decision: Users have full control over the listening mode. Unlike always-listening devices (Alexa, Google Home), our app requires manual activation via Bluetooth buttons. This addressed privacy concerns and gave drivers explicit control – critical for automotive trust.


Build & Test: Simulating Voice UX in Real Driving

We created four user scenarios aligned with our personas, then built prototypes combining graphical interfaces (Axure) with simulated voice interaction.

Test Setup

  • Real driving conditions (not simulated)
  • GoPro camera recorded full interaction
  • GUI tested via clickable Axure prototype on tablet
  • VUI simulated with pre-recorded Cortana audio
  • Bluetooth button for activation control
  • Post-drive interviews with participants

Test Scenarios

  • Scenario 1: Capture quick idea during morning commute
  • Scenario 2: Check calendar and schedule meeting
  • Scenario 3: Listen to email and voice-reply
  • Scenario 4: Add tasks from meeting notes

Wireflow

Voice Interaction Flow

Keyscreen Design


Key Findings: What We Learned from In-Car Testing

1

Seamless Cross-Modal Experience Desired

Users appreciated features but disliked the jarring transition from voice (while driving) to GUI (after parking). Starting a task via voice but needing to finish it on-screen created an unsatisfying „usage break“ that felt incomplete.

Implication: Voice-first design must consider the entire task lifecycle, not just the capture moment. Either complete tasks fully via voice, or design intentional hand-offs that feel natural, not forced by technical limitations.

2

Flexible Work/Personal Separation

Users wanted clear separation between work and personal content, but our solution – logging into separate accounts – proved too rigid. Real life doesn’t fit neat categories: people have personal ideas during work commutes and vice versa.

Implication: Better approach would be neutral capture (no upfront categorization), then tag content by recipient/context later. This matches actual mental models: „idea first, categorize later.“

3

Trial-and-Error Voice Exploration

Remember, this was 2016 – voice interfaces were nascent. Nearly every user entered a trial-and-error mode, testing boundaries and probing what commands the system understood. Users actively explored cognitive capabilities.

Implication: An initial voice menu providing guidance would have slowed task completion but reduced frustration and given users a better mental model of possibilities. Discoverability is critical in voice UX.


Personal Learnings

This was one of my first design sprints in 2016. With no iterations planned, the main outcome was the evidence that emerged – and valuable lessons about sprint methodology and voice-first design.

Most importantly: This sprint taught me that innovation often means building bridges to the future, not final products. Sometimes the value is in the learning, the data, and the validated direction – not the shipped feature.

UX design is fundamentally a cognitive group process, not an individual creative act

Evidence doesn’t always come from effort or time invested in individual tasks, by individuals. In an ideal world, UX design is a dynamic cognitive process that impacts a diverse group of people while insights affect and emerge from collaboration, not just solo work.

Voice UX is Fundamentally Different

You can’t just convert screen workflows to voice commands. Voice requires rethinking information architecture, task flows, error handling, and user mental models from scratch. The modality shift demands new design patterns.

In 2016, we were pioneers figuring this out. Today’s voice-first design principles emerged from sprints like this.

Sprint Constraints Force Creativity

The 10-day limit and lack of planned iterations meant we had one shot to get insights. This constraint forced radical prioritization, scrappiness (pre-recorded audio!), and focus on high-value learning over polish.

Evidence doesn’t require perfection – it requires smart questions and good enough prototypes to get answers.

Voice UX is Fundamentally Different

The 10-day limit and lack of planned iterations meant we had one shot to get insights. This constraint forced radical prioritization, scrappiness (pre-recorded audio!), and focus on high-value learning over polish.

Evidence doesn’t require perfection – it requires smart questions and good enough prototypes to get answers.

Cross-Company Collaboration is Complex

Aligning Volkswagen and Microsoft stakeholders, managing different organizational cultures, and balancing automotive safety concerns with tech innovation ambitions created unique facilitation challenges beyond pure UX design.

The partnership structure itself became part of the design constraint – decisions needed to serve both companies‘ strategic interests while prioritizing user needs.

If I Could Do It Again: I would push for at least one iteration cycle in the sprint structure. The findings from testing were so rich that immediate iteration would have validated solutions to the problems we discovered – making the sprint output actionable rather than just informative.
Privacy Settings
We use cookies to enhance your experience while using our website. If you are using our Services via a browser you can restrict, block or remove cookies through your web browser settings. We also use content and scripts from third parties that may use tracking technologies. You can selectively provide your consent below to allow such third party embeds. For complete information about the cookies we use, data we collect and how we process them, please check our Privacy Policy
Youtube
Consent to display content from - Youtube
Vimeo
Consent to display content from - Vimeo
Google Maps
Consent to display content from - Google