Abstract
Abstract
Despite a global increase in mobile device ownership, older adults continue to use few apps and fewer features. We present results from a qualitative study examining how 17 older adults (60+) use Google Maps — 172 interaction problems were identified. Non-motor issues (discoverability, affordances, information scent) caused more frustration and resignation than motor issues, even though motor issues were numerically more frequent.
Keywords: older adults, mobile maps, accessibility, non-motor issues, information scent, visual saliency, think-aloud, qualitative study
1. Motivation
Mobile Maps as a Window into Older Adults' Mobile Accessibility
Using Google Maps is the most universal smartphone activity across all age groups — yet we know little about how older adults experience it. Unlike simplified lab tasks, mobile maps involve multi-step lookup and exploratory search, voice input, sharing, and frequently-changing UI layouts.
We chose not to study a simplified custom app. We wanted to examine how older adults tackle the complexities of a mainstream app, on their own devices, in their own settings.
2. Method
Ecologically Valid Think-Aloud Study
Pilot tests with surveys and diary studies failed — responses were shallow and diary reports showed family assistance. We shifted to a concurrent think-aloud protocol in participants' own settings, on their own devices, with no time limits.
Task 1
Lookup — Public Transit
Find how to go from Chicago Midway Airport to 540 West Madison using public transportation.
Task 2
Lookup — Constrained (GPS off)
Find how to walk from the Drake Hotel to the Museum of Contemporary Art when GPS is unavailable.
Task 3
Exploratory Search + Share
Find a nearby ice cream shop, choose a transport mode, then share the location with the interviewer.
3. Results
172 Interaction Problems Across Three Non-Motor Categories
105 motor issues and 67 non-motor issues were identified. Older adults encountered more motor issues, but spent more time troubleshooting non-motor issues, got more frustrated, and abandoned tasks 19 times — exclusively due to non-motor causes.
Inadequate Visual Saliency — "I don't see the buses"
UI components frequently went unnoticed despite participants already customizing fonts and contrast. Transit mode selectors, category navbars, and action buttons were missed entirely — not due to a lack of understanding of what they meant, but simply because they were not salient enough to attract attention.
Ambiguous Affordances — "How do I put an address in this?"
In other instances participants noticed the correct UI component but could not use it. Card sliders were tapped rather than swiped, address fields were not recognized, and voice prompts were opened as a help mechanism rather than for input.
Low Information Scent — The Loop Problem
When problem-solving, older adults frequently went back-and-forth the same 3–4 pages and got stuck in a loop. In each recorded instance, the correct feature was always on one of the pages they kept revisiting — but they could not infer the next interaction step.
Core Finding
Older adults can articulate what they are looking for — by feature name or description. The failure is not conceptual but perceptual and navigational: poor discoverability, unclear affordances, and low information scent trap users in repetitive loops.
4. Design Implications
Two Proposed Solutions
- Find on Page Offer a text-search widget to find any UI component by name or label on the current page — invokable via a hardware button or always visible. This directly addresses visual saliency and information scent.
- Intercept All Input Events Program UI components to handle commonly confused input events; on repeated misuse, the app notifies the user how to correctly interact with the component.
These findings directly motivated the design of a voice assistant — since older adults could articulate what they wanted, a voice-based query mapped naturally onto a "Find on Page" concept at the platform level.