Apple (Personal Project)
Overview
Globally, 1.3 billion people live with vision impairments (WHO), and 60% of mobile screen reader users rely on Apple’s VoiceOver (WebAIM). Yet, for the estimated 6.6 million U.S. and 78 million global visually impaired iPhone users, error-prone touch gestures remain a barrier. By transforming the Camera Control Button into a tactile navigation tool, this project empowers accessibility while advancing Apple’s inclusive leadership. Like curb cuts, this intervention benefits all users through intuitive, one-handed ease.
Learn more about VoiceOver →
Globally, 1.3 billion people live with vision impairments (WHO), and 60% of mobile screen reader users rely on Apple’s VoiceOver (WebAIM). Yet, for the estimated 6.6 million U.S. and 78 million global visually impaired iPhone users, error-prone touch gestures remain a barrier. By transforming the Camera Control Button into a tactile navigation tool, this project empowers accessibility while advancing Apple’s inclusive leadership. Like curb cuts, this intervention benefits all users through intuitive, one-handed ease.
Learn more about VoiceOver →
Control Center Navigation
Hypothetical Metrics
While untested, the project’s theoretical impact could include:
40% reduction in navigation errors compared to touch gestures.
25% faster onboarding for new VoiceOver users.
90% user satisfaction in simulated usability studies (based on gesture fatigue reduction).
Role
Product Designer (Conceptual Design, Accessibility Focus)
Duration
6 weeks
Concept Ideation → Interaction Design → Theoretical Validation
Problem Statement
Traditional VoiceOver navigation relies on swipe gestures and double-taps, which can be challenging for users with low vision or motor impairments due to:
Accidental selections from mistimed double-taps.
Fatigue from repetitive swiping.
Difficulty locating on-screen elements without tactile feedback.
✦ Opportunity
Could a hardware button simplify navigation for users who struggle with touch gestures?


Camera Control Button?
While exploring Apple’s hardware features, I noticed the Camera Control Button on the iPhone 16 Pro had untapped potential for accessibility. Traditional VoiceOver navigation relies heavily on touch gestures (swipes, double-taps), which can be challenging for users with low vision or motor impairments
Underutilized Hardware: The button is currently limited to camera functions.
Tactile Advantage: Buttons provide physical feedback, which VoiceOver users often lack.
Goals
Design a speculative feature where the Camera Control Button:
Scrolls through VoiceOver elements (forward/backward).
Allows selection via tactile input.
Reduces reliance on screen gestures.
Key Features:
Key Features:
Button Mapping
Scroll down → Move forward.
Scroll up → Move backward.
Press → Select.
Feedback System
Haptic vibrations for each navigation step.
VoiceOver audio cues (e.g., “Item selected”).
Customization:
Adjust scroll speed (slow/fast) in Accessibility Settings.
Swap button direction (forward/backward) for left-handed users.
Button Mapping
Press → Select.
Scroll down → Move forward
Scroll up → Move backward
Feedback System
Haptic vibrations for each navigation step.
VoiceOver audio cues (e.g., “Item selected”).
Customization
Adjust scroll speed (slow/fast) in
Accessibility Settings.
Swap button direction (forward/backward) for left-handed users.
Key Features:
Button Mapping
Scroll down → Move forward.
Scroll up → Move backward.
Press → Select.
Feedback System
Haptic vibrations for each navigation step.
VoiceOver audio cues (e.g., “Item selected”).
Customization
Adjust scroll speed (slow/fast) in Accessibility Settings.
Swap button direction (forward/backward) for left-handed users.
Key Features:
Key Features:
System-Level Integration
System-Level Integration
New Accessibility Settings
Toggle for enabling/disabling button navigation.
Calibration tools for button sensitivity and scroll speed.
Hardware-Software Syncing
Override the Camera Control Button’s default camera functions when in navigation mode.
Ensure compatibility with VoiceOver’s focus highlighting and rotor commands.
User Education & Onboarding
First-Time Setup: A pop-up tutorial when users enable the feature for the first time.
Interactive Demo: A practice mode in Settings where users can test the button without affecting apps.
Accessibility Hub: Add a section in Apple’s Accessibility website showcasing use cases and testimonials.
New Accessibility Settings
Toggle for enabling/disabling button navigation.
Calibration tools for button sensitivity and scroll speed.
Hardware-Software Syncing
Override the Camera Control Button’s default camera functions when in navigation mode.
Ensure compatibility with VoiceOver’s focus highlighting and rotor commands.
User Education & Onboarding
First-Time Setup:
A pop-up tutorial when users enable the feature for the first time.
Interactive Demo:
A practice mode in Settings where users can test the button without affecting apps.
Accessibility Hub:
Add a section in Apple’s Accessibility website showcasing use cases and testimonials.
System-Level Integration
New Accessibility Settings
Toggle for enabling/disabling button navigation.
Calibration tools for button sensitivity and scroll speed.
Hardware-Software Syncing
Override the Camera Control Button’s default camera functions when in navigation mode.
Ensure compatibility with VoiceOver’s focus highlighting and rotor commands.
User Education & Onboarding
First-Time Setup: A pop-up tutorial when users enable the feature for the first time.
Interactive Demo: A practice mode in Settings where users can test the button without affecting apps.
Accessibility Hub: Add a section in Apple’s Accessibility website showcasing use cases and testimonials.
Key Features:
Key Features:
Prototyping
Prototyping
Designed a micro-interaction video to demonstrate the feature’s potential.
Designed a micro-interaction video to demonstrate the feature’s potential.



Design Philosophy
Inclusive by Default: Leverage existing hardware to solve accessibility challenges.
Tactile Empowerment: Offer a physical alternative to touch-dependent interactions.
Hypothetical User Scenarios
A user with tremors struggles with precise swipes but can reliably press a button.
A user with low vision wants to navigate apps faster without memorizing gestures.
Visualization
Created Figma wireframes showing how the button could integrate with VoiceOver:
Highlighted focus states for VoiceOver elements.
Added a toggle in Settings to enable “Button Navigation Mode.”
Brainstorming
Mapped VoiceOver’s gesture-based workflow to a button-driven interaction:
Short Scroll: Move to next item (mimics swipe right).
Long Scroll: Move to previous item (mimics swipe left).
Double Press: Select item (replaces double-tap).
Explored haptic feedback to confirm actions (e.g., subtle vibrations for navigation steps).
Mapped VoiceOver’s gesture-based workflow to a button-driven interaction:
Short Press: Move to next item (mimics swipe right).
Long Press: Move to previous item (mimics swipe left).
Double Press: Select item (replaces double-tap).
Explored haptic feedback to confirm actions (e.g., subtle vibrations for navigation steps).
Try The Figma Prototype
Theoretical Validation
Theoretical Validation
Theoretical Validation
Referenced WCAG guidelines (e.g., Operable, Understandable) to ensure compliance.
Referenced WCAG guidelines (e.g., Operable, Understandable) to ensure compliance.
Accessibility
Referenced WCAG guidelines (e.g., Operable, Understandable) to ensure compliance.
Compared the concept to Apple’s existing accessibility features (e.g., Back Tap, AssistiveTouch).
Compared the concept to Apple’s existing accessibility features (e.g., Back Tap, AssistiveTouch).
Brand Coherence
Compared the concept to Apple’s existing accessibility features (e.g., Back Tap, AssistiveTouch).
Reduced Cognitive Load: Fewer gestures to memorize.
Reduced Cognitive Load: Fewer gestures to memorize.



Reduced Cognitive Load
Fewer gestures to memorize.
Improved Accuracy:
Physical buttons minimize accidental inputs.
Improved Accuracy:
Physical buttons minimize accidental inputs.



Improved Accuracy
Physical buttons minimize accidental inputs.


Design Rationale
Why This Interaction Model?
Familiarity: Mimics existing VoiceOver gestures (forward/backward).
Scalability: Could extend to other hardware buttons (e.g., volume keys).
Why the Camera Control Button?
Underutilized Hardware: The button is currently limited to camera functions.
Tactile Advantage: Buttons provide physical feedback, which VoiceOver users often lack.
Theoretical Outcomes
Faster Navigation
Users could cycle through menus 20–30% faster (based on gesture vs. button input benchmarks).
Increased Independence
Empowers users who find touch gestures unreliable.
Apple Ecosystem Synergy
Strengthens iPhone’s accessibility suite alongside VoiceOver and Switch Control.


Design Rationale
Underutilized Hardware: The button is currently limited to camera functions.
Tactile Advantage: Buttons provide physical feedback, which VoiceOver users often lack.
Why the Camera Control Button?
Why This Interaction Model?
Familiarity: Mimics existing VoiceOver gestures (forward/backward) and mouse scroll.
Scalability: Could extend to other navigation based interactions.
Theoretical Outcomes
Faster Navigation
Users could cycle through menus 20–30% faster (based on gesture vs. button input benchmarks).
Increased Independence
Empowers users who find touch gestures unreliable.
Apple Ecosystem Synergy
Strengthens iPhone’s accessibility suite alongside VoiceOver and Switch Control.
Other Projects
Overview
Globally, 1.3 billion people live with vision impairments (WHO), and 60% of mobile screen reader users rely on Apple’s VoiceOver (WebAIM). Yet, for the estimated 6.6 million U.S. and 78 million global visually impaired iPhone users, error-prone touch gestures remain a barrier. By transforming the Camera Control Button into a tactile navigation tool, this project empowers accessibility while advancing Apple’s inclusive leadership. Like curb cuts, this intervention benefits all users through intuitive, one-handed ease.
Learn more about VoiceOver →
Role
Product Designer (Conceptual Design, Accessibility Focus)
Duration
6 weeks
Concept Ideation → Interaction Design → Theoretical Validation
Problem Statement
Traditional VoiceOver navigation relies on swipe gestures and double-taps, which can be challenging for users with low vision or motor impairments due to:
Accidental selections from mistimed
double-taps.Fatigue from repetitive swiping.
Difficulty locating on-screen elements without tactile feedback.
✦ Opportunity
Could a hardware button simplify navigation for users who struggle with touch gestures?


Why The Camera Control Button?
While exploring Apple’s hardware features, I noticed the Camera Control Button on the iPhone 15 Pro had untapped potential for accessibility. Traditional VoiceOver navigation relies heavily on touch gestures, which can be challenging for users with low vision or motor impairments
Goals
Design a speculative feature where the Camera Control Button:
Scrolls through VoiceOver elements (forward/backward).
Allows selection via tactile input.
Reduces reliance on screen gestures.