2025
SpeakEasy: Voice-Driven AI for Inclusive XR
Michael Chaves
This thesis explores how voice-driven artificial intelligence can make extended reality (XR) more inclusive for users with limited mobility. Recognizing that most XR systems still depend on handheld controllers, SpeakEasy investigates how natural speech can serve as a primary mode of interaction—reducing physical barriers and fostering emotional ease. Through literature review, participatory co-design sessions, and iterative prototyping, the project developed a Ten-Pillar Framework for Accessible Voice Interaction in Spatial Computing, addressing factors such as feedback clarity, user autonomy, and ethical data control.


The final SpeakEasy prototype demonstrates hands-free, voice-based navigation in a
                  mixed-reality wellness environment, enabling users to complete tasks, receive adaptive
                  feedback, and engage in guided experiences using only speech. User testing showed
                  high satisfaction and strong usability, suggesting that natural language and real-time
                  AI personalization can make immersive technologies more intuitive and equitable. By
                  centering accessibility from the start, SpeakEasy offers designers a practical roadmap
                  toward XR experiences that welcome more bodies, voices, and perspectives.
