Dual-Task Multimodal Web Browsing in Semi-Autonomous Vehicles
Clark, Jarrett William
MetadataShow full item record
Drivers should remain alert when operating semi-autonomous vehicles as human intervention is required when the automated systems fail, or an unusual situation is encountered. However, drivers do not always focus solely on the road and sometimes divert their attention to look at cell phones, laptops, tablet computers, or in-dash infotainment systems. In this research, software was created to enable operators of a simulated self-driving car to browse the Human-Centered Design Institute (HCDI) website using multimodal interaction. The software used synthesized speech to read the website aloud and visually highlighted the content on the screen as it was read. Web analytics were also used to optimize the reading order of the website content. The audio modality of the software was tested in a single-task experiment. The final software prototype with audible and visual output was tested by drivers performing a dual-task in a driving simulator. Workload, usability, and emotion were measured after using the software. Usability increased and workload decreased with added visualization. Driver’s looked at the screen more often when highlighting was added to the screen, but the duration of the glances was shorter. The results suggest the software creates more task switching but allows drivers to shorten the duration of their screen glances when browsing the HCDI website in a simulated semi-autonomous vehicle.