16 results on '"Lewis L. Chuang"'
Search Results
2. Ambient Light Conveying Reliability Improves Drivers’ Takeover Performance without Increasing Mental Workload
- Author
-
Nikol Figalová, Lewis L. Chuang, Jürgen Pichen, Martin Baumann, and Olga Pollatos
- Subjects
ambient light ,reliability ,take-over request ,mental workload ,electroencephalography (EEG) ,transition of control ,Technology ,Science - Abstract
Drivers of L3 automated vehicles (AVs) are not required to continuously monitor the AV system. However, they must be prepared to take over when requested. Therefore, it is necessary to design an in-vehicle environment that allows drivers to adapt their levels of preparedness to the likelihood of control transition. This study evaluates ambient in-vehicle lighting that continuously communicates the current level of AV reliability, specifically on how it could influence drivers’ take-over performance and mental workload (MW). We conducted an experiment in a driving simulator with 42 participants who experienced 10 take-over requests (TORs). The experimental group experienced a four-stage ambient light display that communicated the current level of AV reliability, which was not provided to the control group. The experimental group demonstrated better take-over performance, based on lower vehicle jerks. Notably, perceived MW did not differ between the groups, and the EEG indices of MW (frontal theta power, parietal alpha power, Task–Load Index) did not differ between the groups. These findings suggest that communicating the current level of reliability using ambient light might help drivers be better prepared for TORs and perform better without increasing their MW.
- Published
- 2022
- Full Text
- View/download PDF
3. Visual Demands of Walking Are Reflected in Eye-Blink-Evoked EEG-Activity
- Author
-
Edmund Wascher, Stefan Arnau, Marie Gutberlet, Lewis L. Chuang, Gerhard Rinkenauer, and Julian Elias Reiser
- Subjects
gait control ,eye-blink-evoked potentials ,visual demands ,ERPs ,ERSPs ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Blinking is a natural user-induced response which paces visual information processing. This study investigates whether blinks are viable for segmenting continuous electroencephalography (EEG) activity, for inferring cognitive demands in ecologically valid work environments. We report the blink-related EEG measures of participants who performed auditory tasks either standing, walking on grass, or whilst completing an obstacle course. Blink-related EEG activity discriminated between different levels of cognitive demand during walking. Both behavioral parameters (e.g., blink duration or head motion) and blink-related EEG activity varied with walking conditions. Larger occipital N1 was observed during walking, relative to standing and traversing an obstacle course, which reflects differences in bottom-up visual perception. In contrast, the amplitudes of top-down components (N2, P3) significantly decreased with increasing walking demands, which reflected narrowing attention. This is consistent with blink-related EEG, specifically in Theta and Alpha power that, respectively, increased and decreased with increasing demands of the walking task. This work presents a novel and robust analytical approach to evaluate the cognitive demands experienced in natural work settings, which precludes the use of artificial task manipulations for data segmentation.
- Published
- 2022
- Full Text
- View/download PDF
4. VisME: Visual microsaccades explorer
- Author
-
Tanja Munz, Lewis L. Chuang, Sebastian Pannasch, and Daniel Weiskopf
- Subjects
microsaccades ,visual analytics ,eye movement ,eye tracking ,parameters ,fixations ,Human anatomy ,QM1-695 - Abstract
This work presents a visual analytics approach to explore microsaccade distributions in high-frequency eye tracking data. Research studies often apply filter algorithms and parameter values for microsaccade detection. Even when the same algorithms are employed, different parameter values might be adopted across different studies. In this paper, we present a visual analytics system (VisME) to promote reproducibility in the data analysis of microsaccades. It allows users to interactively vary the parametric values for microsaccade filters and evaluate the resulting influence on microsaccade behavior across individuals and on a group level. In particular, we exploit brushing-and-linking techniques that allow the microsaccadic properties of space, time, and movement direction to be extracted, visualized, and compared across multiple views. We demonstrate in a case study the use of our visual analytics system on data sets collected from natural scene viewing and show in a qualitative usability study the usefulness of this approach for eye tracking researchers. We believe that interactive tools such as VisME will promote greater transparency in eye movement research by providing researchers with the ability to easily understand complex eye tracking data sets; such tools can also serve as teaching systems. VisME is provided as open source software.
- Published
- 2019
- Full Text
- View/download PDF
5. Eye tracking and visualization. Introduction to the Special Thematic Issue
- Author
-
Michael Burch, Lewis L. Chuang, Andrew Duchowski, Weiskopf Daniel, and Rudolf Groner
- Subjects
Eye movement ,eye tracking ,visualization ,vision ,cognition ,perception ,Human anatomy ,QM1-695 - Abstract
There is a growing interest in eye tracking technologies applied to support traditional visualization techniques like diagrams, charts, maps, or plots, either static, animated, or interactive ones. More complex data analyses are required to derive knowledge and meaning from the data. Eye tracking systems serve that purpose in combination with biological and computer vision, cognition, perception, visualization, human-computer-interaction, as well as usability and user experience research. The 10 articles collected in this thematic special issue provide interesting examples how sophisticated methods of data analysis and representation enable researchers to discover and describe fundamental spatio-temporal regularities in the data. The human visual system, supported by appropriate visualization tools, enables the human operator to solve complex tasks, like understanding and interpreting three-dimensional medical images, controlling air traffic by radar displays, supporting instrument flight tasks, or interacting with virtual realities. The development and application of new visualization techniques is of major importance for future technological progress.
- Published
- 2018
- Full Text
- View/download PDF
6. Steering demands diminish the early-P3, late-P3 and RON components of the event-related potential of task-irrelevant environmental sounds
- Author
-
Menja eScheer, Heinrich H. Bülthoff, and Lewis L. Chuang
- Subjects
MMN ,distraction ,steering ,Mental Workload ,early P3 ,late P3 ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
The current study investigates the demands that steering places on mental resources. Instead of a conventional dual-task paradigm, participants of this study were only required to perform a steering task while task-irrelevant auditory distractor probes (environmental sounds and beep tones) were intermittently presented. The event-related potentials (ERPs), which were generated by these probes, were analyzed for their sensitivity to the steering task’s demands. The steering task required participants to counteract unpredictable roll disturbances and difficulty was manipulated either by adjusting the bandwidth of the roll disturbance or by varying the complexity of the control dynamics. A mass univariate analysis revealed that steering selectively diminishes the amplitudes of early P3, late P3, and the re-orientation negativity (RON) to task-irrelevant environmental sounds but not to beep tones. Our findings are in line with a three-stage distraction model, which interprets these ERPs to reflect the post-sensory detection of the task-irrelevant stimulus, engagement, and re-orientation back to the steering task. This interpretation is consistent with our manipulations for steering difficulty. More participants showed diminished amplitudes for these ERPs in the ‘hard’ steering condition relative to the ‘easy’ condition. To sum up, the current work identifies the spatiotemporal ERP components of task-irrelevant auditory probes that are sensitive to steering demands on mental resources. This provides a non-intrusive method for evaluating mental workload in novel steering environments.
- Published
- 2016
- Full Text
- View/download PDF
7. A comparison of geometric- and regression-based mobile gaze-tracking
- Author
-
Björn eBrowatzki, Heinrich H. Bülthoff, and Lewis L. Chuang
- Subjects
Eye Movements ,eye tracking ,active vision ,gaze measurement ,gaussian processes ,Calibration method ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Video-based gaze-tracking systems are typically restricted in terms of their effective tracking space. This constraint limits the use of eyetrackers in studying mobile human behavior. Here, we compare two possible approaches for estimating the gaze of participants who are free to walk in a large space whilst looking at different regions of a large display. Geometrically, we linearly combined eye-in-head rotations and head-in-world coordinates to derive a gaze vector and its intersection with a planar display, by relying on the use of a head-mounted eyetracker and body-motion tracker. Alternatively, we employed Gaussian process regression to estimate the gaze intersection directly from the input data itself. Our evaluation of both methods indicates that a regression approach can deliver comparable results to a geometric approach. The regression approach is favored, given that it has the potential for further optimization, provides confidence bounds for its gaze estimates and offers greater flexibility in its implementation. Open-source software for the methods reported here is also provided for user implementation.
- Published
- 2014
- Full Text
- View/download PDF
8. Adapting visualizations and interfaces to the user
- Author
-
Francesco Chiossi, Johannes Zagermann, Jakob Karolus, Nils Rodrigues, Priscilla Balestrucci, Daniel Weiskopf, Benedikt Ehinger, Tiare Feuchtner, Harald Reiterer, Lewis L. Chuang, Marc Ernst, Andreas Bulling, Sven Mayer, and Albrecht Schmidt
- Subjects
General Computer Science ,ddc:004 ,Physiological computing, adaptive systems, hybrid user interfaces, multimodal interaction - Abstract
Adaptive visualization and interfaces pervade our everyday tasks to improve interaction from the point of view of user performance and experience. This approach allows using several user inputs, whether physiological, behavioral, qualitative, or multimodal combinations, to enhance the interaction. Due to the multitude of approaches, we outline the current research trends of inputs used to adapt visualizations and user interfaces. Moreover, we discuss methodological approaches used in mixed reality, physiological computing, visual analytics, and proficiency-aware systems. With this work, we provide an overview of the current research in adaptive systems.
- Published
- 2022
- Full Text
- View/download PDF
9. Acoustic Cues Increase Situational Awareness in Accident Situations: A VR Car-Driving Study
- Author
-
Uijong Ju, Christian Wallraven, and Lewis L. Chuang
- Subjects
Accident (fallacy) ,Situation awareness ,Warning system ,Mechanical Engineering ,Automotive Engineering ,Applied psychology ,Driving simulation ,Virtual reality ,Car driving ,Psychology ,Physiological responses ,Computer Science Applications - Abstract
Our work for the first time evaluates the effectiveness of visual and acoustic warning systems in an accident situation using a realistic, immersive driving simulation. In a first experiment, 70 participants were trained to complete a course at high speed. The course contained several forks where a wrong turn would lead to the car falling off a cliff and crashing - these forks were indicated either with a visual warning sign for a first, no-sound group or with a visual and auditory warning cue for a second, sound group. In a testing phase, right after the warning signals were given, trees suddenly fell on the road, leaving the (fatal) turn open. Importantly, in the no-sound group, 18 out of 35 people still chose this turn, whereas in the sound group only 5 out of 35 people did so - the added sound therefore had a large and significant increase in situational awareness. We found no other differences between the groups concerning age, physiological responses, or driving experience. In a second replication experiment, the setup was repeated with another 70 participants without emphasis on driving speed. Results fully confirmed the previous findings with 17 out of 35 people in the no-sound group versus only 6 out of 35 in the sound group choosing the turn to the cliff. With these two experiments using a one-shot design to avoid pre-meditation and testing naive, rapid decision-making, we provide clear evidence for the advantage of visual-auditory in-vehicle warning systems for promoting situational awareness.
- Published
- 2022
- Full Text
- View/download PDF
10. Interpersonal Distance in the SARS-CoV-2 Crisis
- Author
-
Heiko Hecht, Lewis L. Chuang, Robin Welsch, and Christoph von Castell
- Subjects
discomfort ,2019-20 coronavirus outbreak ,Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) ,Pneumonia, Viral ,Behavioural sciences ,050109 social psychology ,Human Factors and Ergonomics ,Interpersonal communication ,Mandatory Programs ,050105 experimental psychology ,Behavioral Neuroscience ,Personal Space ,Personal space ,Pandemic ,Humans ,0501 psychology and cognitive sciences ,Social Change ,Social Behavior ,Pandemics ,Applied Psychology ,SARS-CoV-2 ,Social distance ,Communication ,05 social sciences ,Social change ,COVID-19 ,interpersonal distance ,Psychological Distance ,Visual Perception ,How The HF/E Knowledge Base Can Contribute to Addressing The Coronavirus Crisis ,proxemics ,Psychology ,Coronavirus Infections ,Social psychology ,Behavioral Sciences - Abstract
Background Mandatory rules for social distancing to curb the SARS-CoV-2 pandemic require individuals to maintain a critical interpersonal distance above 1.5 m. However, this contradicts our natural preference, which is closer to 1 m for non-intimate encounters, for example, when asking a stranger for directions. Objective This review addresses how humans typically regulate interpersonal distances, in order to highlight the challenges of enforcing atypically large interpersonal distances. Method To understand the challenges posed by social distancing requirements, we integrate relevant contributions from visual perception, social perception, and human factors. Results To date, research on preferred interpersonal distances suggests that social distancing could induce discomfort, heighten arousal, and decrease social signaling in the short term. While the protracted effects of social distancing are unclear, we propose hypotheses on the mid- to long-term consequences of violating preferred norms of interpersonal distances. Conclusion We suggest that enforcing a physical distance of 1.5–2 m presents a serious challenge to behavioral norms. Application We address how notifications, architectural design, and visualizations could be effectively applied to promote interpersonal distance requirements.
- Published
- 2020
11. A Survey of Viewpoint Selection Methods for Polygonal Models.
- Author
-
Xavier Bonaventura, Miquel Feixas, Mateu Sbert, Lewis L. Chuang, and Christian Wallraven
- Published
- 2018
- Full Text
- View/download PDF
12. Take-over requests during highly automated driving: How should they be presented and under what conditions?
- Author
-
Stefan Brandenburg and Lewis L. Chuang
- Subjects
Time headway ,050210 logistics & transportation ,Operations research ,Computer science ,05 social sciences ,Control (management) ,Driving simulator ,Human factors and ergonomics ,Poison control ,Transportation ,Take over ,Visualization ,Geometric design ,0502 economics and business ,Automotive Engineering ,0501 psychology and cognitive sciences ,050107 human factors ,Applied Psychology ,Civil and Structural Engineering - Abstract
Even with highly automated driving, there will exist the occasional need for a human driver to resume control. The design of take-over requests (TOR) and the driving scenario upon their presentation are factors that can be expected to determine the effective resumption of vehicle control. We conducted a driving simulator study in the laboratory that examined how either an abstract or skeuomorphic TOR design influenced driving resumption, given different conditions of road curvature and time headway. We found that drivers react faster and more accurately when experiencing a skeuomorphic visualisation of a TOR prompt compared to an abstract interface, especially when there is less time headway. Moreover, participants reported subjective preferences for the skeuomorphic TOR design and considered them to be more intuitively comprehensible. Furthermore, more accurate TOR responses were found when TORs of lateral control were presented before longitudinal control. Based on these results, we formulated five conclusions relating to the design of take-over requests. These conclusions should be validated in more realistic settings. However, in the future, they could be the basis for more research developing design guidelines that could create safe and reliable communications between drivers and vehicles.
- Published
- 2019
- Full Text
- View/download PDF
13. Object Perception, Attention, and Memory 2005 Conference Report
- Author
-
Lewis L. Chuang, Heinrich H. Bülthoff, Ian M. Thornton, and Quoc C. Vuong
- Subjects
Cognitive science ,Arts and Humanities (miscellaneous) ,Cognitive Neuroscience ,Experimental and Cognitive Psychology ,Psychology - Published
- 2006
- Full Text
- View/download PDF
14. Why are moving faces easier to recognize?
- Author
-
Lewis L. Chuang and Karen Lander
- Subjects
Communication ,business.industry ,Cognitive Neuroscience ,Work (physics) ,Experimental and Cognitive Psychology ,Body movement ,Motion (physics) ,Identity (music) ,Arts and Humanities (miscellaneous) ,Static image ,Face (geometry) ,Frame (artificial intelligence) ,Computer vision ,Optimal distinctiveness theory ,Artificial intelligence ,Psychology ,business - Abstract
Previous work has suggested that seeing a famous face move aids the recognition of identity, especially when viewing conditions are degraded (Knight Johnston, 1997; Lander, Christie, Bruce, 1999). Experiment 1 investigated whether the beneficial effects of motion are related to a particular type of facial motion (expressing, talking, or rigid motion). Results showed a significant beneficial effect of both expressive and talking movements, but no advantage for rigid motion, compared with a single static image. Experiment 2 investigated whether the advantage for motion is uniform across identity. Participants rated moving famous faces for distinctiveness of motion. The famous faces (moving and static freeze frame) were then used as stimuli in a recognition task. The advantage for face motion was significant only when the motion displayed was distinctive. Results suggest that a reason why moving faces are easier to recognize is because some familiar faces have characteristic motion patterns, which act as an additional cue to identity.
- Published
- 2005
- Full Text
- View/download PDF
15. Seeing: The Computational Approach to Biological Vision. Second Edition. By John P. Frisby and James V. Stone. Cambridge (Massachusetts): MIT Press. $55.00 (paper). xiv + 562 p.; ill.; index. ISBN: 978-0-262-51427-9. 2010
- Author
-
Heinrich H. Buelthoff and Lewis L. Chuang
- Subjects
Index (economics) ,media_common.quotation_subject ,Art ,General Agricultural and Biological Sciences ,Humanities ,media_common - Published
- 2011
- Full Text
- View/download PDF
16. Looking for discriminating is different from looking for looking's sake.
- Author
-
Hans-Joachim Bieg, Jean-Pierre Bresciani, Heinrich H Bülthoff, and Lewis L Chuang
- Subjects
Medicine ,Science - Abstract
Recent studies provide evidence for task-specific influences on saccadic eye movements. For instance, saccades exhibit higher peak velocity when the task requires coordinating eye and hand movements. The current study shows that the need to process task-relevant visual information at the saccade endpoint can be, in itself, sufficient to cause such effects. In this study, participants performed a visual discrimination task which required a saccade for successful completion. We compared the characteristics of these task-related saccades to those of classical target-elicited saccades, which required participants to fixate a visual target without performing a discrimination task. The results show that task-related saccades are faster and initiated earlier than target-elicited saccades. Differences between both saccade types are also noted in their saccade reaction time distributions and their main sequences, i.e., the relationship between saccade velocity, duration, and amplitude.
- Published
- 2012
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.