Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels

Many gaze data visualization techniques intuitively show eye movement together with visual stimuli. The eye tracker records a large number of eye movements within a short period. Therefore, visualizing raw gaze data with the visual stimulus appears complicated and obscured, making it difficult to ga...

Full description

Bibliographic Details
Main Authors: Sangbong Yoo, Seongmin Jeong, Yun Jang
Format: Article
Language:English
Published: MDPI AG 2021-07-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/14/4686
id doaj-9ca47798a4e3492d9b683c3668f42b13
record_format Article
spelling doaj-9ca47798a4e3492d9b683c3668f42b132021-07-23T14:05:18ZengMDPI AGSensors1424-82202021-07-01214686468610.3390/s21144686Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction LevelsSangbong Yoo0Seongmin Jeong1Yun Jang2Computer Engineering and Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, KoreaComputer Engineering and Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, KoreaComputer Engineering and Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, KoreaMany gaze data visualization techniques intuitively show eye movement together with visual stimuli. The eye tracker records a large number of eye movements within a short period. Therefore, visualizing raw gaze data with the visual stimulus appears complicated and obscured, making it difficult to gain insight through visualization. To avoid the complication, we often employ fixation identification algorithms for more abstract visualizations. In the past, many scientists have focused on gaze data abstraction with the attention map and analyzed detail gaze movement patterns with the scanpath visualization. Abstract eye movement patterns change dramatically depending on fixation identification algorithms in the preprocessing. However, it is difficult to find out how fixation identification algorithms affect gaze movement pattern visualizations. Additionally, scientists often spend much time on adjusting parameters manually in the fixation identification algorithms. In this paper, we propose a gaze behavior-based data processing method for abstract gaze data visualization. The proposed method classifies raw gaze data using machine learning models for image classification, such as CNN, AlexNet, and LeNet. Additionally, we compare the velocity-based identification (I-VT), dispersion-based identification (I-DT), density-based fixation identification, velocity and dispersion-based (I-VDT), and machine learning based and behavior-based modelson various visualizations at each abstraction level, such as attention map, scanpath, and abstract gaze movement visualization.https://www.mdpi.com/1424-8220/21/14/4686gaze data visualizationgaze behaviormachine learning
collection DOAJ
language English
format Article
sources DOAJ
author Sangbong Yoo
Seongmin Jeong
Yun Jang
spellingShingle Sangbong Yoo
Seongmin Jeong
Yun Jang
Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
Sensors
gaze data visualization
gaze behavior
machine learning
author_facet Sangbong Yoo
Seongmin Jeong
Yun Jang
author_sort Sangbong Yoo
title Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
title_short Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
title_full Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
title_fullStr Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
title_full_unstemmed Gaze Behavior Effect on Gaze Data Visualization at Different Abstraction Levels
title_sort gaze behavior effect on gaze data visualization at different abstraction levels
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2021-07-01
description Many gaze data visualization techniques intuitively show eye movement together with visual stimuli. The eye tracker records a large number of eye movements within a short period. Therefore, visualizing raw gaze data with the visual stimulus appears complicated and obscured, making it difficult to gain insight through visualization. To avoid the complication, we often employ fixation identification algorithms for more abstract visualizations. In the past, many scientists have focused on gaze data abstraction with the attention map and analyzed detail gaze movement patterns with the scanpath visualization. Abstract eye movement patterns change dramatically depending on fixation identification algorithms in the preprocessing. However, it is difficult to find out how fixation identification algorithms affect gaze movement pattern visualizations. Additionally, scientists often spend much time on adjusting parameters manually in the fixation identification algorithms. In this paper, we propose a gaze behavior-based data processing method for abstract gaze data visualization. The proposed method classifies raw gaze data using machine learning models for image classification, such as CNN, AlexNet, and LeNet. Additionally, we compare the velocity-based identification (I-VT), dispersion-based identification (I-DT), density-based fixation identification, velocity and dispersion-based (I-VDT), and machine learning based and behavior-based modelson various visualizations at each abstraction level, such as attention map, scanpath, and abstract gaze movement visualization.
topic gaze data visualization
gaze behavior
machine learning
url https://www.mdpi.com/1424-8220/21/14/4686
work_keys_str_mv AT sangbongyoo gazebehavioreffectongazedatavisualizationatdifferentabstractionlevels
AT seongminjeong gazebehavioreffectongazedatavisualizationatdifferentabstractionlevels
AT yunjang gazebehavioreffectongazedatavisualizationatdifferentabstractionlevels
_version_ 1721286007459938304