Input Method Using Divergence Eye Movement

Similar documents
Evaluation of gaze-added target selection methods suitable for general GUIs

The Effects of Semantic Grouping on Visual Search

Natural Viewing 3D Display

Real Time Eye Movement Identification Protocol

Helping people with ICT device control by eye gaze

Convergence Point Adjustment Methods for Minimizing Visual Discomfort Due to a Stereoscopic Camera

Usability Study on the Use of Eye Mouse Based on All the Functions of Conventional Mouse

An Auto-Stereoscopic VRML Viewer for 3D Data on the World Wide Web

VISIOFFICE 2 - USER MANUAL

Mobile Phone Operations using Human Eyes Only and its Applications

Arranging Touch Screen Software Keyboard Split-keys based on Contact Surface

Visual Appeal vs. Usability: Which One Influences User Perceptions of a Website More?

UniGest: Text Entry Using Three Degrees of Motion

Gaze Computer Interaction on Stereo Display

INTERACTIVE ENVIRONMENT FOR INTUITIVE UNDERSTANDING OF 4D DATA. M. Murata and S. Hashimoto Humanoid Robotics Institute, Waseda University, Japan

Augmenting Reality with Projected Interactive Displays

REGION MAP GENERATION OF EYE PUPIL MOVEMENTS TO PRESCRIBE CUSTOMISED PROGRESSIVE LENS

FLY THROUGH VIEW VIDEO GENERATION OF SOCCER SCENE

A Mouse-Like Hands-Free Gesture Technique for Two-Dimensional Pointing

Scroll Display: Pointing Device for Palmtop Computers

Interactive Video Retrieval System Integrating Visual Search with Textual Search

A Robot Recognizing Everyday Objects

Tobii Technology AB. Accuracy and precision Test report. TX300 fw RC Bright Light Illumination Mode

Preliminary Evidence for Top-down and Bottom-up Processes in Web Search Navigation

3D Digitization of a Hand-held Object with a Wearable Vision Sensor

A Comparison of Gaze-Based and Gesture-Based Input for a Point-and-Click Task

EVALUATING DIFFERENT TOUCH-BASED INTERACTION TECHNIQUES IN A PUBLIC INFORMATION KIOSK

EFFECTS OF KEY SIZE AND SPACING ON THE COMPLETION TIME AND ACCURACY OF INPUT TASKS ON SOFT KEYPADS USING TRACKBALL AND TOUCH INPUT

Tobii Technology AB. Accuracy and precision Test report. Tobii T60 XL Eye tracker. Date: Methodology/Software version: 2.1.

Free head motion eye gaze tracking using a single camera and multiple light sources

Sketch-based Interface for Crowd Animation

An Interactive Technique for Robot Control by Using Image Processing Method

Reprint. from the Journal. of the SID

Extended Fractional View Integral Photography Using Slanted Orthogonal Lenticular Lenses

A Method to Measure Eye-Hand Coordination for Extracting Skilled Elements-Simultaneous Measurement of Eye-Gaze and Hand Location-

Experimental Humanities II. Eye-Tracking Methodology

Comparing the Usability of RoboFlag Interface Alternatives*

Dynamic Stereoscopic 3D Parameter Adjustments for Enhanced Depth Discrimination

Basic distinctions. Definitions. Epstein (1965) familiar size experiment. Distance, depth, and 3D shape cues. Distance, depth, and 3D shape cues

Visual Interaction Design for Tools to Think with: Interactive Systems for Designing Linear Information

index of refraction-light speed

Paging vs. Scrolling: Looking for the Best Way to Present Search Results

Proceedings of NTCIR-9 Workshop Meeting, December 6-9, 2011, Tokyo, Japan

User Calibration-free Method using Corneal Surface Image for Eye Tracking

Phone Tilt. Dynamic UI Adaptations for single-handed smartphone interaction. Fasil Negash Mtr No.:

Outdoor Scene Reconstruction from Multiple Image Sequences Captured by a Hand-held Video Camera

Voronoi Diagrams, Vectors and the Visually Impaired

3D Interaction Techniques for Virtual Environments: Selection and Manipulation. Doug A. Bowman

POSE ESTIMATION OF CONSTRUCTION MATERIALS USING MULTIPLE ID DEVICES FOR CONSTRUCTION AUTOMATION

Available online at ScienceDirect. Procedia Computer Science 35 (2014 )

A Breakdown of the Psychomotor Components of Input Device Usage

Toward Part-based Document Image Decoding

Task analysis based on observing hands and objects by vision

Dense 3-D Reconstruction of an Outdoor Scene by Hundreds-baseline Stereo Using a Hand-held Video Camera

Margarita Grinvald. Gesture recognition for Smartphones/Wearables

BINOCULAR DISPARITY AND DEPTH CUE OF LUMINANCE CONTRAST. NAN-CHING TAI National Taipei University of Technology, Taipei, Taiwan

Projection Center Calibration for a Co-located Projector Camera System

Comparison of Accommodation and Convergence by Simultaneous Measurements during 2D and 3D Vision Gaze

Eye Movements With Peripheral Visual Field Loss While Walking

An Initial Seed Selection Algorithm for K-means Clustering of Georeferenced Data to Improve

AR Cultural Heritage Reconstruction Based on Feature Landmark Database Constructed by Using Omnidirectional Range Sensor

Analyzing the Relationship Between Head Pose and Gaze to Model Driver Visual Attention

Improving Stack Overflow Tag Prediction Using Eye Tracking Alina Lazar Youngstown State University Bonita Sharif, Youngstown State University

Analytic Provenance Datasets: A Data Repository of Human Analysis Activity and Interaction Logs

EyEgaZE & other access products

Prof. Feng Liu. Spring /27/2014

AP Physics: Curved Mirrors and Lenses

Virtual Interaction System Based on Optical Capture

Real-time Integral Photography Holographic Pyramid using a Game Engine

EyeLink Data Viewer User s Manual

Domain Specific Search Engine for Students

Rectification of distorted elemental image array using four markers in three-dimensional integral imaging

Department of Photonics, NCTU, Hsinchu 300, Taiwan. Applied Electromagnetic Res. Inst., NICT, Koganei, Tokyo, Japan

Area Type Judgment of Mathematical Document Using Movements of Gaze Point

Capturing Window Attributes for Extending Web Browsing History Records

HeadMouse. Robotic Research Team. University of Lleida

Eye characteristic detection and gaze direction calculation in gaze tracking system

Mousetracking Visitors to Evaluate Efficacy of Web Site Design

Conflict? What conflict?

Reflection & Refraction

Color Universal Design without Restricting Colors and Their Combinations Using Lightness Contrast Dithering

REFLECTION & REFRACTION

From Gaze to Focus of Attention

Tobii Technology AB. Accuracy and precision Test report. X2-30 fw Date: Methodology/Software version: 2.1.7

HOW TO RUN THE SCREENINGS

/$ IEEE

Comparison of Postures from Pen and Mouse Use

Gesture Recognition using Temporal Templates with disparity information

Under My Finger: Human Factors in Pushing and Rotating Documents Across the Table

Motion analysis for broadcast tennis video considering mutual interaction of players

Accuracy and precision of the Tobii X2-30 eye-tracking under non ideal conditions

Non-line-of-sight imaging

Cascading versus Indexed Menu Design

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Web Front Technology: Interstage Interaction Manager

Light: Geometric Optics

Mahdi Amiri. May Sharif University of Technology

Manhattan-World Assumption for As-built Modeling Industrial Plant

the Hick Hyman Law Pearson Addison-Wesley. All rights reserved. 6-1

PastePost: A Web Interface for Editing Instructions with Captured Image

Transcription:

Input Method Using Divergence Eye Movement Shinya Kudo kudo@kaji-lab.jp Hiroyuki Okabe h.okabe@kaji-lab.jp Taku Hachisu JSPS Research Fellow hachisu@kaji-lab.jp Michi Sato JSPS Research Fellow michi@kaji-lab.jp Shogo Fukushima JSPS Research Fellow shogo@kaji-lab.jp Hiroyuki Kajimoto Japan Science and Technology Agency kajimoto@kaji-lab.jp Abstract A gaze input interface offers hands-free operation by using the view-point position as the cursor coordinates on the display. However, the selection operation of a button is indistinguishable from viewing; this is known as the Midas touch problem. We propose a new input method that measures divergence eye movement, thereby enabling users to press a button by moving their viewpoint forward. Comparison of our method and the conventional blinking input method confirms that input speed and accuracy are similar. Keywords Divergence eye movement; gaze input interface; depth input ACM Classification Keywords H.5.2. [Information interfaces and presentation]: Usability Testing and Evaluation General Terms Human Factors Copyright is held by the author/owner(s). CHI 2013 Extended Abstracts, April 27 May 2, 2013, Paris, France. ACM 978-1-4503-1952-2/13/04.

Introduction Gaze input has been proposed as an intuitive input method for a personal computer (PC) [1]. By moving their eyes, users can move the cursor and select buttons. However, there is a well-known Midas touch problem [2], meaning that this method misinterprets straightforward gazing as a selection operation. Several solutions have been proposed for dealing with this problem. Hansen et al. [3], Murata et al. [4], and Dario [5] reduced incorrect selection by using gaze time, while Shaw et al. achieved this by using blinking [6]. However, these methods have a trade-off between selection precision and selection speed. Zhai et al. [7] and Yamato et al. [8] avoided incorrect selection by using the user's gaze as cursor movement only and using a mouse for selection; however, use of a mouse is not applicable in some applications that originally required gaze input. Our goal is to realize intuitive selection by using an action other than gazing or blinking. In this paper, we propose a new input method that measures divergence eye movement, thereby enables users to press a button by moving their viewpoint forward. Method Divergence eye movement is a type of eye movement that occurs when a person moves his/her viewpoint forward (Figure 1). Pfeiffer et al. tried to estimate 3- dimensional (3D) positions of viewpoints by detecting convergence [9]. Sato et al. used the same information to investigate cross-sections of 3D models [10]. We apply this technique as the selection operation by detecting the forward movement of the viewpoint. We call this method the depth input method. Figure 1. Angle of convergence and divergence eye movement First, we obtain the viewpoints of the two eyes on the display plane using an eye-tracking device. When divergence eye movement occurs, the viewpoint of the right eye moves to the right and that of the left eye moves to the left. By setting the distance between the two viewpoints as dx, the interpupillary distance as dp, and the distance between the eyes and display as d, the depth of viewpoint is calculated as follows (Figure 2). Depth= d dx (1) dp dx

Figure 2. Estimation of the depth viewpoint position Figure 3. Presentation technique A 3D-shaped button is presented on the display, moving vertically to the display according to the change in estimated depth of viewpoint (Figure 3). A preliminary experiment revealed that intentionally staring at empty space is difficult without training. Therefore, we provided a visual marker on the rear of the display using a half mirror on the display and a marker in the front. Adjusting the angle of convergence becomes easier by watching a mirror image of the marker. This setup can be simplified by using a 3D monitor, but this is beyond the scope of this paper. Experiment We conducted experiments to verify the operability of the proposed method, compared with the conventional blinking input method. Experimental setup The experimental setup is shown in Figure 5. A chin support was set up on the desk 700mm in front of the display to fix the eye position. A half mirror was fixed to the display. A visual marker was placed 300mm in front of the display so that the mirror image of the marker from the half mirror was 300 mm behind the display. The distance between the two viewpoints (dx) was 20 mm when the participants looked at the image of the marker.

as a selection operation. In the blinking type method, the button is pressed when the participant's eyes remain closed for 0.5 s. (This duration is typically used to discriminate between natural blinking and selection [6]). Figure 4. Overview of the experiment Figure 5. Experiment setup Nine buttons numbered 1 to 9, 40 mm square (3.27 deg in terms of viewing angle) were arranged in a 3 3 format on the display (Figure 6). Five randomly chosen numbers were displayed above the central button. Participants were asked to sequentially press the buttons corresponding to the five displayed numbers ignoring incorrect input. They were also asked to complete the task as quickly and accurately as possible. Input of five numbers was regarded as one trial, and each participant conducted ten trials for each method. Interpupillary distance (dp) was set to 65mm. We used a desktop type eye tracking device (TM3 EyeTech Digital Systems), which uses the pupil-corneal reflection method. Viewpoint resolution was 1.0 deg. We conducted a calibration of the viewpoints on the display plane for each participant before the experiment. The display resolution was 1920 1080 and viewing angle 35.7 deg 23.3 deg. We used OpenGL for drawing. An image of the actual experiment is shown in Figure 4. Experimental condition We compared the depth input method and the conventional blinking input method. In the former method, a button is pressed as the participant moves his/her viewpoint forward to the image of the marker. When the viewpoint reaches the image, it is regarded Figure 6. Display during experiment

Figure 7. Correct answer rates (%). Figure 8. Input time (s). Procedure First, the participants practiced the depth input method for about 1-2 min to familiarize themselves with the forward movement of their viewpoints. After the experiment started, participants were instructed to enter the five numbers, while the input speed and accuracy were recorded. Participants were given a 5- second break before starting the next trial. After completion of the ten trials, participants were given a 5-min break. Thereafter, they carried out the experiment using the blinking input method in the same way as the depth input method. The participants comprised four males (one male with glasses, two males wearing soft contact lenses and the other is naked eye) and 4 females (one female with glasses, two females wearing soft contact lenses and the other is naked eye), respectively, aged between 21 to 28 years. Results The correct answer rates and input times for all participants are shown in Figure 7 and Figure 8. Correct answer rates for both methods were close to 90%, no significant difference was observed (Chisquare test: X-squared = 0.26, df = 1, p-value = 0.61 > 0.05). Input times for the blinking input method were mostly 2.1 s faster than those of the depth input method. However, there was no significant difference between the input times of the two methods (t-test: t = 0.9093, df = 13.354, p-value = 0.3793 > 0.05). Discussion The experiment revealed that the depth input method produces similar performance to the conventional blinking input method. As the blinking input method distinguishes selection and natural blinking by the duration of blinking, it prevents users from closing their eyes for a while, which is a natural behavior when the eyes become fatigued. Conversely, the depth input method can be distinguished from other eye movements, but still, retains some intuitiveness. During the experiment, we observed a relatively large deviation among the participants. This was probably due to the difference in accuracy of the calibration as a result of wearing glasses or contact lenses. The depth input method requires more accurate calibration than the blinking input method because it uses the distance between the two viewpoints. Therefore, any difference in the calibration accuracy significantly affects the result. Additionally, the calibration was done on a 2D display plane only, which is not sufficient for 3D viewpoint movement. Pfeiffer et al. carried out a 3D calibration on a 3D display [9], which should be done in our system. After the experiment, there was a comment that eye fatigue occurred during the depth input method rather than the blinking input method. This is due to the unnaturally large eye movements in the depth direction. There was also a comment that the button appeared in double vision during input, which was not comfortable. These problems should be solved by shortening the depth input length through more precise calibration, and by using a 3D display. Conclusion In this paper, we proposed a new input method that measures divergence eye movements, enabling user to "press" a button by moving his/her viewpoint forward. We compared this method with the conventional

blinking input method and confirmed that input speed and accuracy are comparable. Thus, the depth input method achieves similar performance to the conventional blinking input method. In the future, we will continue to improve the depth input method by using a 3D display, thereby making it easier to understand. References [1] Sibert, L.E. and Jacob, R.J.K. Evaluation of eye gaze interaction. In Proc. CHI 2000, ACM Press (2000), 281-288. [2] Jacob, R.J.K. What you look at is what you get: eye movement-based interaction techniques. In Proc. CHI 1990, ACM Press (1990). [3] Hansen, J.P., Anderson, A.W., and Roed, P. Eyegaze control of multimedia systems. In Symbiosis of human and artifact, Vol. 20A. Anzai, Y., Ogawa, K., and Mori, H. (Eds.), Elsevier Science (1995), 37-42. [4] Murata, A. and Miyake, T. Effectiveness of eye-gaze input system: identification of conditions that assures high pointing accuracy and movement directional effect. In Fourth International Workshop on Computational Intelligence & Applications, IEEE (2008), 127-132. [5] Dario D. Inferring intent in eye-based interfaces: tracing eye movements with process model. In Proc. CHI 1999, ACM Press (1999), 15-20. [6] Shaw, R., Crisman, E., Loomis, A., and Laszewski, Z. The eye wink control interface: using the computer to provide the severely disabled with increased flexibility and comfort. In Third Annual IEEE Symposium on Computer-Based Medical Systems, IEEE (1990), 105-111. [7] Zhai, S., Morimoto, C., and Ihde, S. Manual and gaze input cascaded (MAGIC) pointing. In Proc. CHI 1999, ACM Press (1999), 246-253. [8] Yamato, M., Monden, A., Matsumoto, K., Inoue, K., and Torii, K. Quick button selection with eye Gazing for general GUI environment. In Proc. International Conference on Software: Theory and Practice (2000). [9] Pfeiffer, T., Latoschik, Marc, E., and Wachsmuth, I. Evaluation of binocular eye trackers and algorithms for 3D gaze interaction in virtual reality environments. Journal of Virtual Reality and Broadcasting, 5, 16 (2008). [10] Sato, M. and Kajimoto, H. Dynamic stereograms based on eye convergence for displaying multilayered images. In SIGGRAPH ASIA, ACM Press (2012).