Citation and metadata
Recommended citation
Thies Pfeiffer, Marc E. Latoschik, and Ipke Wachsmuth, Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments. JVRB - Journal of Virtual Reality and Broadcasting, 5(2008), no. 16. (urn:nbn:de:0009-6-16605)
Download Citation
Endnote
%0 Journal Article %T Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments %A Pfeiffer, Thies %A Latoschik, Marc E. %A Wachsmuth, Ipke %J JVRB - Journal of Virtual Reality and Broadcasting %D 2008 %V 5(2008) %N 16 %@ 1860-2037 %F pfeiffer2008 %X Tracking user’s visual attention is a fundamental aspectin novel human-computer interaction paradigmsfound in Virtual Reality. For example, multimodal interfacesor dialogue-based communications with virtualand real agents greatly benefit from the analysis ofthe user’s visual attention as a vital source for deicticreferences or turn-taking signals. Current approachesto determine visual attention rely primarily on monoculareye trackers. Hence they are restricted to the interpretationof two-dimensional fixations relative to adefined area of projection.The study presented in this article compares precision,accuracy and application performance of twobinocular eye tracking devices. Two algorithms arecompared which derive depth information as requiredfor visual attention-based 3D interfaces. This informationis further applied to an improved VR selectiontask in which a binocular eye tracker and an adaptiveneural network algorithm is used during the disambiguationof partly occluded objects. %L 004 %K Eye Tracking %K Human-Computer Interaction %K Object Selection %K Virtual Reality %R 10.20385/1860-2037/5.2008.16 %U http://nbn-resolving.de/urn:nbn:de:0009-6-16605 %U http://dx.doi.org/10.20385/1860-2037/5.2008.16Download
Bibtex
@Article{pfeiffer2008, author = "Pfeiffer, Thies and Latoschik, Marc E. and Wachsmuth, Ipke", title = "Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments", journal = "JVRB - Journal of Virtual Reality and Broadcasting", year = "2008", volume = "5(2008)", number = "16", keywords = "Eye Tracking; Human-Computer Interaction; Object Selection; Virtual Reality", abstract = "Tracking user's visual attention is a fundamental aspectin novel human-computer interaction paradigmsfound in Virtual Reality. For example, multimodal interfacesor dialogue-based communications with virtualand real agents greatly benefit from the analysis ofthe user's visual attention as a vital source for deicticreferences or turn-taking signals. Current approachesto determine visual attention rely primarily on monoculareye trackers. Hence they are restricted to the interpretationof two-dimensional fixations relative to adefined area of projection.The study presented in this article compares precision,accuracy and application performance of twobinocular eye tracking devices. Two algorithms arecompared which derive depth information as requiredfor visual attention-based 3D interfaces. This informationis further applied to an improved VR selectiontask in which a binocular eye tracker and an adaptiveneural network algorithm is used during the disambiguationof partly occluded objects.", issn = "1860-2037", doi = "10.20385/1860-2037/5.2008.16", url = "http://nbn-resolving.de/urn:nbn:de:0009-6-16605" }Download
RIS
TY - JOUR AU - Pfeiffer, Thies AU - Latoschik, Marc E. AU - Wachsmuth, Ipke PY - 2008 DA - 2008// TI - Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments JO - JVRB - Journal of Virtual Reality and Broadcasting VL - 5(2008) IS - 16 KW - Eye Tracking KW - Human-Computer Interaction KW - Object Selection KW - Virtual Reality AB - Tracking user’s visual attention is a fundamental aspectin novel human-computer interaction paradigmsfound in Virtual Reality. For example, multimodal interfacesor dialogue-based communications with virtualand real agents greatly benefit from the analysis ofthe user’s visual attention as a vital source for deicticreferences or turn-taking signals. Current approachesto determine visual attention rely primarily on monoculareye trackers. Hence they are restricted to the interpretationof two-dimensional fixations relative to adefined area of projection.The study presented in this article compares precision,accuracy and application performance of twobinocular eye tracking devices. Two algorithms arecompared which derive depth information as requiredfor visual attention-based 3D interfaces. This informationis further applied to an improved VR selectiontask in which a binocular eye tracker and an adaptiveneural network algorithm is used during the disambiguationof partly occluded objects. SN - 1860-2037 UR - http://nbn-resolving.de/urn:nbn:de:0009-6-16605 DO - 10.20385/1860-2037/5.2008.16 ID - pfeiffer2008 ER -Download
Wordbib
<?xml version="1.0" encoding="UTF-8"?> <b:Sources SelectedStyle="" xmlns:b="http://schemas.openxmlformats.org/officeDocument/2006/bibliography" xmlns="http://schemas.openxmlformats.org/officeDocument/2006/bibliography" > <b:Source> <b:Tag>pfeiffer2008</b:Tag> <b:SourceType>ArticleInAPeriodical</b:SourceType> <b:Year>2008</b:Year> <b:PeriodicalTitle>JVRB - Journal of Virtual Reality and Broadcasting</b:PeriodicalTitle> <b:Volume>5(2008)</b:Volume> <b:Issue>16</b:Issue> <b:Url>http://nbn-resolving.de/urn:nbn:de:0009-6-16605</b:Url> <b:Url>http://dx.doi.org/10.20385/1860-2037/5.2008.16</b:Url> <b:Author> <b:Author><b:NameList> <b:Person><b:Last>Pfeiffer</b:Last><b:First>Thies</b:First></b:Person> <b:Person><b:Last>Latoschik</b:Last><b:First>Marc E.</b:First></b:Person> <b:Person><b:Last>Wachsmuth</b:Last><b:First>Ipke</b:First></b:Person> </b:NameList></b:Author> </b:Author> <b:Title>Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments</b:Title> <b:Comments>Tracking user’s visual attention is a fundamental aspectin novel human-computer interaction paradigmsfound in Virtual Reality. For example, multimodal interfacesor dialogue-based communications with virtualand real agents greatly benefit from the analysis ofthe user’s visual attention as a vital source for deicticreferences or turn-taking signals. Current approachesto determine visual attention rely primarily on monoculareye trackers. Hence they are restricted to the interpretationof two-dimensional fixations relative to adefined area of projection.The study presented in this article compares precision,accuracy and application performance of twobinocular eye tracking devices. Two algorithms arecompared which derive depth information as requiredfor visual attention-based 3D interfaces. This informationis further applied to an improved VR selectiontask in which a binocular eye tracker and an adaptiveneural network algorithm is used during the disambiguationof partly occluded objects.</b:Comments> </b:Source> </b:Sources>Download
ISI
PT Journal AU Pfeiffer, T Latoschik, M Wachsmuth, I TI Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments SO JVRB - Journal of Virtual Reality and Broadcasting PY 2008 VL 5(2008) IS 16 DI 10.20385/1860-2037/5.2008.16 DE Eye Tracking; Human-Computer Interaction; Object Selection; Virtual Reality AB Tracking user’s visual attention is a fundamental aspectin novel human-computer interaction paradigmsfound in Virtual Reality. For example, multimodal interfacesor dialogue-based communications with virtualand real agents greatly benefit from the analysis ofthe user’s visual attention as a vital source for deicticreferences or turn-taking signals. Current approachesto determine visual attention rely primarily on monoculareye trackers. Hence they are restricted to the interpretationof two-dimensional fixations relative to adefined area of projection.The study presented in this article compares precision,accuracy and application performance of twobinocular eye tracking devices. Two algorithms arecompared which derive depth information as requiredfor visual attention-based 3D interfaces. This informationis further applied to an improved VR selectiontask in which a binocular eye tracker and an adaptiveneural network algorithm is used during the disambiguationof partly occluded objects. ERDownload
Mods
<mods> <titleInfo> <title>Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments</title> </titleInfo> <name type="personal"> <namePart type="family">Pfeiffer</namePart> <namePart type="given">Thies</namePart> </name> <name type="personal"> <namePart type="family">Latoschik</namePart> <namePart type="given">Marc E.</namePart> </name> <name type="personal"> <namePart type="family">Wachsmuth</namePart> <namePart type="given">Ipke</namePart> </name> <abstract>Tracking user’s visual attention is a fundamental aspect in novel human-computer interaction paradigms found in Virtual Reality. For example, multimodal interfaces or dialogue-based communications with virtual and real agents greatly benefit from the analysis of the user’s visual attention as a vital source for deictic references or turn-taking signals. Current approaches to determine visual attention rely primarily on monocular eye trackers. Hence they are restricted to the interpretation of two-dimensional fixations relative to a defined area of projection. The study presented in this article compares precision, accuracy and application performance of two binocular eye tracking devices. Two algorithms are compared which derive depth information as required for visual attention-based 3D interfaces. This information is further applied to an improved VR selection task in which a binocular eye tracker and an adaptive neural network algorithm is used during the disambiguation of partly occluded objects.</abstract> <subject> <topic>Eye Tracking</topic> <topic>Human-Computer Interaction</topic> <topic>Object Selection</topic> <topic>Virtual Reality</topic> </subject> <classification authority="ddc">004</classification> <relatedItem type="host"> <genre authority="marcgt">periodical</genre> <genre>academic journal</genre> <titleInfo> <title>JVRB - Journal of Virtual Reality and Broadcasting</title> </titleInfo> <part> <detail type="volume"> <number>5(2008)</number> </detail> <detail type="issue"> <number>16</number> </detail> <date>2008</date> </part> </relatedItem> <identifier type="issn">1860-2037</identifier> <identifier type="urn">urn:nbn:de:0009-6-16605</identifier> <identifier type="doi">10.20385/1860-2037/5.2008.16</identifier> <identifier type="uri">http://nbn-resolving.de/urn:nbn:de:0009-6-16605</identifier> <identifier type="citekey">pfeiffer2008</identifier> </mods>Download
Full Metadata
Bibliographic Citation | JVRB, 5(2008), no. 16. |
---|---|
Title |
Evaluation of Binocular Eye Trackers and Algorithms for 3D Gaze Interaction in Virtual Reality Environments (eng) |
Author | Thies Pfeiffer, Marc E. Latoschik, Ipke Wachsmuth |
Language | eng |
Abstract | Tracking user’s visual attention is a fundamental aspect in novel human-computer interaction paradigms found in Virtual Reality. For example, multimodal interfaces or dialogue-based communications with virtual and real agents greatly benefit from the analysis of the user’s visual attention as a vital source for deictic references or turn-taking signals. Current approaches to determine visual attention rely primarily on monocular eye trackers. Hence they are restricted to the interpretation of two-dimensional fixations relative to a defined area of projection. The study presented in this article compares precision, accuracy and application performance of two binocular eye tracking devices. Two algorithms are compared which derive depth information as required for visual attention-based 3D interfaces. This information is further applied to an improved VR selection task in which a binocular eye tracker and an adaptive neural network algorithm is used during the disambiguation of partly occluded objects. |
Subject | Eye Tracking, Human-Computer Interaction, Object Selection, Virtual Reality |
Classified Subjects |
|
DDC | 004 |
Rights | DPPL |
URN: | urn:nbn:de:0009-6-16605 |
DOI | https://doi.org/10.20385/1860-2037/5.2008.16 |