Thanks to visit codestin.com
Credit goes to link.springer.com

Skip to main content

Eye and Head Tracking for Focus of Attention Control in the Cockpit

  • Chapter
  • First Online:
Automotive User Interfaces

Part of the book series: Human–Computer Interaction Series ((HCIS))

  • 4561 Accesses

Abstract

The driver’s focus of attention is a key factor to be considered for building novel, intuitive user interaction concepts, and enhancing the current infotainment and safety applications in the vehicle. In this chapter we present several topics related to the development of application and systems that incorporate the user’s visual focus of attention. In the presented real-life experiments, 3D representations of both the vehicle’s interior and the outside environment are used. A real-time evaluation concerning the object in the driver’s visual focus in these environments is also performed. We describe the functionality and the accuracy of the presented systems, which is integrated in a fully functional vehicle in an actual traffic setting. In addition, several analyses concerning accuracy of the off-the-shelf eye trackers regarding peripheral vision or direct interaction with urban objects are presented.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+
from £29.99 /Month
  • Starting from 10 chapters or articles per month
  • Access and download chapters and articles from more than 300k books and 2,500 journals
  • Cancel anytime
View plans

Buy Now

Chapter
GBP 19.95
Price includes VAT (United Kingdom)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
GBP 127.50
Price includes VAT (United Kingdom)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
GBP 159.99
Price includes VAT (United Kingdom)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
GBP 159.99
Price includes VAT (United Kingdom)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    http://www.dssmining.com/what-we-do/results/.

  2. 2.

    https://unity3d.com/.

  3. 3.

    http://www.tobii.com/xperience/.

  4. 4.

    https://www.asus.com/3D-Sensor/Xtion_PRO/.

  5. 5.

    http://www.faceshift.com/.

  6. 6.

    http://structure.io/.

  7. 7.

    The object is visible and not too far or too close.

References

  • Alt, F., Schneegass, S., Auda, J., Rzayev, R., & Broy, N. (2014). Using eye tracking to support interaction with layered 3D interfaces on stereoscopic displays. In Proceedings of the 19th international conference on Intelligent User Interfaces (p. 267272). ACM.

    Google Scholar 

  • Dinh, B. -T. (2015). Accuracy measurement of low-cost eye trackers in 2D and 3D environments. Bachelors thesis, Computer Science Institute, Saarland University, Saarbrücken, Germany.

    Google Scholar 

  • Dinh, B. -T. (2015). Framework for analyzing drivers focus-of-attention based on dynamic 3D Map Data. Bachelors thesis, Computer Science Institute, Saarland University, Saarbrücken, Germany.

    Google Scholar 

  • Duchowski, A. (2007). Eye tracking methodology: Theory and practice (Vol. 373). Springer Science & Business Media.

    Google Scholar 

  • Fanelli, G., Weise, T., Gall, J., & Gool, L. J. V. (2011). Real time head pose estimation from consumer depth cameras. In R. Mester & M. Felsberg (Eds.), DAGM-Symposium. Lecture Notes in Computer Science (Vol. 6835, p. 101110). Springer.

    Google Scholar 

  • Fletcher, L., & Zelinsky, A. (2009). Driver inattention detection based on eye gaze–road event correlation. International Journal of Robotics Research, 28(6), 774801.

    Article  Google Scholar 

  • Gernoth, T., Martínez, K. A., Gooßen, A., & Grigat, R. -R. (2010). Facial pose estimation using active appearance models and a generic face model. In P. Richard & J. Braz (Eds.), VISAPP (Vol. 2, p. 499506). INSTICC Press.

    Google Scholar 

  • Hatada, T., Sakata, H., & Kusaka, H. (1980). Psychophysical analysis of the "sensation of reality" induced by a visual wide-field display. SMPTE Journal, 89(8), 560569.

    Article  Google Scholar 

  • Jiménez, P., Bergasa, L. M., Nuevo, J., & Alcantarilla, P. F. (2012). Face pose estimation with automatic 3D model creation in challenging scenarios. Image and Vision Computing, 30(9), 589602.

    Article  Google Scholar 

  • Kang, S., Kim B., Han, S., & Kim, H. (2015). Do you see what I see: Towards a gaze-based surroundings query processing system. In Proceedings of the 7th International Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI ’15) (pp. 93–100.). New York, NY, USA: ACM. http://dx.doi.org/10.1145/2799250.2799285

  • Kern, D., Mahr, A., Castronovo, S., Schmidt, A., & Müller, C. A. (2010). Making use of drivers’ glances onto the screen for explicit gaze-based interaction. In A. K. Dey, A. Schmidt, S. Boll, & A. L. Kun (Eds.), AutomotiveUI (p. 110116). ACM.

    Google Scholar 

  • Math, R., Mahr, A., Moniri, M. M., & Müller, C. (2012). Opends: A new open-source driving simulator for research. Adjunct Proceedings of the 4th International Conference on Automotive User Interfaces and Interactive Vehicular Appilcations (p. 78). Portsmouth: NH, USA.

    Google Scholar 

  • Merkel, D. (2015). System für die Aufmerksamkeitsanalyse des Fahrers basierend auf Blickdaten. Bachelors thesis, Computer Science Institute, Saarland University, Saarbrücken, Germany.

    Google Scholar 

  • Moniri, M. M. (2011). Multimodal reference resolution for mobile spatial interaction in urban environments (Masters thesis, Computer Science Institute, Saarland University, Saarbrücken, Germany).

    Google Scholar 

  • Moniri, M. M., & Müller, C. (2014). EyeVIUS: intelligent vehicles in intelligent urban spaces. In Adjunct Proceedings of the 6th International Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI ’14) (pp. 1–6). New York, NY, USA: ACM. http://dx.doi.org/10.1145/2667239.2667265

  • Moniri, M. M., Feld, M., & Müller, C. A. (2012). Personalized in-vehicle information systems: Building an application infrastructure for smart cars in smart spaces (p. 379382). In Intelligent Environments: IEEE.

    Google Scholar 

  • Nakao, M., Terada, T., & Tsukamoto, M. (2014). An information presentation method for head mounted display considering surrounding environments. In Proceedings of the 5th Augmented Human International Conference, (p. 47). ACM.

    Google Scholar 

  • Toyama, T., Dengel, A., Suzuki, W., & Kise, K. (2013). Wearable reading assist system: Augmented reality document combining document retrieval and eye tracking. In 2013 12th International Conference on Document Analysis and Recognition (ICDAR) (p. 34). IEEE.

    Google Scholar 

  • Toyama, T., Sonntag, D., Orlosky, J., & Kiyokawa, K. (2014). A natural interface for multi-focal plane head mounted displays using 3D gaze. In Proceedings of the 2014 International Working Conference on Advanced Visual Interfaces (p. 2532). ACM.

    Google Scholar 

  • Toyama, T., Sonntag, D., Orlosky, J., & Kiyokawa, K. (2015). Attention engagement and cognitive state analysis for augmented reality text display functions. In Proceedings of the 20th International Conference on Intelligent User Interfaces (p. 322332). ACM .

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Michael Feld .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this chapter

Cite this chapter

Moniri, M.M., Feld, M. (2017). Eye and Head Tracking for Focus of Attention Control in the Cockpit. In: Meixner, G., Müller, C. (eds) Automotive User Interfaces. Human–Computer Interaction Series. Springer, Cham. https://doi.org/10.1007/978-3-319-49448-7_9

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-49448-7_9

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-49447-0

  • Online ISBN: 978-3-319-49448-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Publish with us

Policies and ethics