ADAPT: A predictive cognitive model of user visual attention and action planning

Stephanie M. Doane, Young Woo Sohn

Research output: Contribution to journalArticlepeer-review

22 Citations (Scopus)


We present a computational cognitive model of novice and expert aviation pilot action planning called ADAPT that models performance in a dynamically changing simulated flight environment. We perform rigorous tests of ADAPT's predictive validity by comparing the performance of individual human pilots to that of their respective models. Individual pilots were asked to execute a series of flight maneuvers using a flight simulator, and their eye fixations and control movements were recorded in a time-synched database. Computational models of each of the 25 individual pilots were constructed, and the individual models simulated execution of the same flight maneuvers performed by the human pilots. The time-synched eye fixations and control movements of individual pilots and their respective models were compared, and rigorous tests of ADAPT's predictive validity were performed. The model explains and predicts a significant portion of pilot visual attention and control movements during flight as a function of piloting expertise. Implications for adaptive training systems are discussed.

Original languageEnglish
Pages (from-to)1-45
Number of pages45
JournalUser Modelling and User-Adapted Interaction
Issue number1
Publication statusPublished - 2000

Bibliographical note

Funding Information:
This research was supported by funding from the Office of Naval Research. The authors wish to thank Drs Susan Chipman, Harold Hawkins, and Terry Allerd from ONR for their support of this effort. We would also like to thank Drs Gary Bradshaw, Alfred Kobsa, and three anonymous reviewers for their comments on previous versions of this manuscript.

All Science Journal Classification (ASJC) codes

  • Education
  • Human-Computer Interaction
  • Computer Science Applications


Dive into the research topics of 'ADAPT: A predictive cognitive model of user visual attention and action planning'. Together they form a unique fingerprint.

Cite this