You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

Dataset Card: GazeIntent = RadSeq & RadExplore & RadHybrid

Dataset Name: phamtrongthang/GazeIntent
Repository: UARK‑AICV/RadGazeIntent
License: CC BY-NC-SA 4.0


1. Dataset Summary

GazeIntent is the first intention-labeled eye-tracking dataset for radiological interpretation, capturing radiologist's diagnostic intentions during chest X-ray analysis. It includes:

  • 3,562 chest X-ray samples with expert radiologist eye-tracking data
  • Fine-grained intention labels for each fixation point
  • Three distinct intention modeling paradigms representing different visual search behaviors
  • Multi-label annotations for 13 radiological findings

This dataset supports research in intention interpretation, gaze-informed diagnosis, cognitive modeling, and explainable AI in medical imaging.

🏅 This work was accepted at ACM MM 2025 - A top-tier international conference on multimedia research.


2. Dataset Structure

Attribute Description
Total Samples 3,562 chest X-rays
Sources EGD (1,079) + REFLACX (2,483)
Modality Chest X-ray images
Gaze Data 2D coordinates + fixation duration + intention labels
Intention Classes 13 radiological findings
Radiologists Multiple expert radiologists

3. Three Intention Paradigms

RadSeq (Systematic Sequential Search)

  • Models radiologists following a structured diagnostic checklist
  • One finding examined at a time in sequential order
  • Reflects systematic, methodical visual search patterns

RadExplore (Uncertainty-driven Exploration)

  • Captures opportunistic visual search behavior
  • Radiologists consider multiple findings simultaneously
  • Represents exploratory, uncertainty-driven attention

RadHybrid (Hybrid Pattern)

  • Combines initial broad scanning with focused examination
  • Two-phase approach: overview → targeted search
  • Reflects real-world diagnostic behavior patterns

4. Intended Uses

  • Radiologist intention interpretation and prediction
  • Gaze-informed medical diagnosis systems
  • Cognitive modeling of expert visual reasoning
  • Medical education and training assessment
  • Explainable AI for radiology applications
  • Human-AI collaboration in medical imaging

5. Tasks and Benchmarks

Primary Task: Fixation-based Intention Classification

  • Baseline: RadGazeIntent (transformer-based architecture)
  • Input: Fixation sequences + chest X-ray images
  • Output: Intention confidence scores for 13 findings

Evaluation Metrics:

  • Classification: Accuracy, F1-score, Precision, Recall
  • Multi-label: Per-class and macro-averaged metrics

Findings Covered: Atelectasis, Cardiomegaly, Consolidation, Edema, Enlarged Cardiomediastinum, Fracture, Lung Lesion, Lung Opacity, Pleural Effusion, Pleural Other, Pneumonia, Pneumothorax, Support Devices


6. Data Availability

The processed intention-labeled datasets are publicly available via Hugging Face under CC BY-NC-SA 4.0 license.

Access Requirements: Users must agree to share contact information and accept the license terms to access the dataset files.


7. Technical Details

Data Processing: Three datasets derived from existing eye-tracking sources (EGD, REFLACX) using different intention modeling assumptions:

  • Uncertainty Filtering: Assigns labels based on temporal alignment with radiologist transcripts
  • Sequential Constraints: Applies GazeSearch methodology for systematic search modeling
  • Hybrid Integration: Combines initial scanning phase with focused examination periods

8. Citation

Please cite this dataset using the following BibTeX entry:

@article{pham2025interpreting,
  title={Interpreting Radiologist's Intention from Eye Movements in Chest X-ray Diagnosis},
  author={Pham, Trong-Thang and Nguyen, Anh and Deng, Zhigang and Wu, Carol C and Nguyen, Hien and Le, Ngan},
  journal={arXiv preprint arXiv:2507.12461},
  year={2025}
}

9. Acknowledgments

This work is supported by:

  • National Science Foundation (NSF) Award No OIA-1946391, NSF 2223793 EFRI BRAID
  • National Institutes of Health (NIH) 1R01CA277739-01
  • Built upon EGD and REFLACX eye-tracking datasets

Contact: Trong Thang Pham ([email protected])

Downloads last month
21