Gaze Meets CV workshop in conjunction with WACV 2025 Call for Papers
***************************************************************************
The 2025 Gaze Meets CV workshop in conjunction with WACV 2025
****************************************************************************
Webpage: https://sites.google.com/view/gmcv-workshop-wacv2025
Twitter Handle: https://twitter.com/Gaze_Meets_ML
Submission site: https://cmt3.research.microsoft.com/GMCV2025
Submission deadline: November 30th, 2024
Date: Feb 28th - Mar 4th
Location: Tucson, Arizona, USA
** Overview **
We are excited to host the Gaze Meets Computer Vision (GMCV) Workshop,
in conjunction with WACV 2025 (Feb 28th - Mar 4th). The workshop
will take place in person at Tucson, Arizona! We've got a great
lineup of speakers.
** Background **
The rise of big data and human-centered technologies has brought
exciting advancements and challenges, such as data annotation,
multimodal fusion, and enhancing human-computer interaction. Wearable
eye-tracking devices like Meta Quest 3 and Apple Vision Pro promise to
revolutionize the field by enabling eye-gaze data collection in
real-world settings, offering new ways to study human cognition and
develop gaze-aware ML models.
Eye gaze is a cost-effective way to gather physiological data,
revealing attentional patterns in various domains like radiology,
marketing, and UX. Recently, it's been used for data labeling and
analysis in computer vision, with growing interest in using gaze as a
cognitive signal to train models. Key challenges remain, including
data quality and decoding, but advancements in eye-tracking are
opening new possibilities for egocentric perception, embodied AI, and
multimodality. This workshop aims to bring together experts to address
core issues in gaze-assisted computer vision.
** Call for Papers **
We invite submissions to the "Gaze meets Computer Vision (GMCV):
Bridging Human Attention and Machine Perception" workshop at WACV
2025. The workshop seeks original research contributions, as well as
surveys and position papers, that focus on the integration of gaze
data with computer vision tasks. We welcome papers addressing a broad
range of topics, including but not limited to:
Gaze-Informed Visual Understanding
Gaze-based Human-AI Interaction
Attention Modeling in Vision Systems
Gaze-Driven Annotation and Labeling
Egocentric Vision and Embodied AI
Gaze Enhanced Medical Imaging
Understanding Human intention and Goal inference
Eye-tracking in Visual Search and Navigation
Explainable AI and Trustworthy Vision Systems
Ethical Considerations of using eye-tracking data
Gaze Data Quality and Integration
State-of-the-art method integrating Gaze in ML
Gaze applications in cognitive psychology, radiology, neuroscience, AR/VR, autonomous cars, privacy, etc.
Gaze-Driven Behavioral Analytics
Cross-Modal Learning with Gaze and Vision
Real-Time Gaze Prediction and Analysis
Gaze-Guided Object Detection and Recognition
?Learning from Noisy Gaze Data
?Temporal Dynamics of Gaze in Video Analysis
Privacy-Preserving Gaze Analysis
Gaze in Low-Light and Challenging Environments
Personalization in Vision Systems via Gaze Data
Other Applications of Gaze and Computer Vision
Submission Tracks:
We are accepting submissions for two distinct tracks: Full Paper Track
and Extended Abstract Track. Both offer unique opportunities to
showcase your work at the workshop.
Full Paper Track (Archival). This track is for original research
contributions that will be published in the conference proceedings
and included in IEEE Xplore. Full papers in this track undergo
rigorous peer review and are indexed separately from the main
conference proceedings, ensuring visibility and recognition in the
field.
Page Limit: Up to 8 pages (excluding references and appendices)
Review Process: Double-blind peer review
Publication: IEEE Xplore, archival indexing
Extended Abstract Track (Non-Archival). This track is for
late-breaking research, and preliminary results, or if you wish to
present previously published work. Submissions in this track will
also undergo double-blind peer review, without committing your
work to archival publication. This means that presenting at Gaze
Meets ML does not preclude future submissions to other journals or
conferences.
Page Limit: Up to 4 pages (excluding references and appendices)
Review Process: Double-blind peer review
Publication: Non-archival, no restrictions on future publication
Submission Guidelines:
Formatting: All submissions must adhere to the WACV template and
guidelines.
References & Appendices: Include references and any appendices
within the same PDF document. These sections are excluded from the
page count limit.
Review Process: All submissions, regardless of track, will undergo
a double-blind peer review to ensure quality and fairness.
** Awards and Funding **
We are offering two GP3 SD UX eye-tracking devices from Gazepoint as
Best Paper Awards and, depending on funding availability, we will
cover the registration fees for presenting authors, with a focus on
supporting underrepresented minorities.
** Important dates for Workshop paper submission **
Paper submission deadline: November 22nd, 2024 November 30th, 2024
Notification of acceptance: December 18th, 2024
Camera-ready: January 10th, 2025
Workshop: (Feb 28th or Mar 4th)
** Organizing Committee **
Dario Zanca (FAU Erlangen-Nürnberg)
Ismini Lourentzou (Illinois Urbana-Champaign)
Joy Tzung-yu Wu (Stanford)
Bertram Emil SHI (HKUST)
Elizabeth Krupinski (Emory School of Medicine)
Jimin Pi (Google)
Alexandros Karargyris (MLCommons)
Amarachi Mbakwe (Virginia Tech)
Satyananda Kashyap (IBM)
Abhishek Sharma (Google)
** Contact **
All inquiries should be sent to dario.zanca@fau.de or akarargyris@gmail.com