Call for participation for Workshop on Human and Computer Models of Video Understanding

 

Workshop on 15 May 2024, University of Surrey.

 

Dear Colleagues,

 

We invite participants to a workshop on “Human and Computer Models of Video Understanding”. We aim to bring together human and computer vision scientists to share the latest knowledge and collaborate.

 

Participants can submit a 500 word (approx) abstract describing the work they would like to present and their preference for an oral or poster presentation. Given time constraints, accommodating all desired oral presentations might not be possible, so we will review abstract submissions and assign some to oral presentation slots and some to posters.

 

Please visit the workshop’s website to submit an abstract and/or register:

https://www.ias.surrey.ac.uk/event/human-and-computer-models-of-video-understanding/

 

It is free to attend but there will be an optional evening dinner at an additional cost (details and link for payment to be sent nearer the date).

 

Key Dates:

Abstract submission deadline: 29 March

Notification: 5 April

Workshop date: 15 May 2024, University of Surrey.

 

Invited Speakers:

Professor Shaogang Gong, Queen Mary University of London, Queen Mary Computer Vision Laboratory

Professor Frank Pollick, University of Glasgow, School of Psychology and Neuroscience

 

 

Call for Research Contributions

 

We invite participants for a multi-disciplinary workshop on “Human and Computer Models of Video Understanding”. The core research question we are concerned with is: How does the human brain understand people's activities in a video much better than existing computer systems? We invite participants from the science of human vision (psychology or brain sciences) and computer vision, focusing on understanding activities from video.

 

To give some concrete examples: Humans can very quickly make accurate judgements about the activity happening in a video even if the quality of the video is poor, or the motions observed are ambiguous, for example, to discriminate hugging from fighting, or smoking from eating finger food. Computers cannot match human performance in these tasks, which are critical for applications in surveillance, monitoring safety and welfare in a care setting, or removing inappropriate videos from social media. We do not yet fully understand how humans perform these feats, nor how to make computer vision systems reach their performance.

 

We invite human and computer vision science participants to present the latest advances in their fields, in a language accessible to a multi-disciplinary audience. We intend to foster a cross-fertilisation of ideas between the different scientific communities, where each can see ways to incorporate insights and techniques from a foreign field in their models. Moreover we intend to act as a “matchmaker” for new cross-disciplinary partnerships on projects that can incorporate the techniques of separate communities. We plan a future special issue of the journal Transactions in Cognitive and Developmental Systems for the topic “Special issue on Vision Sciences for Video Understanding in Cognitive Systems”, where multi-disciplinary teams emerging from this workshop can publish a paper in a presentation accessible to a multi-disciplinary audience.

 

We welcome participation from both junior and senior researchers from academia and industry.

 

Potential topics include, but are not limited to:

Eye Tracking in Video,

Visual Attention and Salient Features in Video,

Cognitive Models for Video Understanding,

Perceptual Quality in Video Streaming,

Multimodal Perception in Videos,

Emotion and Affect in Video Viewing,

Neuroscience and Brain Imaging in Video Perception,

Visual Cognition in Virtual Reality (VR) and Augmented Reality (AR),

Attentional Shifts and Change Detection in Video,

Video Activity Recognition,

Video Object Detection and Tracking,

Video Segmentation and Scene Understanding,

Event Detection and Recognition in Videos,

Video-Based Surveillance and Security,

Spatiotemporal Action Localization,

Human Vision-Inspired Video Models

 

 

ORGANISERS

Dr Frank Guerin, University of Surrey

Dr Andrew Gilbert, University of Surrey

Dr Quoc Vuong, Newcastle University

 

 

 

===================================

Biosciences Institute

Newcastle University

Newcastle upon Tyne, UK, NE2 4HH

+44 (0)191 208 6183

https://www.staff.ncl.ac.uk/q.c.vuong/

===================================