See the MediaEval 2025 webpage for information on how to register and participate.
The goal of this task is to study the long-term memory performance when recognizing small movie excerpts or commercial videos. We provide the videos, precomputed features or EEG features for the challenges proposed in the task such as How memorable a video, if a person familiar with a video or if you can predict the brand memorability?
Subtask 1: Movie Memorability. This task studies the long-term memory performance when recognizing small movie excerpts from weeks to years after having viewed them.
Subtask 2: Commercial/Ad Memorability. This task evaluates long-term memory performance in recognizing commercial videos. Participants will use the VIDEM dataset, which contains commercial videos along with their memorability and brand memorability scores, to train their systems. The trained models will then predict the scores for new, unseen commercial videos (product, brand, and concept presentations and discussions). This challenge does not include EEG data.
In an era where visual content, such as movies and commercials, permeates our daily lives, understanding and predicting the memorability of multimedia content is becoming increasingly important. For marketers, filmmakers, and content creators, selecting and designing media that effectively captures attention and leaves a lasting impression is crucial for success. Commercials, in particular, need to engage viewers immediately and remain memorable to drive brand recognition and influence consumer behavior. However, the potential applications of memorability prediction extend beyond commercial and advertising sectors.
This task aims to develop models that predict the memorability of multimedia content by leveraging various content features. While the results can directly benefit professionals in advertising and film, the insights gained can also be applied to other fields, such as education, content retrieval, and beyond. For instance, educators can use memorability predictions to create more engaging learning materials, while content retrieval systems can enhance search and recommendation accuracy by prioritizing content with higher memorability potential.
This year’s task extends the state of the art by focusing on the memorability of multimedia content within the specific domains of movies and commercials. While previous research has explored the general memorability of videos and images, there has been limited focus on how this concept applies to the nuanced structure of films and advertisements. By addressing this gap, we aim to deepen our understanding of how human cognition interacts with multimedia, providing valuable insights into what makes content memorable and how it can be optimized for various applications across different industries, including both commercial and non-commercial use cases.
New for 2025. In 2025, the MediaEval Media Memorability Task introduces two new datasets: the Movie Memorability dataset and the VIDEM dataset. These additions offer exciting opportunities for participants to explore the memorability of movie excerpts and commercial videos across various real-world contexts. This year, the task continues to build on past efforts by integrating multimodal data, including video content, memorability scores, and EEG data collected during memorability experiments, while encouraging innovative approaches to improve prediction accuracy. Additionally, a new challenge is introduced, focusing on brand memorability prediction. In this challenge, participants are not tasked with predicting the memorability of videos but with predicting a brand memorability score for commercial videos. This new challenge seeks to deepen our understanding of how brands are remembered within multimedia content, adding an intriguing layer of complexity to the task.
Researchers interested in this task include those working in areas such as human perception, multimedia content analysis, cognitive science, and machine learning, particularly in the fields of image and video analysis, memorability, emotional response to media, aesthetics, and multimedia affective computing (though not limited to).
This includes scholars focused on predictive modeling, user experience, and the cognitive impact of media, with a specific interest in movies, commercials, and educational content. Signal processing researchers can also bring valuable insights to this task by leveraging EEG signals to enhance the memorability predictive models. Additionally, researchers exploring content retrieval, recommendation systems, and multimedia interaction, as well as those studying the influence of media on memory and learning, will find the task valuable. It will also appeal to those working on improving machine learning algorithms for content classification and understanding, especially in video and image domains, and those interested in applying these models across both commercial and non-commercial media, including educational and informational content.
One dataset will be provided for each subtask.
For subtask 1, a subset of the Movie Memorability dataset will be used. This is a collection of movie excerpts and corresponding ground-truth files based on the measurement of long-term memory performance when recognizing small movie excerpts from weeks to years after having viewed them. It is accompanied with audio and video features extracted from the movie excerpts. EEG data recorded while viewing this subset will be also provided. EEG data were recorded while 27 participants viewed a subset of clips from the dataset. The clips were selected to include both previously seen and unseen movies. After viewing each clip, participants were asked if they remembered seeing it before. In total 3484 epochs of 64 channel EEG data are available, of which 2122 were not recognised and 1362 were remembered.
For subtask 2, the VIDEM (VIDeo Effectiveness and Memorability) dataset will be used. It focuses on video and brand memorability in commercial advertisements, including some educational or explanatory videos. Developed through a university-business collaboration between the University of Essex and Hub, with support from Innovate UK’s Knowledge Transfer Partnership (grant agreement No. 11071), This is a collection of commercial advertisements and corresponding ground-truth files based on the measurement of long-term memory performance when recognizing them from 24 to 72 hours after having viewed them. Each video is accompanied with metadata such as titles, descriptions, number of views, and duration and audio and video features extracted from the commercial advertisements. The dataset consists of 424 commercial videos sampled from a larger collection of 4791 videos published on YouTube between June 2018 and June 2021. Video lengths range from 7 seconds to 94 minutes. For longer videos, users are allowed to watch only 1 minute.
Submissions for the video-based prediction challenges will be evaluated using Spearman’s rank correlation coefficient. Additional metrics, such as Mean Squared Error (MSE), may also be used to assess prediction accuracy. For Challenge 1.2 (EEG-based detection of recall), submissions will be evaluated based on accuracy.
Here are several research questions related to this challenge that participants can strive to answer in order to go beyond just looking at the evaluation metrics:
More details will follow.
[1] 2018 R.Cohendet, K. Yadati, N. Q. Duong and C.-H. Demarty. Annotating, understanding, and predicting long-term video memorability. In Proceedings of the ICMR 2018 Conference, Yokohama, Japan, June 11-14, 2018.
[2] 2014. Phillip Isola, Jianxiong Xiao, Devi Parikh, Antonio Torralba, and Aude Oliva. What makes a photograph memorable? IEEE Transactions on Pattern Analysis and Machine Intelligence 36, 7 (2014), 1469–1482.
[3] 2023. Dumont, T., Hevia, J. S., & Fosco, C. L. Modular memorability. Tiered representations for video memorability prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10751-10760).
[4] 2025. Kumar, P. et al. Eye vs. AI: Human Gaze and Model Attention in Video Memorability. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Tucson, AR, USA, 2025 (In press).
[5] 2025. SI, H.et al. Long-Term Memorability On Advertisements. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Tucson, AR, USA, 2025 (In press).
The program will be updated with the exact dates.
More details will follow.