Skip to content

Commit 914c15d

Browse files
Merge pull request #177 from lucienheitz/patch-1
Update newsimages.md
2 parents a4286e6 + 45f430f commit 914c15d

File tree

1 file changed

+62
-29
lines changed

1 file changed

+62
-29
lines changed

_editions/2025/tasks/newsimages.md

Lines changed: 62 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -15,49 +15,82 @@ blurb: "Participants receive a large set of articles (including the headline and
1515

1616
#### Task description
1717

18+
This year’s NewsImages challenge explores matching news articles with retrieved or generated images. Participants receive a large set of English-language articles from international publishers. Given the text of a news article, the goal of this task is to retrieve and/or generate a fitting image recommendation. In the **Image Retrieval** subtask, teams retrieve images from a larger collection. In the **Image generation** subtask, the teams use generative AI to produce an image that can be used as a thumbnail. (Teams can take part in one or both of two subtasks; separate submissions are required.) The tasks are as follows:
19+
20+
1. **Image retrieval**, where participants design and implement approaches to retrieve relevant images in an existing image pool that fits a given news headline and lead.
21+
2. **Image generation**, where participants need to use and develop techniques for creating appropriate visuals for news articles.
22+
23+
The main criterion for evaluation (i.e., ranking of teams) is the image fit and relevance, i.e., whether the images capture key attributes of the text article, but do not depict any important elements not present in the article. The image retrieval/generation pipeline must not rely on any third-party closed-sourced APIs or resources. We provide a comprehensive dataset of news articles with article title, article lead, URL (to retrieve the full text), and an editorially assigned image.
24+
25+
In the final phase of this challenge, participants will take part in a crowdsourced evaluation event to rate the quality and relevance of the submissions. We provide an online interface where teams will be shown images produced by their peers. Each team will need to rate a subset of submitted images.
26+
27+
Furthermore, the images will be evaluated for their media policy compliance. We are particularly interested in non-photorealistic images that do not suggest that they accurately represent real events so as to not mislead and/or deceive readers.
28+
1829
#### Motivation and background
30+
News publishers and recommender systems depend on images and thumbnails to engage readers with news articles. Technology has advanced to the point that it is possible to automatically find a matching image (image retrieval) or generate an appropriate image (generative AI) for a news article. Although these techniques present an opportunity for the news media, they also present a great number of technical and ethical challenges. It is of critical importance that the match between the news and the image is accurate. And images should not mislead and/or deceive readers into assuming it represents a real-life situation when it does not.
31+
32+
Online news articles are inherently multimodal. The text of an article is often accompanied by an image and/or other multimedia items. This image, however, is not only important for illustrating and complementing the text of news articles. It plays a critical role in capturing the readers’ attention; it is often the first thing readers see when browsing a news platform.
33+
34+
Research in multimedia and recommender systems generally assumes a simple relationship between images and text occurring together. For example, in image captioning [1], the caption is often assumed to describe the literally depicted content of the image. In contrast, when images accompany news articles, the relationship becomes less clear [2]. Since there are often no images available for the most recent news messages, stock images, archived photos, or even generated images are used. Here, preliminary studies showed that users prefer AI-generated content over stock images [3, 4]. The goal of this task is to investigate these intricacies in more depth, in order to understand the implications that it may have for the areas of journalism and news personalization.
1935

2036
#### Target group
37+
This task targets researchers who are interested in investigating the retrieval and generation of images for news and the connection between images and text. This includes people working in the areas of computer vision, recommender systems, cross-modal information retrieval, as well as in the area of news analysis.
2138

2239
#### Data
23-
24-
#### Ground truth
40+
This challenge uses the open-source [GDELT](https://www.gdeltproject.org/) news dataset. We create and share a subset of 8,500 news articles collected during 2022 and 2023. The news articles are all in English. Each item includes the article title, article lead, and the original image. The article text itself is not shared, but participants are free to retrieve it from the original source. We ask participants to use the [Yahoo-Flickr Creative Commons 100 Million (YFCC100M)](https://www.multimediacommons.org/) dataset to source the images for the retrieval task.
2541

2642
#### Evaluation methodology
43+
The generated and retrieved image submissions of the participants will be evaluated during a crowd-sourced event. To that end, the organizers provide an online tool and distribute accounts to participants. The participants are then requested to rate the quality of submitted images of their peers. The winning team is determined by the overall average rating for submitted images. In addition to the qualitative questions outlined above, the rating procedure for generated images will include a check for policy compliance. The policy applied within the scope of this challenge is based on existing AI guidelines in news media: Generated image submissions should adhere to current editorial standards and not suggest that they accurately represent real events, deceive, or mislead readers in any other way.
44+
45+
Two subsets of image submissions will be rated during the evaluation event. We have a small subset of predetermined articles (will be communicated in advance by the organizers) and a larger subset of randomly picked articles. This allows the participants to focus on creating hand-crafted solutions for the predetermined article-image pairs as well as fully automated and scalable solutions for the random sample.
46+
47+
Staying true to the principles of MediaEval of promoting reproducible research, the submissions of both retrieved and generated images must include the entire workflow. For image generation, we encourage participants to use tools that can automatically embed the workflow used for their generation in the image file, e.g., [ComfyUI](https://github.com/comfyanonymous/ComfyUI) or [WebUI](https://github.com/AUTOMATIC1111/stable-diffusion-webui)). Furthermore, we ask participants to use open-source solutions that can be set up and run locally.
2748

2849
#### Quest for insight
29-
Here are several research questions related to this challenge that participants can strive to answer in order to go beyond just looking at the evaluation metrics:
30-
* <!-- # First research question-->
31-
* <!-- # Second research question-->
32-
<!-- # and so on-->
50+
This task invites participants to further reflect on the role and potential of news images in terms of generation and retrieval. The organizers encourage participants to explore further topics related to the task. Possible research questions are:
51+
52+
* Are generated images better suited than real archival images?
53+
* What parameters are most important for generating perfectly matching images?
54+
* Which categories of news are best suited for image generation/retrieval?
55+
* What do users expect from images embedded in news articles?
56+
* Do news readers care more about generated or non-generated images?
57+
* Are photographs or illustrations more effective in capturing the readers’ attention?
58+
* What are existing policies and guidelines for retrieving/generating images and are they in line with readers' expectations?
59+
* What is the impact of item properties (e.g., image style or selection of subject) across the dimension of:
60+
* Trustworthiness: Does the image look like it originates from a trustworthy source?
61+
* Promoting stereotypes or Biases: Does the image promote any stereotypes or biases?
62+
* Attractiveness/clickbaitness: Does the image capture the reader’s attention?
63+
* Visual quality: Does the image look professional and does not include any flaws?
3364

3465
#### Participant information
35-
<!-- Please contact your task organizers with any questions on these points. -->
36-
<!-- # * Signing up: Fill in the [registration form]() and fill out and return the [usage agreement](). -->
37-
<!-- # * Making your submission: To be announced (check the task read me) <!-- Please add instructions on how to create and submit runs to your task replacing "To be announced." -->
38-
<!-- # * Preparing your working notes paper: Instructions on preparing you working notes paper can be found in [MediaEval 2023 Working Notes Paper Instructions]().-->
66+
* You are requested to release your code and/or workflow to reproduce your submission.
67+
* Signing up: Fill in the [registration form](https://forms.gle/FFS8V3zbijXoqEWJ6) and fill out and return the [usage agreement](https://multimediaeval.github.io/editions/2023/docs/MediaEval2023_UsageAgreement.pdf).
68+
* Making your submission: Details to be announced together with the dataset release.
69+
* Evaluate event: Details to be announced once all teams handed in their runs.
70+
* Preparing your working notes paper: Instructions on preparing you working notes paper can be found in [MediaEval 2023 Working Notes Paper Instructions](https://docs.google.com/document/d/1HcAx14RVuxqDEi-1SJJRwhHhzC_V-Ktpw-9jn5dg0-0).
71+
* Please contact your task organizers with any questions on these points.
3972

4073
#### References and recommended reading
41-
<!-- # Please use the ACM format for references https://www.acm.org/publications/authors/reference-formatting (but no DOI needed)-->
42-
<!-- # The paper title should be a hyperlink leading to the paper online-->
74+
[1] Hossain, M. Z., Sohel, F., Shiratuddin, M. F., & Laga, H. (2019). [A comprehensive survey of deep learning for image captioning](https://dl.acm.org/doi/abs/10.1145/3295748). ACM Computing Surveys (CSUR), 51(6), p. 1-36.
4375

44-
#### Task organizers
45-
* <!-- # First organizer-->
46-
* <!-- # Second organizer-->
47-
<!-- # and so on-->
76+
[2] Oostdijk, N., van Halteren, H., Bașar, E., & Larson, M. (2020). [The Connection between the Text and Images of News Articles: New Insights for Multimedia Analysis](https://www.aclweb.org/anthology/2020.lrec-1.535/). In Proceedings of The 12th Language Resources and Evaluation Conference, p. 4343-4351.
4877

49-
#### Task auxiliaries
50-
<!-- # optional, delete if not used-->
51-
* <!-- # First auxiliary-->
52-
* <!-- # Second auxiliary-->
53-
<!-- # and so on-->
78+
[3] Heitz, L., Bernstein, A. & Rossetto, L. (2024). [An Empirical Exploration of Perceived Similarity between News Article Texts and Images](https://ceur-ws.org/Vol-3658/paper8.pdf). MediaEval 2023 Working Notes Proceedings.
79+
80+
[4] Heitz, L., Chan, Y. K., Li, H., Zeng, K., Bernstein, A. & Rossetto, L. (2024). [Prompt-based Alignment of Headlines and Images Using OpenCLIP](https://ceur-ws.org/Vol-3658/paper7.pdf). MediaEval 2023 Working Notes Proceedings.
81+
82+
#### Task organizers
83+
* Lucien Heitz, University of Zurich, Switzerland, [email protected]
84+
* Luca Rossetto, Dublin City University, Ireland
85+
* Benjamin Kille, NTNU, Trondheim, Norway
86+
* Andreas Lommatzsch, TU Berlin, Germany
87+
* Duc-Tien Dang-Nguyen, University of Bergen, Norway
88+
* Mehdi Elahi, University of Bergen, Norway
5489

5590
#### Task schedule
56-
* XX May 2025: Development Data release <!-- * XX May 2025: Data release <!-- # Replace XX with your date. We suggest setting the date in May - of course if you want to realease sooner it's OK. -->
57-
* XX June 2025: Development Data release <!-- * XX June 2025: Data release <!-- # Replace XX with your date. We suggest setting the date in June - of course if you want to realease sooner it's OK. -->
58-
* XX September 2025: Runs due and results returned. Exact dates to be announced. <!--* XX September 2025: Runs due <!-- # Replace XX with your date. We suggest setting enough time in order to have enough time to assess and return the results by the Results returned.-->
59-
* 08 October 2025: Working notes paper <!-- Fixed. Please do not change.-->
60-
* 25-26 October 2025: MediaEval Workshop, Dublin, Ireland and Online.. <!-- Fixed. Please do not change.-->
61-
62-
#### Acknowledgements
63-
<!-- # optional, delete if not used-->
91+
* 30 April 2025: News dataset released.
92+
* 10 September 2025: Deadline for team submissions.
93+
* 17-24 September 2025: Online evaluation event.
94+
* 30 September 2025: Return of the evaluation results.
95+
* 08 October 2025: Deadline for submitting working notes paper (incl. code/pipeline).
96+
* 25-26 October 2025: MediaEval Workshop, Dublin, Ireland and Online.

0 commit comments

Comments
 (0)