SemEval-2022 Task 09: R2VQ - Competence-based Multimodal Question Answering

Event Notification Type: 
Call for Participation
Abbreviated Title: 
R2VQ
Contact: 
James Pustejovsky
Roberto Navigli
Jingxuan Tu
Marco Maru
Simone Conia
Kyeongmin Rim
Kelley Lynch
Richard Brutti
Eben Holderness

FIRST CALL FOR PARTICIPATION

SemEval-2022 Task 09: R2VQ - Competence-based Multimodal Question Answering

We invite you to participate in the SemEval-2022 Task 9: Competence-based Multimodal Question Answering (R2VQ).

The task is being held as part of SemEval-2022, and all participating team will be able to publish their system description paper in the proceedings published by ACL.

Codalab (Data download): https://competitions.codalab.org/competitions/34056

Motivation
================================================
When we apply our existing knowledge to new situations, we demonstrate a kind
of understanding of how the knowledge (through tasks) is applied. When viewed
over a conceptual domain, this constitutes a competence. Competence-based
evaluations can be seen as a new approach for designing NLP challenges, in
order to better characterize the underlying operational knowledge that a
system has for a conceptual domain, rather than focusing on individual tasks.
In this shared task, we present a challenge that is reflective of linguistic
and cognitive competencies that humans have when speaking and reasoning.

Task Overview
================================================
Given the intuition that textual and visual information mutually inform each
other for semantic reasoning, we formulate the challenge as a competence-
based question answering (QA) task, designed to involve rich semantic
annotation and aligned text-video objects. The task is structured as question
answering pairs, querying how well a system understands the semantics of
recipes.

We adopt the concept of "question families" as outlined in the CLEVR dataset
(Johnson et al., 2017). While some question families naturally transfer over
from the VQA domain (e.g., integer comparison, counting), other concepts such
as ellipsis and object lifespan must be employed to cover the full extent of
competency within procedural texts.

Data Content
================================================
We have built the R2VQ (Recipe Reading and Video Question Answering) dataset, a dataset consisting of a collection of recipes sourced from https://recipes.fandom.com/wiki/Recipes_Wiki and foodista.com, and labeled according to three distinct annotation layers: (i) Cooking Role Labeling (CRL), (ii) Semantic Role Labeling (SRL), and (iii) aligned image frames taken from creative commons cooking videos downloaded from YouTube. It consists of 1,000 recipes, with 800 to be used as training, and 100 recipes each for validation and testing. Participating systems will be exposed to the aforementioned multimodal training set, and will be asked to provide answers to unseen queries exploiting (i) visual and textual information jointly, or (ii) textual information only.

Task Website and Codalab Submission site: https://competitions.codalab.org/competitions/34056
Mailing List: semeval-2022-task9 [at] googlegroups.com

Important Dates
================================================
Training data available: October 15, 2021

Validation data available: December 3, 2021

Evaluation data ready: December 3, 2021

Evaluation start: January 10, 2021

Evaluation end: January 31, 2022

System Description Paper submissions due: February 23, 2022

Notification to authors: March 31, 2022

Organization
================================================
James Pustejovsky, Brandeis University, jamesp [at] brandeis.edu
Jingxuan Tu, Brandeis University, jxtu [at] brandeis.edu
Marco Maru, Sapienza University of Rome, maru [at] di.uniroma1.it
Simone Conia, Sapienza University of Rome, conia [at] di.uniroma1.it
Roberto Navigli, Sapienza University of Rome, navigli [at] diag.uniroma1.it
Kyeongmin Rim, Brandeis University, krim [at] brandeis.edu
Kelley Lynch, Brandeis University, kmlynch [at] brandeis.edu
Richard Brutti, Brandeis University, richardbrutti [at] brandeis.edu
Eben Holderness, Brandeis University, egh [at] brandeis.edu
Please contact the organizers at semeval-2022-task9 [at] googlegroups.com, or post questions at the Forum page in Codalab.