Call for Participation in TRECVID 2019
CALL FOR PARTICIPATION in the
2019 TREC VIDEO RETRIEVAL EVALUATION (TRECVID 2019)
February 2019 - November 2019
Conducted by the National Institute of Standards and Technology (NIST)
with additional funding from other US government agencies.
I n t r o d u c t i o n:
The TREC Video Retrieval Evaluation series (trecvid.nist.gov) promotes
progress in content-based analysis of and retrieval from digital video
via open, metrics-based evaluation. TRECVID is a laboratory-style
evaluation that attempts to model real world situations or significant
component tasks involved in such situations. In its 19th annual evaluation
cycle TRECVID will evaluate participating systems on 5 different video
analysis and retrieval tasks using various types of real world datasets.
D a t a:
In TRECVID 2019 NIST will use at least the following data sets:
* Vimeo Creative Commons Collection (V3C)
The V3C is a large-scale video dataset that has been collected from high-quality
web videos with a time span over several years in order to represent true videos
in the wild. It consists of 28,450 videos with a duration of 3,801 hours in total.
The first part of this dataset (V3C1) has been used by the Video Browser Showdown
(VBS) 2019 and will be used for the Ad-Hoc Video Search (AVS) task at TRECVID 2019
as well. For both campaigns V3C1 will serve as a basis over three years (VBS 2019-2021
and TRECVID 2019-2021) before it is planned to be extended with further parts of the
V3C dataset. V3C1 contains 1,000 hours of video content and approximately one million
shots that were created by the authors of the dataset using the open-source multimedia
retrieval engine Cineast.
The IACC.3 was introduced in 2016 and consists of approximately 4600 Internet
Archive videos (144 GB, 600 h) with Creative Commons licenses in MPEG-4/H.264
format with duration ranging from 6.5 min to 9.5 min and a mean duration
of almost 7.8 min. Most videos will have some metadata provided by the
donor available e.g., title, keywords, and description.
* BBC EastEnders
Approximately 244 video files (totally 300GB, 464 hours) with
associated metadata, each containing a week's worth of BBC EastEnders
programs in MPEG-4/H.264 format.
* Twitter Vine videos
Approximately 50,000 video clips URLs have been collected by NIST
from the public Twitter stream of Vine videos. Only a subset of at
least 2,000 videos will be used with their human annotated text descriptions.
* Gatwick and i-LIDS MCT airport surveillance video
The data consist of about 150 hours obtained from airport
surveillance video data (courtesy of the UK Home Office). The
Linguistic Data Consortium has provided event annotations for
the entire corpus. The corpus was divided into development and
evaluation subsets. Annotations for 2008 development and test
sets are available.
* VIRAT dataset
The VIRAT Video Dataset is a large-scale surveillance video dataset designed to
assess the performance of activity detection algorithms in realistic scenes.
The dataset was collected to facilitate both detection of activities and to localize
the corresponding spatio-temporal location of objects associated with activities from
a large continuous video. The VIRAT dataset are closely aligned with real-world video
T a s k s:
In TRECVID 2019 NIST will evaluate systems on the following tasks
using the [data] indicated:
* AVS: Ad-hoc Video Search (automatic, manually-assisted, relevance feedback) [V3C1]
A new Ad-hoc search task started in TRECVID 2016 and will continue in 2019
(with new dataset) to model the end user search use-case, who is looking for
segments of video containing persons, objects, activities, locations, etc.,
and combinations of the former. Given about 30 multimedia topics created at
NIST, return for each topic all the shots which meet the video need expressed
by it, ranked in order of confidence. Although all evaluated submissions will be
for automatic runs, Interactive systems will have the opportunity to
participate in the Video Browser Showdown (VBS) in 2020 using
the same testing data (V3C1).
* ActEV: Activities in Extended Video [VIRAT]
ActEV is a series of evaluations to accelerate development of robust, multi-camera,
automatic activity detection algorithms for forensic and real-time alerting applications.
ActEV is an extension of the annual TRECVID Surveillance Event Detection (SED) evaluation
where systems will also detect, and track objects involved in the activities. Each evaluation
will challenge systems with new data, system requirements, and/or new activities.
* INS: Instance search (interactive, automatic) [BBB EastEnders]
An important need in many situations involving video collections
(archive video search/reuse, personal video organization/search,
surveillance, law enforcement, protection of brand/logo use) is to
find more video segments of a certain specific person, object,
or place, given a visual example. In 2016-2018 new query type was tested
by asking systems to retrieve specific persons in specific locations.
A new query type will start in 2019 asking systems to retrieve specific persons
doing specific actions. A set of defined actions with various image/video
examples will be given and each topic will include few examples (image and video)
of a person and ask systems to find that person doing one of the defined actions.
* VTT: Video to Text Description [Twitter Vine videos]
Automatic annotation of videos using natural language text descriptions
has been a long-standing goal of computer vision. The task involves
understanding of many concepts such as objects, actions, scenes, person-object
relations, temporal order of events and many others. In recent years there have
been major advances in computer vision techniques which enabled researchers to
start practically to work on solving such problem.
Given a set of Vine videos URLs and number of reference sets of text descriptions,
systems are asked to work and submit results for two subtasks.The core "Description Generation" subtask requires
systems to automatically generate a text description (1 sentence) for each video URL.
An optional "Matching and Ranking" subtask requires systems to
return for each video URL a ranked list of the most likely text
description that correspond (was annotated) to the video from each of the reference sets.
In addition to the data, TRECVID will provide uniform scoring
procedures, and a forum for organizations interested in comparing
their approaches and results.
Participants will be encouraged to share resources and intermediate
system outputs to lower entry barriers and enable analysis of various
components' contributions and interactions.
* You are invited to participate in TRECVID 2019 *
The evaluation is defined by the Guidelines. A draft version is
available: http://www-nlpir.nist.gov/projects/tv2019/index.html and
further feedback input from the participants are welcomed till April,2019.
You should read the guidelines carefully before applying to participate in one or more tasks:
P l e a s e n o t e:
1) Dissemination of TRECVID work and results other than in the
(publicly available) conference proceedings is welcomed, but the
conditions of participation specifically preclude any advertising claims based on TRECVID results.
2) All system output and results submitted to NIST are published in
the Proceedings or on the public portions of TRECVID web site archive.
3) The workshop is open only to participating groups that submit
results for at least one task and to selected government personnel
from sponsoring agencies and data donors.
4) Each participating group is required to submit before the November
workshop a notebook paper describing their experiments and results.
This is true even for groups who may not be able to attend the
5) It is the responsibility of each team contact to make sure that
information distributed via the call for participation and the
firstname.lastname@example.org email list is disseminated to all team members with
a need to know. This includes information about deadlines and
restrictions on use of data.
6) By applying to participate you indicate your acceptance of the
above conditions and obligations.
There is a tentative schedule for the tasks included in the Guidelines
W o r k s h o p f o r m a t
Plans are for a 2 days workshop at NIST
in Gaithersburg, Maryland - just outside Washington, DC. Confirmation
and details will be provided to participants as soon as available.
The TRECVID workshop is used as a forum both for presentation of
results (including failure analyses and system comparisons), and for
more lengthy system presentations describing retrieval techniques
used, experiments run using the data, and other issues of interest to
researchers in information retrieval. As there is a limited amount of
time for these presentations, the evaluation coordinators and NIST
will determine which groups are asked to speak and which groups will
present in a poster session. Groups that are interested in having a
speaking slot during the workshop will be asked to submit a short
abstract before the workshop describing the experiments they
performed. Speakers will be selected based on these abstracts.
H o w t o r e s p o n d t o t h i s c a l l
Organizations wishing to participate in TRECVID 2019 must respond
to this call for participation by submitting an on-line application by
1 April. Only ONE APPLICATION PER TEAM please, regardless of how
many organizations the team comprises.
*PLEASE* only apply if you are able and fully intend to complete the
work for at least one task. Taking the data but not submitting any
runs threatens the continued operation of the workshop and the
availability of data for the entire community.
Here is the application URL:
You will receive an immediate automatic response when your application
is received. NIST will respond with more detail to all applications submitted
before the end of March. At that point you'll be
given the active participant's userid and password, be subscribed to
the tv19.list email discussion list, and can participate in finalizing
the guidelines as well as sign up to get the data, which is controlled
by separate passwords.
T R E C V I D 2 0 1 9 e m a i l d i s c u s s i o n l i s t
The tv19.list email discussion list (email@example.com) will serve as
the main forum for discussion and for dissemination information about
TRECVID 2019. It is each participant's responsibility to monitor the
tv19.list postings. It accepts postings only from the email addresses
used to subscribe to it. At the bottom of the guidelines there is a
link to an archive of past postings available using the active
Q u e s t i o n s
Any administrative questions about conference participation,
application format/content, subscriptions to the tv19.list,
etc. should be sent to george.awad at nist.gov.
TRECVID 2019 organizers team
updated: [an error occurred while processing this directive]
For further information contact