CALL FOR PARTICIPATION in the
2011 TREC VIDEO RETRIEVAL EVALUATION (TRECVID 2011)
February 2011 - December 2011
Conducted by the National Institute of Standards and Technology (NIST)
With support from other US government agencies
Please note that due to the timing of the ACM Multimedia Conference
TRECVID will occur later than usual, namely 5-7. December.
I n t r o d u c t i o n:
The TREC Video Retrieval Evaluation series (trecvid.nist.gov) promotes
progress in content-based analysis of and retrieval from digital video
via open, metrics-based evaluation. TRECVID is a laboratory-style
evaluation that attempts to model real world situations or significant
component tasks involved in such situations.
D a t a:
In TRECVID 2011 NIST will use the following data sets:
* IACC.1.B
Approximately 8000 Internet Archive videos (50GB, 200 hours)
with Creative Commons licenses in MPEG-4/H.264 with duration
between 10 seconds and 3.5 minutes. Most videos will have some
metadata provided by the donor available e.g., title, keywords,
and description
* IACC.1.A
Approximately 8000 Internet Archive videos (50GB, 200 hours)
with Creative Commons licenses in MPEG-4/H.264 with duration
between 10 seconds and 3.5 minutes. Most videos will have some
metadata provided by the donor available e.g., title, keywords,
and description
* IACC.1.tv10.training
Approximately 3200 Internet Archive videos (??GB, 200 hours)
with Creative Commons licenses in MPEG-4/H.264 with durations
between 3.6 and 4.1 minutes. Most videos will have some metadata
provided by the donor available e.g., title, keywords, and
description
* BBC rushes
Approximately 460 hours divided into 30-second clips (files) of
unedited video (MPEG-1) provided by the BBC Archive. Rushes are
the raw material for programming and in the case of this data
include mostly travel video and material from several dramatic
series.
* Gatwick surveillance video
The data consist of about 150 hours obtained from Gatwick
Airport surveillance video data (courtesy of the UK Home
Office). The Linguistic Data Consortium has provided event
annotations for the entire corpus. The corpus was divided into
development and evaluation subsets. Annotations for 2008
development and test sets are available.
* HAVIC
HAVIC is designed to be a large new collection of Internet
multimedia. Construction by the Linguistic Data Consortium and
NIST began in 2010.
T a s k s:
In TRECVID 2011 NIST will evaluate systems on the following tasks
using the [data] indicated:
* Known-item search task (interactive, manual, automatic) [IACC.1]
The known-item search task models the situation in which someone
knows of a video, has seen it before, believes it is contained
in a collection, but doesn't know where to look. To begin the
search process, the searcher formulates a text-only description,
which captures what the searcher remembers about the target
video. 300 topics are planned for automatic systems, a
subset of 24 for human-in-the-loop systems.
* Semantic indexing [IACC.1]
Automatic assignment of semantic tags to video segments can be
fundamental technology for filtering, categorization, browsing,
search, and other video exploitation. New technical issues to be
addressed include methods needed/possible as collection size and
diversity increase, when the number of features increases, and
when features are related by an ontology.
* Content-based multimedia copy detection [IACC.1]
As used here, a copy is a segment of video derived from another
video, usually by means of various transformations such as
addition, deletion, modification (of aspect, color, contrast,
encoding, ...), camcording, etc. Detecting copies is important
for copyright control, business intelligence and advertisement
tracking, law enforcement investigations, etc. Content-based
copy detection offers an alternative to watermarking.
* Surveillance video event detection [Gatwick]
Detecting human behaviors efficiently in vast amounts
surveillance video, both retrospectively and in realtime, is
fundamental technology for a variety of higher-level
applications of critical importance to public safety and
security. In light of results for 2009/10, in 2011 we will rerun
the 2009 task/data using the 2009 ground truth but on a subset
of the 2009 events.
* Instance search [BBC rushes] (interactive, automatic)
An important need in many situations involving video collections
(archive video search/reuse, personal video organization/search,
surveillance, law enforcement, protection of brand/logo use) is
to find more video segments of a certain specific person,
object, or place, given a visual example.
In 2011 this will still be a pilot task - evaluated by NIST but
intended mainly to explore task definition and evaluation issues
using data and an evaluation framework in hand - in a first
approximation to the desired full task using a smaller number of
topics, a simpler identification of the target entity, and less
accuracy in locating the instance than would be desirable in a
full evaluation of the task.
* Multimedia event detection [HAVIC]
Exploding multimedia content in the Internet necessitates
development of new technologies for content understanding and
search for a wide variety of commerce, research, and government
applications.
Much like TREC, TRECVID will provide, in addition to the data, uniform
scoring procedures, and a forum for organizations interested in
comparing their approaches and results.
Participants will be encouraged to share resources and intermediate
system outputs to lower entry barriers and enable analysis of various
components' contributions and interactions.
*********************************************
* You are invited to participate in TRECVID 2011 *
*********************************************
The evaluation is defined by the Guidelines. A draft version is
available: http://www-nlpir.nist.gov/projects/tv2011/tv2011.html and
details will be worked out starting in February based in part on input
from the participants.
You should read the guidelines carefully before applying to
participate.
Organizations may choose to participate in one or more of the tasks.
TRECVID participants must submit results for at least one task in
order to attend the TRECVID workshop in Gaithersburg in November.
*PLEASE* only apply if you are able and fully intend to complete the
work for at least one task. Taking the data but not submitting any
runs threatens the continued operation of the workshop and the
availability of data for the entire community.
P l e a s e n o t e:
1) Dissemination of TRECVID work and results other than in the
(publicly available) conference proceedings is welcomed, but the
conditions of participation specifically preclude any advertising
claims based on TRECVID results.
2) All system results submitted to NIST are published in the
Proceedings and on the public portions of TRECVID web site archive.
3) The workshop is open only to participating groups that submit
results for at least one task and to selected government personnel
from sponsoring agencies and data donors.
4) By applying to participate you indicate your acceptance of the
above restrictions.
T e n t a t i v e s c h e d u l e
There is a tentative schedule for the tasks included in the guidelines
webpage, which may be changed as part of defining the final guidelines.
Please note that due to the timing of the ACM Multimedia Conference
TRECVID will occur later than usual, namely 5-7. December.
Here is a snapshot of that schedule:
*************************************************
FILL IN SCHEDULE WHEN MED AND SED DATES ARE KNOWN
*************************************************
*************************************************
FILL IN SCHEDULE WHEN MED AND SED DATES ARE KNOWN
*************************************************
*************************************************
FILL IN SCHEDULE WHEN MED AND SED DATES ARE KNOWN
*************************************************
W o r k s h o p f o r m a t
The 2 1/2 day workshop itself, December 5-7 at NIST in Gaithersburg,
Maryland near Washington,DC, will be used as a forum both for
presentation of results (including failure analyses and system
comparisons), and for more lengthy system presentations describing
retrieval techniques used, experiments run using the data, and other
issues of interest to researchers in information retrieval. As there
is a limited amount of time for these presentations, the evaluation
coordinators and NIST will determine which groups are asked to speak
and which groups will present in a poster session. Groups that are
interested in having a speaking slot during the workshop will be asked
to submit a short abstract before the workshop describing the
experiments they performed. Speakers will be selected based on these
abstracts.
As some organizations may not wish to describe their proprietary
algorithms, TRECVID defines two categories of participation:
*Category A: Full participation*
Participants will be expected to present full details of system
algorithms and various experiments run using the data, either in a talk
or in a poster session.
*Category C: Evaluation only* Participants in this category will be
expected to submit results for common scoring and tabulation. They
will not be expected to describe their systems in detail, but will be
expected to provide a general description and report on time and
effort statistics in a notebook paper.
H o w t o r e s p o n d t o t h i s c a l l
Organizations wishing to participate in TRECVID 2011 should respond
to this call for participation by submitting an application by
18. February. Only ONE APPLICATION PER TEAM please, regardless of how
many organizations the team comprises.
An application consists of an email to Lori.Buckland at nist.gov
with the following parts. Please send the application as part of
the body of an email - in plain ASCII text.
1) Name of the TRECVID 2011 main contact person
2) Mailing address of main contact person (no post office box, please)
3) Phone for main contact person
4) Fax for main contact person
5) Complete (unique) team name (if you know you are one of multiple
groups from one organization, PLEASE consult with your colleagues
to make your name unique)
6) Short (unique) team name (20 chars or less) that you will use
to identify yourself in ALL email to NIST
7) Optional - names and email addresses of additional team members you
would like added to the tv11.list mailing list.
8) What years, if any, has your team participating in TRECVID before?
9) A one paragraph description of your technical approaches
10) A list of tasks you plan to participate in:
KIS Known-item search task
SIN Semantic indexing
CCD Content-based multimedia copy detection
SED Surveillance video event detection
INS Instance search
MED Multimedia event detection
11) Participation category:
Category A: Full participation - Participants will be expected to
present full details of system algorithms and various experiments
run using the data, either in a talk or in a poster session.
Category C: Evaluation only - Participants in this category will
be expected to submit results for common scoring and tabulation.
They will not be expected to describe their systems in detail,
but will be expected to provide a general description and report
on time and effort statistics in a notebook paper.
Once you have applied, you'll be subscribed to the tv11.list email
discussion list, can participate in finalizing the guidelines, and
sign up to get the data. The tv11.list email discussion list will
serve as the main forum for such discussion and for dissemination of
other information about TRECVID 2011. It accepts postings only from
the email addresses used to subscribe to it.
All applications must be submitted by *February 18, 2011* to
Lori.Buckland at nist.gov. Any administrative questions about
conference participation, application format, content, etc. should
be sent to the same address.
If you would like to contribute to TRECVID in one or both of the
following ways, please contact Paul Over (info at bottom of page)
directly as soon as possible:
- agree to host 2011 test video data for download by other
participants on a fast, password-protected site. (Asian
and European sites especially needed)
- agree to provide the output of your automatic speech recognition
system run on the IACC test/development video (at least for the
English speech)
Best regards,
Paul Over
Alan Smeaton
Wessel Kraaij
Georges Quenot
Last
updated: