Skip to main content
U.S. flag

An official website of the United States government

Official websites use .gov
A .gov website belongs to an official government organization in the United States.

Secure .gov websites use HTTPS
A lock ( ) or https:// means you’ve safely connected to the .gov website. Share sensitive information only on official, secure websites.

MED 2016 Evaluation

The Multimedia Event Detection (MED) track is part of the TRECVID Evaluation. The 2016 evaluation will be the seventh MED evaluation which was preceded by the 2011, 2012, 2013, 2014, and 2015 evaluations and the 2010 Pilot evaluation.

The goal of MED is to assemble core detection technologies into a system that can search multimedia recordings for user-defined events based on pre-computed metadata. The metadata stores developed by the systems are expected to be sufficiently general to permit re-use for subsequent user defined events.

A user searching for events in multimedia material may be interested in a wide variety of potential events. Since it is an intractable task to build special purpose detectors for each event a priori, technology is needed that can take as input a human-generated definition of the event that a system will use to search a collection of multimedia clips. The MED evaluation series will define events via an event kit which consists of an event name, definition, explication (textual exposition of the terms and concepts), evidential descriptions, and illustrative video exemplars.

Information Dissemination

NIST maintains an email discussion list to disseminate information. Send requests to join the list to med_poc at nist dot gov.

Evaluation Plan

MED system performance will be evaluated as specified in the evaluation plan, MED16 Evaluation Plan.  Please note that some content (specifically the submission format and instructions) is subject to change as we update our submission/scoring infrastructure to handle the addition of the YFCC100M dataset.

Data Resources

The HAVIC collection of Internet multimedia (i.e., clips containing both audio and video streams) will be provided to registered MED participants. The data, which was collected by the Linguistic Data Consortium, consists of publicly available, user-generated content posted to the various Internet video hosting sites.

The Yahoo Flickr Creative Commons 100M dataset (YFCC100M) is a large collection of images and video available on Yahoo! Flickr.  All photos and videos listed in the collection are licensed under one of the Create Commons copyright licenses.  This dataset is available directly from Yahoo! here.  Only a subset of the YFCC100M videos will be used for this evaluation, this subset is to be determined.

The evaluation plan and license information will specify usage rules of the data resources in full detail.

Video data

HAVIC clips will be provided in MPEG-4 formatted files. The video will be encoded to the H.264 standard. The audio will be encoded using MPEG-4'S Advanced Audio Coding (AAC) standard.  Please refer to the YFCC100M documentation regarding video file format.

Data Licensing

In order to obtain the HAVIC corpora, ALL (including previous participants) TRECVID-registered sites must complete an evaluation license with the LDC. Each site who requires access to the HAVIC data, either as part of a team or as a standalone researcher, must complete a license.
To complete the evaluation license follow these steps:

  1. Download the license MED16 Data License.
  2. Return the completed license to LDC's Membership Office via email at ldc at ldc dot upenn dot edu.  Alternatively, you may fax the completed license to the LDC at 215-573-2175.
  3. When you send the completed license to the LDC, include the following information:
  • Registered TRECVID Team name
  • Site/organization name
  • Data contact person's name
  • Data contact person's email
  • Specifically which Corpora/Data listed in the agreement you need from the LDC, as participants in last years MED evaluation may already have some of the data on site

The designated data contact person for each site will receive instructions from the LDC about the specific procedures for obtaining the HAVIC data packages when they are released.

Input Files

The following archives contain the input files for the Pre-Specified Event, and Ad-Hoc Event portions of the MED16 evaluation respectively.

Dry Run Evaluation

There are two relevant archives for the dry run evaluation:

In addition to the data files, F4DE-3.3.0 contains a scoring primer for MED '16 in the file DEVA/doc/TRECVid-MED16-ScoringPrimer.html that demonstrates how to score a system and prepare a submission file.

Evaluation Tools

Evaluation scripts to support the MED evaluation are within the NIST Framework for Detection Evaluations (F4DE) Toolkit, a link to the latest release can be found on the NIST MIG tools page.
MED16 requires a F4DE version of 3.3.0 or later. The list of F4DE releases can be found on the F4DE GitHub page here.


Consult the TRECVID Master Schedule

Revision History

  • Aug 31, 2016 - Added MED16 Ad-Hoc input files
  • July 26, 2016 - Added MED16 Dry Run information
  • July 1, 2016 - Added MED16 Pre-Specified input files
  • April 1, 2016 - Initial version of evaluation plan added
  • Mar 14, 2016 - Added Data Licensing section
  • Jan 28, 2016 - Initial page created
Created January 28, 2016, Updated November 29, 2019