The Multimedia Event Detection (MED) track is part of the TRECVID Evaluation. The 2016 evaluation will be the seventh MED evaluation which was preceded by the 2011, 2012, 2013, 2014, and 2015 evaluations and the 2010 Pilot evaluation.
The goal of MED is to assemble core detection technologies into a system that can search multimedia recordings for user-defined events based on pre-computed metadata. The metadata stores developed by the systems are expected to be sufficiently general to permit re-use for subsequent user defined events.
A user searching for events in multimedia material may be interested in a wide variety of potential events. Since it is an intractable task to build special purpose detectors for each event a priori, technology is needed that can take as input a human-generated definition of the event that a system will use to search a collection of multimedia clips. The MED evaluation series will define events via an event kit which consists of an event name, definition, explication (textual exposition of the terms and concepts), evidential descriptions, and illustrative video exemplars.
NIST maintains an email discussion list to disseminate information. Send requests to join the list to med_poc at nist dot gov.
MED system performance will be evaluated as specified in the evaluation plan, MED16 Evaluation Plan. Please note that some content (specifically the submission format and instructions) is subject to change as we update our submission/scoring infrastructure to handle the addition of the YFCC100M dataset.
The HAVIC collection of Internet multimedia (i.e., clips containing both audio and video streams) will be provided to registered MED participants. The data, which was collected by the Linguistic Data Consortium, consists of publicly available, user-generated content posted to the various Internet video hosting sites.
The Yahoo Flickr Creative Commons 100M dataset (YFCC100M) is a large collection of images and video available on Yahoo! Flickr. All photos and videos listed in the collection are licensed under one of the Create Commons copyright licenses. This dataset is available directly from Yahoo! here. Only a subset of the YFCC100M videos will be used for this evaluation, this subset is to be determined.
The evaluation plan and license information will specify usage rules of the data resources in full detail.
HAVIC clips will be provided in MPEG-4 formatted files. The video will be encoded to the H.264 standard. The audio will be encoded using MPEG-4'S Advanced Audio Coding (AAC) standard. Please refer to the YFCC100M documentation regarding video file format.
In order to obtain the HAVIC corpora, ALL (including previous participants) TRECVID-registered sites must complete an evaluation license with the LDC. Each site who requires access to the HAVIC data, either as part of a team or as a standalone researcher, must complete a license.
To complete the evaluation license follow these steps:
- Download the license MED16 Data License.
- Return the completed license to LDC's Membership Office via email at ldc at ldc dot upenn dot edu. Alternatively, you may fax the completed license to the LDC at 215-573-2175.
- When you send the completed license to the LDC, include the following information:
- Registered TRECVID Team name
- Site/organization name
- Data contact person's name
- Data contact person's email
- Specifically which Corpora/Data listed in the agreement you need from the LDC, as participants in last years MED evaluation may already have some of the data on site
The designated data contact person for each site will receive instructions from the LDC about the specific procedures for obtaining the HAVIC data packages when they are released.
The following archives contain the input files for the Pre-Specified Event, and Ad-Hoc Event portions of the MED16 evaluation respectively.
Dry Run Evaluation
There are two relevant archives for the dry run evaluation:
- MED16DRYRUN_Files.tar.bz2 (md5) - Contains the index files for three events (E006, E009, and E013) tested against the MEDTest testing collection
- MED16_testTEAM_MED16DRYRUN_PS_1.tar.bz2 (md5) - A demonstration submission file for the dry run
In addition to the data files, F4DE-3.3.0 contains a scoring primer for MED '16 in the file DEVA/doc/TRECVid-MED16-ScoringPrimer.html that demonstrates how to score a system and prepare a submission file.
Evaluation scripts to support the MED evaluation are within the NIST Framework for Detection Evaluations (F4DE) Toolkit, a link to the latest release can be found on the NIST MIG tools page.
MED16 requires a F4DE version of 3.3.0 or later. The list of F4DE releases can be found on the F4DE GitHub page here.
Consult the TRECVID Master Schedule.
- Aug 31, 2016 - Added MED16 Ad-Hoc input files
- July 26, 2016 - Added MED16 Dry Run information
- July 1, 2016 - Added MED16 Pre-Specified input files
- April 1, 2016 - Initial version of evaluation plan added
- Mar 14, 2016 - Added Data Licensing section
- Jan 28, 2016 - Initial page created