TREC'24 ActEV Self-Reported Leaderboard (SRL) Challenge
- The Leaderboard is reporting results for TREC '24
- TREC'24 conference: Nov 18 — 22 (hybrid)
- TREC'24 ActEV test dataset release: Same as for TRECVID ActEV SRL 2023 and 2022
- TREC'24 ActEV SRL Challenge starts from June 01, 2024.
- TREC'24 ActEV SRL Challenge results submission deadline : October 7, 2024: 4:00 PM EST
- Primary Metric is Activity and Object Detection (AOD) and is based on Pmiss@0.1RFA.
You can download the public MEVA resources (training video, training annotations and the test set) as described on the SRL Data Tab.
ActEV SRL Test Datasets
The TREC 2024 ActEV SRL test dataset is the same as for TRECVID'23 ActEV SRL, TRECVID'22 ActEV SRL and CVPR'22 ActivityNet ActEV SRL challenge.
The ActEV Self-Reported Leaderboard (SRL) Challenge will report system performance scores on a public leaderboard on this website.
Please see details in the ActEV Self-Reported Leaderboard (SRL) Challenge evaluation plan below or check out the Updated ActEV Scoring Software GitHub repo.
For ActEV Evaluation information, please email: actev-nist@nist.gov
For ActEV Evaluation Discussion, please visit our ActEV Slack.
TREC 2024 conference Task: ActEV Self-Reported Leaderboard (SRL) Challenge
- TREC'24 ActEV test dataset release: May 1, 2024 (same as for TRECVID'23 ActEV SRL)
- ActEV SRL Challenge Opens: June 01, 2024
- Deadline for ActEV SRL Challenge results submission: October 7, 2024: 4:00 PM EST
- All teams invited for TREC workshop based on the participation TREC 2024 conference : Nov 18 — 22 (hybrid)
The ActEV Self-Reported Leaderboard (SRL) Challenge is based on the Multiview Extended Video with Activities (MEVA) Known Facility (KF) dataset. The MEVA KF data was collected at the Muscatatuck Urban Training Center (MUTC) with a team of over 100 actors performing in various scenarios. The MEVA KF dataset has two parts: (1) the public training and development data and (2) ActEV SRL test dataset .
The MEVA KF data were collected and annotated for the Intelligence Advanced Research Projects Activity (IARPA) Deep Intermodal Video Analytics (DIVA) program. A primary goal of the DIVA program is to support activity detection in multi-camera environments for both DIVA performers and the broader research community.
In December 2019, the public MEVA dataset has been released with 328 hours of ground-camera data and 4.2 hours of Unmanned Arial Vehicle video have been released. 160 hours of the ground camera video have been annotated by the same team that has annotated the ActEV test set. Additional annotations have been performed by the public and are also available in the annotation repository.
1)The TREC 2024 ActEV SRL test dataset has been released and is the same as for TRECVID 2023 ActEV SRL, TRECVID 2022 ActEV SRL and CVPR'22 ActivityNet ActEV SRL challenge.
There are four locations of data pertaining to the MEVA data resources and the evaluation. The sections below document how to obtain and use the data for the TREC 2024 ActEV SRL (same as HADCV evaluation).
- http://mevadata.org - general information about MEVA.
- MEVA AWS Video Data Bucket - The AWS bucket contains the video data for download.
- https://gitlab.kitware.com/meva/meva-data-repo - The GIT repo for public annotations.
- https://gitlab.kitware.com/actev/actev-data-repo - The GIT repo for files pertaining to ActEV and HADCV evaluations. This repo is the distribution mechanism for the TREC'24 ActEV SRL evaluation-related materials. The evaluations make use of multiple data sets. This repo is a nexus point between the evaluations and the utilized data sets. The repo consists of partition definitions (e.g., train, validation, or test) to be used for the evaluations.
- The TREC 2024 ActEV SRL test dataset is the same as for TRECVID 2023 ActEV SRL, TRECVID 2022 ActEV SRL and CVPR'22 ActivityNet ActEV SRL challenge
You can download the public MEVA video for free from the mevadata.org website (http://mevadata.org/) by completing these steps:
SRL Test data
- Get an up-to-date copy of the ActEV Data Repo via GIT. You'll need to either clone the repo (the first time you access it) or update a previously downloaded repo with 'git pull'.
- Clone: git clone https://gitlab.kitware.com/actev/actev-data-repo.git
- Update: cd "Your_Directory_For_actev-data-repo"; git pull
- Follow the steps in the top-level README.
- Download the TREC 2024 ActEV SRL Test dataset (same as for the HADCV'22 workshop) into ./partitions/HADCV22-Test-20211010 using the command:
% python scripts/actev-corpora-maint.py --regex ".*drop-4-hadcv22.*" --operation download
- Get an up-to-date copy of the MEVA Data Repo via GIT. You'll need to either clone the repo (the first time you access it) or update a previously downloaded repo with 'git pull'.
- Clone: git clone https://gitlab.kitware.com/meva/meva-data-repo
- Update: cd "Your_Directory_For_meva-data-repo"; git pull
- Download the training data collection found in the MEVA AWS Video Data Bucket within the directories: drops-123-r13, examples, mutc-3d-model, uav-drop-01, and updates-r13. (NOTE: directory drop-4-hadcv22 is NOT a training resource).
In the ActEV SRL evaluation, there is one primary task:
The Activity and Object Detection (AOD) task for detecting and
temporal/spatially localizes all instances of the activity from predefined activity
classes. For a system-identified activity instance to be evaluated as correct, the
type of activity must be correct, and the temporal/spatial overlap must fall within a
minimal requirement.
In the ActEV SRL evaluation, there is one secondary task:
The Activity Detection (AD) task for detecting and temporally localizing activities.
Given a target activity, a system automatically detects and temporally localizes all
instances of the activity from predefined activity classes. For a system-identified activity
instance to be evaluated as correct, the type of activity must be correct and must meet minimal temporal overlap requirement.
Systems will be tested on MEVA Known Facility test data. Facility data is available at https://mevadata.org including a site map with approximate camera locations and sample FOVs, camera models, a 3D site model, and additional metadata and site information. Sample representative video from the known facility is also provided, with over 160 hours of video annotated for leaderboard activities. All available metadata and site information may be used during system development.
For the MEVA Known Activities (KA) tests, developers are provided a list of activities in advance for use during system development (e.g., training) for the system to automatically detect and localize all instances of the activities.
Detailed activity definitions are in the ActEV Annotation Definitions for MEVA Data document.
The names of the 20 Known Activities for ActEV SRL (subset of the the SDL names ) :
person_closes_vehicle_door | person_reads_document |
person_enters_scene_through_structure | person_sits_down |
person_enters_vehicle | person_stands_up |
person_exits_scene_through_structure | person_talks_to_person |
person_exits_vehicle | person_texts_on_phone |
person_interacts_with_laptop | >person_transfers_object |
person_opens_facility_door | vehicle_starts |
person_opens_vehicle_door | vehicle_stops |
person_picks_up_object | vehicle_turns_left |
person_puts_down_object | vehicle_turns_right |
Datasets
- ActEV-supported data sets
-
Multiview Extended Video with Activities (MEVA) mevadata.org
- Video data is availabe as described in "Accessing and using MEVA" and "MEVA Download Instructions"
- MEVA Annotations GIT Repo: NEW: additional annotations posted weekly
- VIRAT
- ActEV Annotations: ActEV Data Repo. Access credentials provided during signup
-
Multiview Extended Video with Activities (MEVA) mevadata.org
- Kinetics
- AVA
- Moments-in-Time
- ActivityNet
- NVIDIA's CityFlow dataset
-
Live Datasets for Visual AI (Visym Collector)
- People in Public - 175k: 184,379 video clips of the ActEV activity classes for training in the unknown facility use-case.
Framework
The DIVA Framework is a software framework designed to provide an architecture and a set of software modules which will facilitate the development of activity recognition analytics. The Framework is developed as a fully open source project on GitHub. The following links will help you get started with the framework:- DIVA Framework Github Repository This is the main DIVA Framework site, all development of the framework happens here.
- DIVA Framework Issue Tracker Submit any bug reports or feature requests for the framework here.
- DIVA Framework Main Documentation PageThe source code for the framework documentation is maintained in the Github repository using Sphinx. A built version is maintained on ReadTheDocs at this link. A good place to get started in the documentation, after reading the Introduction is the UseCase section which will walk you though a number of typical use cases with the framework.
- KWIVER Github Repository This is the main KWIVER site, all development of the framework happens here.
- KWIVER Issue Tracker Submit any bug reports or feature requests for the KWIVER here. If there's any question about whether your issues belongs in the KWIVER or DIVA framework issues tracker, submit to the DIVA tracker and we'll sort it out..
- KWIVER Main Documentation Page The source for the KWIVER documentation is maintained in the Github repository using Sphinx. A built version is maintained on ReadTheDocs at this link. A good place to get started in the documentation, after reading the Introduction are the Arrows and Sprokit sections, both of which are used by the KWIVER framework.
Baseline Algorithms
KITWARE has adapted two "baseline" activity recognition algorithms to work within the DIVA Framework:Visualization Tools
Annotation Tools
- Kitware annotation tool (the tool natively supports the DIVA format)
- The VGG Image Annotator
- Scalabel (used for annotation of Berkeley DeepDrive project)
- VATIC - Video Annotation Tool
- BeaverDam
- VoTT: Visual Object Tagging Tool
- Computer Vision Annotation Tool (CVAT)
- Efficient Annotation of Segmentation Datasets with Polygon-RNN++