A new dataset with both first and third-person videos—Charades-Ego—is now available.
Charades is dataset composed of 9848 videos of daily indoors activities collected through Amazon Mechanical Turk. 267 different users were presented with a sentence, that includes objects and actions from a fixed vocabulary, and they recorded a video acting out the sentence (like in a game of Charades). The dataset contains 66,500 temporal annotations for 157 action classes, 41,104 labels for 46 object classes, and 27,847 textual descriptions of the videos. This work was presented at ECCV2016.
Each video has been exhaustively annotated using consensus from 4 workers on the training set, and from 8 workers on the test set. Please refer to the updated accompanying publication for details. Please contact firstname.lastname@example.org for questions about the dataset.
|Asynchronous Temporal Fields||22.4% mAP [*]|
|LSTM w/ post-processing||10.4% mAP|
|Two-Stream w/ post-processing||10.9% mAP|
|Asynchronous Temporal Fields||12.8% mAP [*]|