|
Data description
RAW images
All frames are recorded in RAW format using a GRGB Bayer filter and have to be interpolated to construct an RGB image. This can for example be done using the OpenCV function cvtColor(RAW_IM, RGB_IM, CV_BayerGR2BGR, 3), where RAW_IM is the input file and RGB_IM is the output file.
Naming
Frames for each camera are provided in a separate ZIP file per camera. For the Train station data, there are 14 sub folders for the 14 scenarios. Because the Hall data consists of one very long scene, the data has been devided over 38 sub folders numbered 000000 - 000037. Each of these sub folders contains 250 images (1000 images if all cameras are put together), except for the first and last sub foldersub, which contain 80 and 71 files respectively.
All files have been named according to the following scheme: frame<XXXXXX>_cam<X>_<XXXXXXXXXX.XXXXXX>.raw, where the first X-es are the frame number, the next X is the camera number and the last set of X-es is the frame timestamp, provided as a UNIX timestamp in seconds (with microsecond accuracy).
Background images
Together with the frames for all scenarios, frames showing the empty scene are provided to learn background models. 10 frames are provided in RAW format for the Hall dataset and most Train station scenarios. For two Train station scenarios (scenario12-2 and scenario13-2), no frames of the empty scene with matching lighting conditions were available. For these scenarios, an empty background image was constructed for each camera. These images are provided as PPM files named <X>_bgim.ppm, where <X> is the camera number. Background files can be found in the "backgrounds" folder in each camera folder.
Camera calibration
All cameras are calibrated intrinsicly as well as extrinsicly. Calibration parameters are provided in sparate ZIP files and are stored in files named <data>_c<x>_calib.txt, where <data> is either "train" for the train station data or "hall" for the hall data, and <x> is the camera number. Together with the data for the intrinsic and extrinsic camera calibration matrices, these files also provide the distortion coefficients k1-k5.
Furthermore, groundPlane.gpl files defining the ground plane are provided, in which the ground plane is defined as the normal vector to the plane (3 coordinates) and an offset on that vector. Finally, files named <data>_roi.txt (with <data> either "train" or "hall") are provided, defining the position and orientation of the regions of interest in world coordinates.
Annotations
Annotations are provided in separate ZIP files, as a single TXT file per scenario. Person positions are given as XY ground plane coordinates, in millimeters. Each file is structured as follows:
frame
|
p0_x
|
p0_y
|
p1_x
|
p1_y
|
p2_x
|
p2_y
|
...
|
650
|
9732
|
3931
|
7380
|
3882
|
NaN
|
NaN
|
...
|
651
|
9733
|
3935
|
7388
|
3880
|
7229
|
3246
|
...
|
...
|
...
|
...
|
...
|
...
|
...
|
...
|
...
|
The first row contains column labels. The columns p<X>_x and p<X>_y represent the annotated x and y ground plane coordinate for person <X>. When no annotation is available for a certain person at a certain time step, it's XY position is specified as <NaN, NaN>.
The file evaluation.zip contains a Matlab script to compute the CLEAR_MOT metrics used in the above publication. The script calculate_metrics.m takes two TXT files as input, containing tracking results and annotations, and computes the CLEAR-MOT metrics from these files. In order to use this script, the tracking results should be provided in the same way as teh annotation data, where tracks that last for periods of time shorter than the full scenario length are padded with NaN values, similar to missing values in the annotation data. Note that the script assumes the first line of each file to contain column labels, and as such, skips the first line when reading the file!
Similar to the publication mentioned above, the evaluation of tracking results is done in the area covered by all cameras. If a detection or annotation is outside this area, or is matched to an annotation or detection outside this area, it is not counted as an correct match, nor as an error. The area covered by all cameras is specified as a binary lookup image matching the region of interest. The images are found in the evaluation folder, named boundim_3_cam_trainstation.png (Train station data) and boundim_4_cam_hall.png (Hall data) and are automatically used by the evaluation script when specifying 'trainstation' or 'hall' as input to the compute_metrics() function.
License Terms
This dataset is made available to the scientific community for non-commercial research purposes such as academic research, teaching, scientific publications, or personal experimentation. Permission is granted to use, copy, and distribute the data given that you agree:
- That the dataset comes "AS IS", without express or implied warranty. Although every effort has been made to ensure accuracy, the University of Amsterdam (or this webhost) does not accept any responsibility for errors or omissions.
- That you include a reference to the above publication in any published work that makes use of the dataset.
- That if you have altered the content of the dataset or created derivative work, prominent notices are made so that any recipients know that they are not receiving the original data.
- That you may not use or distribute the dataset or any derivative work for commercial purposes as, for example, licensing or selling the data, or using the data with a purpose to procure a commercial gain.
- That this original license notice is retained with all copies or derivatives of the dataset.
- That all rights not expressly granted to you are reserved by the University of Amsterdam.
I DO NOT AGREE with the above license terms I AGREE with the above license terms
|