- 07/2021: RAD tensors of the CARRADA dataset are now available here.
- 04/2021: A new version of CARRADA is released ! You have now access to the angle-Doppler view of the RAD tensor, each view are also available with different processing (have a look into the README of the dataset).
- 01/2021: Our paper has been presented at ICPR2020
CARRADA Dataset: Camera and Automotive Radar with Range-Angle-Doppler Annotations, ICPR 2020.
Arthur Ouaknine, Alasdair Newson, Julien Rebut, Florence Tupin, Patrick Pérez
If you find this code or the dataset useful for your research, please cite our paper:
@INPROCEEDINGS{9413181,
author={Ouaknine, Arthur and Newson, Alasdair and Rebut, Julien and Tupin, Florence and Pérez, Patrick},
booktitle={2020 25th International Conference on Pattern Recognition (ICPR)},
title={CARRADA Dataset: Camera and Automotive Radar with Range- Angle- Doppler Annotations},
year={2021},
volume={},
number={},
pages={5068-5075},
doi={10.1109/ICPR48806.2021.9413181}
}
It is strongly recommanded to use Docker with the provided Dockerfile containing all the dependencies.
- Clone the repo:
$ git clone https://github.com/valeoai/carrada_dataset.git
- Create the Docker image:
$ cd carrada_dataset/
$ docker build . -t "carrada_dataset:Dockerfile"
Important note: The CARRADA dataset will be downloaded and unzipped in the Docker image. Make sure that the Docker user has sufficent rights and has around 90GB of available space on its hard drive. The allocated space to an image can be modified using the option --memory="100g"
.
- Run a container and join an interactive session, you will be able to run the code and the tests on this session:
$ docker run -d -it --name containername carrada_dataset:Dockerfile sleep infinity
$ docker exec -it containername bash
You can either use Docker with the provided Dockerfile containing all the dependencies, or follow these steps.
- Clone the repo:
$ git clone https://github.com/valeoai/carrada_dataset.git
- Install this repository and the dependencies using pip:
$ cd carrada_dataset/
$ pip install -e .
With this, you can edit the carrada_dataset code on the fly and import function and classes of carrada_dataset in other project as well.
- Install OpenCV if you don't already have it:
$ conda install -c menpo opencv
-
Install all the dependencies using pip and conda, please take a look at the Dockerfile for the list and versions of the dependencies.
-
Optional. To uninstall this package, run:
$ pip uninstall carrada_dataset
You can take a look at the Dockerfile if you are uncertain about steps to install this project.
If you are using Docker, the downloading and extraction of the dataset is already contained in the Dockerfile.
Otherwise, the dataset is available on Arthur Ouaknine's personal web page using this link: https://arthurouaknine.github.io/codeanddata/carrada
Importante note: the Carrada.tar.gz file is 23GB but once it is extracted, the dataset is around 90GB. Please be sure you have at least 90GB avalaible on your hard drive.
The CARRADA dataset contains the camera images, the raw radar data and the generated annotations. A README.md
file in the dataset provides details about all the files. It is not mandatory to run the entire pipeline to obtain the annotations.
The .bold[RAD tensors] of each recorded sequence are also available in a separated file due to its large size (176GB compressed, 198GB otherwise). This part of the dataset will not be automatically downloaded using this repo, however it is publicly accessible on Arthur Ouaknine's personal web page.
The annotation generation pipeline is composed of 4 blocks:
- generate and track instances using images,
- generate range-Doppler points from the segmented instance,
- generate and track instances in the DoA-Doppler representation,
- generate the annotation files.
- Run the entire pipeline:
It is mandatory to specify the path where the CARRADA dataset is located. Example: I put the Carrada
folder in /datasets/
, the path I should specify is /datasets/
. If you are using Docker, the CARRADA dataset is extracted in the /datasets/
folder by default.
$ cd scripts/
$ bash run_annotation_pipeline.sh /datasets/
- Run the blocks independently:
If the user didn't set the path to the Carrada
folder yet, the following lines must be executed:
$ cd carrada_dataset/scripts/
$ python set_path.py /datasets/
Then, each script can be executed independently. Note that each step generates mandatory data for the next one, it is important to keep the pipeline order if the user doesn't have the intermediate data. All intermediate data are provided in the CARRADA dataset, thus the user is able to run any step if the data are downloaded correctly.
$ cd scripts/
$ python name_of_the_script.py
Tests have been implemented to ensure the consistency of the pipeline.
- Run all the tests:
It is mandatory to specify the path where the CARRADA dataset is located. Example: I put the Carrada
folder in /datasets/
, the path I should specify is /datasets/
. If you are using Docker, the CARRADA dataset is extracted in the /datasets/
folder by default.
$ cd tests/
$ bash run_tests.sh /datasets/
- Run the tests independently
If the user didn't set the path to the Carrada
folder yet, the following lines must be executed:
$ cd carrada_dataset/scripts/
$ python set_path.py /datasets/
Then, each test can be executed independently. Note that the script test_transform_data.py
requires RAD tensors which will be available in a next release.
$ cd tests/
$ python name_of_the_test.py
A Jupyter Notebook visualize_samples.ipynb
is provided to visualize samples of the CARRADA dataset with annotations.
Note that this notebook also uses RAD tensors, please comment or modify the code if necessary.
Run Jupyter on Docker: a command line is already written in the Dockerfile to execute Jupyter on a container. The user should map the port of the container on its local machine using the following command:
$ docker run -p 8889:8889 carrada_dataset:Dockerfile
The Jupyter interface is then available on your local machine at localhost:8889
.
- Special thanks to the SensorCortex team which has recorded the data.
- Special thanks for @gabrieldemarmiesse for his valuable technical help.
- Code for Simple Online and Realtime Tracking (SORT) is borrowed from https://github.com/abewley/sort
- Code for Fully Convolutional Network (FCN) baseline is borrowed from https://github.com/wkentaro/pytorch-fcn
- The carrada_dataset repo is released under the GNU GPL 3.0 license.
- The CARRADA dataset is released under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International Public License (“CC BY-NC-SA 4.0”).
- Migration to Pathlib for multi os support