In this paper, we present a baseline workflow for behavioral cloning of human driving behavior using deep learning, specifically an end-to-end implementation. The proposed workflow involves training an artificial neural network of the convolutional type, which is then deployed on two different driving behavior models. These models are subsequently simulated in a virtual environment. The training process encompasses data collection, augmentation, preprocessing, and the training of a neural network. The trained model is then deployed onto an ego vehicle to predict steering commands based on real-time camera feed input.
Keywords: Autonomous Vehicles, Deep Learning, Behavioral Cloning
Behavior planning in autonomous driving involves designing high-level driving actions or maneuvers to safely achieve driving objectives under various scenarios. This planning must consider interactions with static and dynamic elements in the environment while ensuring safety and efficiency. This paper focuses on end-to-end imitation learning for autonomous driving, which has shown potential but also limitations. The study explores the behavior planning system in the context of diverse driving scenarios.
The proposed pipeline is validated using an open-source simulator by Udacity.
The study involves cloning two distinct driving behaviors, basic driving, and rigorous driving, using end-to-end learning. These behaviors are tested on specialized simulated environments tailored to their characteristics.
The training phase comprises data collection, augmentation, preprocessing, and neural network training.
Datasets are acquired by manually driving the ego vehicle in different scenarios. Each dataset includes timestamped frames from the center, left, and right cameras, along with corresponding steering angle measurements.
Various augmentation techniques are applied to the collected data, including zoom, pan, brightness, and flip, to improve the robustness of the training dataset.
Images are preprocessed through scaling and standardization operations to prepare them for input into the neural network.
The training pipeline is implemented using a Jupyter notebook. Data is shuffled, balanced, and augmented in real-time during training.
A shallow CNN architecture is adopted to predict steering angles from camera frames.
The deployment phase involves data preprocessing and longitudinal vehicle control loop.
Images are resized and normalized to prepare them for input into the trained neural network.
Throttle commands are determined based on the predicted steering angles, enabling vehicle control.
The computational details, including training time and deployment latencies, are provided. Simulated field results show activation maps and demonstrate the neural network's ability to make intelligent steering predictions based on learned features.
The paper presents a lightweight pipeline for training and deploying strong driving behavior models using end-to-end imitation learning. It emphasizes cooperative control and proposes evaluation criteria for assessing the pipeline's efficiency and resilience. Future work includes investigating hardware implementations, gathering diverse datasets, and exploring alternative techniques for generalization.
For video demonstrations of the experiments, visit YouTube link.