Skip to content

Elios-Lab/Driving-Scenario-Classification

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 

Repository files navigation

Design, implementation and testing of deep-learning-based driving scenario classifiers for automated vehicles

The aim of this project is to design, implement and test the performance of different neural networks for driving scenarios classification.

The first implemented networks are two versions of a residual 3D convolutional network:

  • the first one (R3D) uses fully 3D kernels
  • the second one (R(2+1)D) separates, in each constituting block, the 2D spatial convolution from the temporal convolution.

Finally, a (YOLOP + CNN) + LSTM stack was realized, which spatially processes each frame only once (instead than every time it appears in a sliding window of scenario analysis), in order to try and obtain latency times compatible with online analysis of driving scenarios. The networks performance were tests using a synthetic dataset composed by four simple class (approaching static leading vehicle, change lane, following leading vehicle and free ride) in three different conditions (day, night and rain with fog).

Networks’ structure

The three networks are implemented using pytorch and tensorflow:

R(2+1)D and R3D structure

(YOLOP + CNN) + LSTM structure

Dataset:

Syntetic dataset obtained using Carla simulator:

Approaching static leading vehicle Change lane Following leading vehicle Free ride

Example of previous dataset's sample processed with YOLOP

Example of sample od previous dataset processed with YOLOP

Results

The experimental results obtained seem to suggest two different use cases for the analyzed networks:

  • Offline:
    • R(2+1)D that is more suited when the scenario changes quickly or a low latency is required
    • R3D that is better for slow-changing scenarios or with low FPS rates.
  • Online:
    • (YOLOP + CNN) + LSTM

Static analysis: Confusion matrix (full video)

R(2+1)D confusion matrix (full video) R3D confusion matrix (full video) (YOLOP + CNN) + LSTM confusion matrix (full video)

Dinamic analysis (Different length video-clip and conditions)

R(2+1)D R3D (YOLOP + CNN) + LSTM

Accuracy and precision performances with different number of testing and training frames

Residual convolutional neural networks tested with 2 seconds video-clip length:

R(2+1)D and R3D

(YOLOP + CNN) + LSTM tested with 4 seconds video-clip length:

(YOLOP + CNN) + LSTM

Time inference performances:

R(2+1)D time inference R3D time inference (YOLOP + CNN) + LSTM time inference

Requirements

see requirements.txt

Acknowledge

YOLOP

R(2+1)D

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published