Skip to content

Latest commit

 

History

History
76 lines (60 loc) · 4.5 KB

README.md

File metadata and controls

76 lines (60 loc) · 4.5 KB

This repository contains Python scripts showing how you can use Microsoft AirSim to collect image data from a moving vehicle, then use that data to train and test a deep-learning neural net in TensorFlow.

Prerequisites

  • Recommended hardware for running UnrealEngine4, required for AirSim. Although it is possible build AirSim on OS X and Linux, we found it easiest to use the pre-compiled Windows binaries in the Neighborhood example.

  • Python3 for 64-bit Windows

  • TensorFlow. To run TensorFlow on your GPU as we and most people do, you'll need to follow the directions for installing CUDA and CuDNN. We recommend setting aside at least an hour to make sure you do this right.

Instructions

  1. Clone this repository.
  2. Download and unzip the Neighborhood example, open it, and click run.bat to launch AirSim.
  3. When prompted, go with the default car simulation. If you press the 3 key on your keyboard, you will see the little image on which the neural net will be trained.
  4. From the repository, run the image_collection.py script. It will start the car moving and stop when the car collides with the fence, creating a carpix folder containing the images on which you will train the network in the next step.
  5. From the repository, run the collision_training.py script. Running on an HP Z440 workstation with NVIDIA GeForce GTX 1080 Ti GPU, we were able to complete the 500 training iterations in a few seconds.
  6. From the repository, run the collision_testing.py script. This should drive the car forward as before, but but the car should stop right before it hits the fence, based on the collision predicted by the neural net.

How it works

The image_collection script maintains a queue of the ten most recent images and saves them to numbered files in the carpix folder. The collision_training script converts these color images to grayscale, then builds a training set in which all images but the final one are labeled as safe (no collision; code [1 0]), and the final one is labeled as a collision (code [0 1]).
Finally, this training script uses Python's built-in pickle library to save the trained network parameters (weights and biases). The collision_testing script uses pickle to restore these parameters, then reconstructs the TensorFlow neural net from them. (We found this approach easier than using TensorFlow's save-and-restore API.) Finally, the collision_testing script moves the vehicle forward, converting the live image into grayscale and running it through the network to make a collision/no-collision prediction. When the value of the “collision bit” exceeds 0.5, the script stops the vehicle by applying the brakes.

Future work

Our single-layer logistic regression network provides a simple proof-of-concept example; however, for a more realistic data set involving collisions with different types of objects, a convolutional network would make more sense. AirSim also provides access to depth images (just press the 1 key during the simulation) which, like the Lidar on today's self-driving cars, would provide a valuable additional source of information for avoiding collisions.

Credits

This code represents the combined work of two teams in Prof. Simon D. Levy's fall 2017 AI course (CSCI 315) at Washington and Lee University (listed alphabetically):

  • Jack Baird
  • Alex Cantrell
  • Keith Denning
  • Rajwol Joshi
  • Will McMurtry
  • Jacob Rosen

Acknowledgement

We thank David Pfaff of the W&L IQ Center for providing the hardware on which we developed this project.