Skip to content

An autonomous wildlife camera (Jetson ChallengeRocket submission)

Notifications You must be signed in to change notification settings

mynameisjohn/rAInger

Repository files navigation

rAInger

An autonomous wildlife camera
(NVIDIA® Jetson™ Developer Challenge submission)

The goal of this project is to design a program that enables an Nvidia Jetson computer equipped with a camera to autonomously capture and analyze pictures. These pictures are fed to a convolutional neural network (CNN) trained to recognized various types of animals, allowing the Jetson to independently track wildlife in the field.

That's the theory, anyway. I don't have easy access to wildlife training set, so my networks are trained on cats and dogs. The high level purpose and design of this program is described further this short video, which I've included as part of my submission. Rather than repeat that information, I'll use this README to describe the actual implementation.

rAInger submission video


0. File Index and Libraries

There are three source files in this repository

  • rAInger.py: the main script, which runs the code that trains the network and uses it to analyze images
  • rAInger_Util.py: an auxilliary script where most of the functions used in rAInger.py are implemented
  • rAInger_pyl.cpp: A C++ source file that communicates with an onboard radio to transmit analysis data

Additionally there are two subrepositories (lora_gateway and PyLiaison). The former is used for radio communication, and the latter allows the python code to communicate with C++ code. Pyliaison should behave nicely with CMake. There is no CMakeLists.txt for libloragw, but if you follow the instructions in lora_gateway/libloragw/install_ftdi.txt and build within that directory you should be good to go.

The python code uses the following libraries

  • numpy (array and math operations)
  • opencv (for image IO and processing)
  • PIL (to load GIF images)
  • tensorflow-gpu (underlying network and CUDA implementation)
  • tflearn (high level DNN that wraps tensorflow code)
  • matplotlib (optional, used for testing and verification)

argparse is also used quite heavily to control program parameters. A full listing of parameters can be seen in rAInger_Util.get_arguments.


1. Training

The first thing we have to do is train our CNN. The data used to train the network was obtained from kaggle as a part of their Dogs vs. Cats challenge. Because the training step is computationally intense I chose to run it on my desktop computer which has an Nvidia 980 Ti graphics card. Once the network was trained I was able to save it to a file and load it pre-trained on the Jetson.

The training algorithm works by splitting the kaggle data set into two parts - one we use to train and one we use to validate. There are 25000 kaggle images - we use 500 as a validation set and train on the remaining 24500. By iteratively changing the network so as to match the training set with the validation set our network gradually becomes more accurate. Information output by tflearn gives us an indication of how accurate our net is - after running it with our default parameters (specified in rAInger_Util.get_arguments) we get this result:

| Adam | epoch: 010 | loss: 0.33058 - acc: 0.8508 | val_loss: 0.55176 - val_acc: 0.7860 -- iter: 24500/24500

Indicating an accuracy of 85.08% - not too bad!


2. Prediction

Once the network is trained we can use it to analyze the content of images. Because it was trained on cats and dogs that's what it can recognize, but with a more diverse training set it could be used to analyze images of other types of wildlife (or anything, really. )

We run the prediction code on live input. The code uses a simple OpenCV motion detector to wait for something interesting to go by, and when it does it runs the prediction code to determine what it's seeing.

In the field the input will be coming from a camera, but for testing purposes I optionally allow reading a .GIF file as input.

The rectangles you see are the areas in which motion was detected. We take motion as a trigger to start running images through the prediction network - while motion occurs we find the image with the strongest prediction(in this case, which image is the most "cat or dog"-like). That strength gets returned to us as a 2-element array - the first element is how certain we are it's a cat, and the second how certain we are it's a dog. We want the image with the maximum element - we don't care if it's the first of the second, just so long as it's the max. That gets us this image:

The vector we're returned is [1.3075967, -1.1696696]. Adding 'softmax' layer to our network ensures the values sum to 1, giving us something like a percent certainty: [0.9225327 0.07746734]. Given that this is the strongest prediction we got during the motion window, we can say we're about 92% certain there's a cat in front of us (it must be that itty bitty nose!)

The reason we choose not to use the softmax layer is that we can't always be certain that what triggered the motion is something we recognize. If, for example a hamster were to cross in front of the camera, the softmax layer would still feel certain that it was either a cat or a dog. This is useful while training because we know we're feeding it something we'll recognize, but in the wild we want to know if it's something we don't recognize.

For this reason we use a softmax layer and a regression when training so as to reach an accurate model more quickly, but in the wild we use either a linear or rectified linear layer to avoid classifying junk as something we know.

# Use a softmax layer when training because we can be certain the input is either a cat or a dog
if bTrain:
    convnet = fully_connected(convnet, 2, activation='softmax')
# Otherwise use a rectified linear layer to force unlikely results to zero
else:
    convnet = fully_connected(convnet, 2, activation='relu')

3. Communication

Now that we've figured we've seen something, we've got to send it over to our "central database". Admittedly this is an abstract concept for the time being, but what we can do is transmit some data over a LoRaWAN antenna to some remote gateway. The easiest way I know of doing this is using the lora_gateway library (libloragw). This kind of thing could be implemented in Python, but I honestly don't trust myself to do it on such short notice.

However, libloragw is a C library, and in order for us to use tensorflow and tflearn we've got to be using Python. My solution is to use PyLiaison, a library I wrote to make it easy to call C++ from Python and vice versa. I use any opportunity I can to plug the library, but in situations like this it's quite useful.

What we really need is a way of sending raw bytes over the antenna. We've conveniently given cats an label of 0 and dogs an label of 1, so all we need to do is send a 0 or 1 each time we see a cat or dog, respectively. With this function,

// Sends raw bytes over the antenna to our gateway
bool send_loro_data( std::vector<char> vData )
{
	// create packet from template, store packet
	struct lgw_pkt_tx_s txPkt = g_TxPktTemplate;
	txPkt.size = std::min<uint16_t>( vData.size(), sizeof( txPkt.payload ) );
	memcpy( txPkt.payload, vData.data(), txPkt.size );

	// send - this is a non-blocking call so don't bunch up...
	return LGW_HAL_SUCCESS == lgw_send( txPkt );
}

We can invoke this function from Python by exposing it into a module like so

// Create python module named pylLoRaWAN with send_loro_data
pyl::ModuleDef * pLLGWDef = pylCreateMod( pylLoRaWAN );
pylAddFnToMod( pLLGWDef, send_loro_data );

and calling it from Python

import pylLoRaWAN

def send_lora_data(data):
    if not (isinstance(data, bytes) or isinstance(data, bytearray)):
        raise RuntimeError('We can only send bytes over LoRaWAN')
    return pylLoRaWAN.send_lora_data(data)

The data variable, if it is a bytes or bytearray type, will be converted to a std::vector<char> and used by the C++ code in the send_loro_data function. In this way we can string together a bytearray of what we've seen and send it over the antenna.


4. Conclusion

Hopefully the code here will demonstrate a working proof of concept for what rAInger could be. I hope you consider it a valid submission to the Jetson Developer Challenge, and thank you for your time.


5. Acknowledgements
  • Medium for the tutorial upon which this work is based
  • kaggle for the training data set
  • PyImageSeach for the motion detecion logic
  • Kurt Keville for loaning me one of his Jetson TX2s as well as the LoRaWAN antenna and guidance.
  • mirakonta for libloragw
  • Humble Bundle for giving out Sony Vegas for a very low price - without it I would've struggled to make the video!
  • Thanks as well to markdownify, which I'm using right now to write this readme.

About

An autonomous wildlife camera (Jetson ChallengeRocket submission)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published