DeepOSM 
Classify roads and features in satellite imagery, by training neural networks with OpenStreetMap (OSM) data.
DeepOSM can:
- Download a chunk of satellite imagery
- Download OSM data that shows roads/features for that area
- Generate training and evaluation data
- Display predictions of mis-registered roads in OSM data, or display raw predictions of ON/OFF
Running the code is as easy as install Docker, make dev, and run a script.
Contributions are welcome. Open an issue if you want to discuss something to do, or email me.
Default Data/Accuracy
By default, DeepOSM will analyze about 200 sq. km of area in Delaware. DeepOSM will
- predict if the center 9px of a 64px tile contains road.
- use the infrared (IR) band and RGB bands.
- be 75-80% accurate overall, training only for a minute or so.
- use a single fully-connected relu layer in TensorFlow.
- render, as JPEGs, “false positive” predictions in the OSM data – i.e. where OSM lists a road, but DeepOSM thinks there isn’t one.

Background on Data – NAIPs and OSM PBF
For training data, DeepOSM cuts tiles out of NAIP images, which provide 1-meter-per-pixel resolution, with RGB+infrared data bands.
For training labels, DeepOSM uses PBF extracts of OSM data, which contain features/ways in binary format that can be munged with Python.
The NAIPs come from a requester pays bucket on S3 set up by Mapbox, and the OSM extracts come from geofabrik.
Install Requirements
DeepOSM has been run successfully on both Mac (10.x) and Linux (14.04 and 16.04). You need at least 4GB of memory.
AWS Credentials
You need AWS credentials to download NAIPs from an S3 requester-pays bucket. This only costs a few cents for a bunch of images, but you need a credit card on file.
export AWS_ACCESS_KEY_ID='FOO'
export AWS_SECRET_ACCESS_KEY='BAR'
Install Docker
First, install a Docker Binary.
I also needed to set my VirtualBox default memory to 4GB, when running on a Mac. This is easy:
- start Docker, per the install instructions
- stop Docker
- open VirtualBox, and increase the memory of the VM Docker made
(GPU Only) Install nvidia-docker
In order to use your GPU to accelerate DeepOSM, you will need to download and install the latest NVIDIA drivers for your GPU, and (after first installing docker itself), install nvidia-docker
.
First, find the latest NVIDIA drivers for your GPU on NVIDIA’s website. Make sure you check the version number of the driver, as the most recent release isn’t always the latest version.
Once you have downloaded the appropriate NVIDIA-*.run
file, install it as follows (based on these instructions):
Ensure your system is up-to-date and reboot to ensure the latest installed kernel is loaded:
# ensure your packages are up-to-date
sudo apt-get update
sudo apt-get dist-upgrade
# and reboot
sudo reboot
Once your system has rebooted, install build-essential
and the linux-headers
package for your current kernel version (or equivalents for your linux distribution):
sudo apt-get install build-essential linux-headers-$(uname -r)
Then run the NVIDIA driver install you downloaded earlier, and reboot your machine afterwards:
<div class="snippet-clipboard-content position-relative" data-snippet-clipboard-copy-content="sudo bash
sudo reboot
“>
sudo bash <location of ./NVIDIA-Linux-*.run file>
sudo reboot
Finally, verify that the NVIDIA drivers are installed correctly, and your GPU can be located using nvidia-smi
:
Thu Mar 9 03:40:33 2017
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 367.57 Driver Version: 367.57 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 GRID K520 Off | 0000:00:03.0 Off | N/A |
| N/A 54C P0 45W / 125W | 0MiB / 4036MiB | 0% Default |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: GPU Memory |
| GPU PID Type Process name Usage |
|=============================================================================|
| No running processes found |
+-----------------------------------------------------------------------------+
Now that the NVIDIA drivers are installed, nvidia-docker
can be downloaded and installed as follows (based on these instructions):
wget -P /tmp https://github.com/NVIDIA/nvidia-docker/releases/download/v1.0.1/nvidia-docker_1.0.1-1_amd64.deb
sudo dpkg -i /tmp/nvidia-docker*.deb && rm /tmp/nvidia-docker*.deb
And you can confirm the installation, by attempting to run nvida-smi
inside of a docker container:
nvidia-docker run --rm nvidia/cuda nvidia-smi
Using default tag: latest
latest: Pulling from nvidia/cuda
d54efb8db41d: Pull complete
f8b845f45a87: Pull complete
e8db7bf7c39f: Pull complete
9654c40e9079: Pull complete
6d9ef359eaaa: Pull complete
cdfa70f89c10: Pull complete
3208f69d3a8f: Pull complete
eac0f0483475: Pull complete
4580f9c5bac3: Pull complete
6ee6617c19de: Pull complete
Digest: sha256:2b7443eb37da8c403756fb7d183e0611f97f648ed8c3e346fdf9484433ca32b8
Status: Downloaded newer image for nvidia/cuda:latest
Thu Mar 9 03:44:23 2017
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 367.57 Driver Version: 367.57 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 GRID K520 Off | 0000:00:03.0 Off | N/A |
| N/A 54C P8 18W / 125W | 0MiB / 4036MiB | 0% Default |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: GPU Memory |
| GPU PID Type Process name Usage |
|=============================================================================|
| No running processes found |
+-----------------------------------------------------------------------------+
Once you have confirmed nvidia-smi
works inside of nvidia-docker
, you should be able to run DeepOSM using your GPU.
Run Scripts
Start Docker, then run:
Or if you don’t have a capable GPU, run:
Download NAIP, PBF, and Analyze
Inside Docker, the following Python scripts will work. This will download all source data, tile it into training/test data and labels, train the neural net, and generate image and text output.
The default data is six NAIPs, which get tiled into 64x64x4 bands of data (RGB-IR bands). The training labels derive from PBF files that overlap the NAIPs.
python bin/create_training_data.py
python bin/train_neural_net.py
For output, DeepOSM will produce some console logs, and then JPEGs of the ways, labels, and predictions overlaid on the tiff.
Testing
There is a very limited test suite available at the moment, that can be accessed (from the host system) by running:
make test
Jupyter Notebook
Alternately, development/research can be done via jupyter notebooks:
make notebook
To access the notebook via a browser on your host machine, find the IP VirtualBox is giving your default docker container by running:
docker-machine ls
NAME ACTIVE DRIVER STATE URL SWARM DOCKER ERRORS
default * virtualbox Running tcp://192.168.99.100:2376 v1.10.3
The notebook server is accessible via port 8888, so in this case you’d go to: http://192.168.99.100:8888
Readings
Also see a work journal here.
Papers – Relevant Maybe
Papers – Not All that Relevant
Papers to Review
Recent Recommendations
Citing Mnih and Hinton
I am reviewing these papers from Google Scholar that both cite the key papers and seem relevant to the topic.
Original Idea
This was the general idea to start, and working with TMS tiles sort of worked (see first 50 or so commits), so DeepOSM got switched to better data:

GitHub
https://github.com/trailbehind/DeepOSM