Skip to content

Commit

Permalink
doc: added image for comparison to YOLOv2 and fixed links in README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
ksakmann committed Feb 24, 2017
1 parent ec02be4 commit bda31cd
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 3 deletions.
10 changes: 7 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@ This pipeline is then used to draw bounding boxes around the cars in the video.
[image6]: ./images/heatmap.png
[image7]: ./images/labels.png
[image8]: ./images/bounding_boxes.png
[image8]: ./images/yolo-hero.png
[video1]: ./output_images/processed_project_video.mp4

# Getting Started
Expand Down Expand Up @@ -119,15 +120,18 @@ Finally the resulting bounding boxes are drawn onto the last frame in the series


## Comparsion to YOLO

While I was happy with the results of an SVM + HOG approach, I also wanted to check what a state of the art deep network could do. This comparison is anything but fair.
We did not use the GPU at all in this project and I am shamelessly comparing apples and pears now. YOLO stands for "You only look once" and tiles an image into a modest number of squares.
Each of the squares is responsible for predicting whether there is an object centered around it and if so predicting the shape of its bounding box, together with a confidence level.


![YOLOv2-result][image9]

For the comparsion I cloned and compiled the original YOLO implementation from the [darknet](https://pjreddie.com/darknet/yolo/) website.
Please read more about YOLO [here](https://arxiv.org/abs/1506.02640). I used the weights of YOLOv2 trained on the Common Objects in Context dataset (COCO) which are also available at the darknet website.
Feeding in the project video on a GTX 1080 averages about 65 FPS or about 20x faster than the current SVM + HOG pipeline. Here is the result of passing the project video through YOLOv2:

[yolo_result.avi](./output_images/yolo_result.avi).
Feeding in the project video on a GTX 1080 averages about 65 FPS or about 20x faster than the current SVM + HOG pipeline.
Here is the result of passing the project video through YOLOv2: [yolo-result.avi](./output_images/yolo-result.avi).

This is an extremely exciting result. Note that false positives are practically absent. So there is no need at all here for a heatmap,
although it certainly could be used to reduce any possible false positives. I vehicle detection with YOLO type networks are an exciting direction to
Expand Down
Binary file added images/yolo-hero.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

0 comments on commit bda31cd

Please sign in to comment.