Skip to content

Commit 38d58ba

Browse files
Christopher B. ChoyChristopher B. Choy
Christopher B. Choy
authored and
Christopher B. Choy
committed
Added analysis
1 parent 9953665 commit 38d58ba

File tree

3 files changed

+8
-5
lines changed

3 files changed

+8
-5
lines changed

README.md

+8-5
Original file line numberDiff line numberDiff line change
@@ -22,11 +22,14 @@ If you find this work useful in your research, please consider citing:
2222

2323
Traditionally, single view reconstruction and multi view reconstruction are disjoint problmes that has been dealt using different approaches. In this work, we first propose a unified framework for both single and multi view reconstruction using a `3D Recurrent Reconstruction Neural Network` (3D-R2N2).
2424

25-
| The schematic of the `3D-Convolutional LSTM` | Inputs (red cells + feature) for each cell (purple) |
26-
|:--------------------------------------------:|:---------------------------------------------------:|
27-
| ![3D-LSTM](imgs/lstm.png) | ![3D-LSTM](imgs/lstm_time.png) |
25+
| 3D-Convolutional LSTM | 3D-Convolutional GRU | Inputs (red cells + feature) for each cell (purple) |
26+
|:-------------------------:|:-----------------------:|:---------------------------------------------------:|
27+
| ![3D-LSTM](imgs/lstm.png) | ![3D-GRU](imgs/gru.png) | ![3D-LSTM](imgs/lstm_time.png) |
2828

29-
We can feed in images in random order since the network is trained to be invariant to the order. The ciritical component that enables the network to be invariant to the order is the `3D-Convolutional LSTM` which we first proposed in this work. The `3D-Convolutional LSTM` selectively updates parts that are visible and keeps the parts that are self occluded (please refer to [http://cvgl.stanford.edu/3d-r2n2/](http://cvgl.stanford.edu/3d-r2n2/) for the supplementary material for analysis).
29+
We can feed in images in random order since the network is trained to be invariant to the order. The ciritical component that enables the network to be invariant to the order is the `3D-Convolutional LSTM` which we first proposed in this work. The `3D-Convolutional LSTM` selectively updates parts that are visible and keeps the parts that are self occluded.
30+
31+
![LSTM Analysis](imgs/analysis.png)
32+
*Visualization of the 3D-Convolutional LSTM input gate activations. The images are fed into the network sequentially from left to right (top row). Visualization of input gate activations. The input gates corresponding to the parts that are visible and mismatch prediction open and update its hidden state (middle row). Corresponding prediction at each time step (bottom row).*
3033

3134
![Networks](imgs/full_network.png)
3235
*We used two different types of networks for the experiments: a shallow network (top) and a deep residual network (bottom).*
@@ -111,6 +114,6 @@ tar -xzf ShapeNetVox32.tgz -C ShapeNet/
111114
./experiments/script/res_gru_net.sh
112115
```
113116

114-
## LICENSE
117+
## License
115118

116119
MIT License

imgs/analysis.png

121 KB
Loading

imgs/gru.png

31.3 KB
Loading

0 commit comments

Comments
 (0)