You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+30-23Lines changed: 30 additions & 23 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,21 +1,21 @@
1
1
sparse-to-dense.pytorch
2
2
============================
3
3
4
-
This repo implements the training and testing of deep regression neural networks for ["Sparse-to-Dense: Depth Prediction from Sparse Depth Samples and a Single Image"](https://arxiv.org/pdf/1709.07492.pdf) by [Fangchang Ma](http://www.mit.edu/~fcma) and [Sertac Karaman](http://karaman.mit.edu/) at MIT. A video demonstration is available on [YouTube](https://youtu.be/vNIIT_M7x7Y).
4
+
This repo implements the training and testing of deep regression neural networks for ["Sparse-to-Dense: Depth Prediction from Sparse Depth Samples and a Single Image"](https://arxiv.org/pdf/1709.07492.pdf) by [Fangchang Ma](http://www.mit.edu/~fcma) and [Sertac Karaman](http://karaman.mit.edu/) at MIT. A video demonstration is available on [YouTube](https://youtu.be/vNIIT_M7x7Y).
5
5
<palign="center">
6
6
<img src="http://www.mit.edu/~fcma/images/ICRA2018.png" alt="photo not available" width="50%" height="50%">
7
7
<img src="https://j.gifs.com/Z4qDow.gif" alt="photo not available" height="50%">
8
8
</p>
9
9
10
-
This repo can be used for training and testing of
10
+
This repo can be used for training and testing of
11
11
- RGB (or grayscale image) based depth prediction
12
12
- sparse depth based depth prediction
13
13
- RGBd (i.e., both RGB and sparse depth) based depth prediction
14
14
15
-
The original Torch implementation of the paper can be found [here](https://github.com/fangchangma/sparse-to-dense). This PyTorch version is under development and is subject to major modifications in the future.
15
+
The original Torch implementation of the paper can be found [here](https://github.com/fangchangma/sparse-to-dense).
16
16
17
17
## Thanks
18
-
Thanks to [Tim](https://github.com/timethy) and [Akari](https://github.com/AkariAsai) for their contribution.
18
+
Thanks to [Tim](https://github.com/timethy) and [Akari](https://github.com/AkariAsai) for their contributions.
19
19
20
20
## Contents
21
21
0.[Requirements](#requirements)
@@ -27,30 +27,31 @@ Thanks to [Tim](https://github.com/timethy) and [Akari](https://github.com/Akari
27
27
28
28
## Requirements
29
29
This code was tested with Python 3 and PyTorch 0.4.0.
30
-
- Install [PyTorch](http://pytorch.org/) on a machine with CUDA GPU.
30
+
- Install [PyTorch](http://pytorch.org/) on a machine with CUDA GPU.
31
31
- Install the [HDF5](https://en.wikipedia.org/wiki/Hierarchical_Data_Format) and other dependencies (files in our pre-processed datasets are in HDF5 formats).
- Download the preprocessed [NYU Depth V2](http://cs.nyu.edu/~silberman/datasets/nyu_depth_v2.html) dataset in HDF5 formats, and place them under the `data` folder. The downloading process might take an hour or so. The NYU dataset requires 32G of storage space.
37
+
- Download the preprocessed [NYU Depth V2](http://cs.nyu.edu/~silberman/datasets/nyu_depth_v2.html)and/or [KITTI Odometry](http://www.cvlibs.net/datasets/kitti/eval_odometry.php)dataset in HDF5 formats, and place them under the `data` folder. The downloading process might take an hour or so. The NYU dataset requires 32G of storage space, and KITTI requires 81G.
The training scripts come with several options, which can be listed with the `--help` flag. Currently this repo only supports training on the NYU dataset.
47
+
The training scripts come with several options, which can be listed with the `--help` flag. Currently this repo only supports training on the NYU dataset.
47
48
```bash
48
49
python3 main.py --help
49
50
```
50
51
51
-
For instance, run the following command to train a network with ResNet50 as the encoder, deconvolutions of kernel size 3 as the decoder, and both RGB and 100 random sparse depth samples as the input to the network.
52
+
For instance, run the following command to train a network with ResNet50 as the encoder, deconvolutions of kernel size 3 as the decoder, and both RGB and 100 random sparse depth samples as the input to the network.
52
53
```bash
53
-
python3 main.py -a resnet50 -d deconv3 -m rgbd -s 100
0 commit comments