Skip to content

KAIST-VICLab/One-Look-is-Enough

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

[ICCV 2025] One Look is Enough: Seamless Patchwise Refinement for Zero-Shot Monocular Depth Estimation on High-Resolution Images


Corresponding author
1KAIST (Korea Advanced Institute of Science and Technology), South Korea

This repository is the official PyTorch implementation of "One Look is Enough: Seamless Patchwise Refinement for Zero-Shot Monocular Depth Estimation on High-Resolution Images". Our proposed method, PRO, achieves state-of-the-art zero-shot depth accuracy on high-resolution datasets with fine-grained details, outperforming existing depth refinement methods.


📧 News

  • Sep 10, 2025: Train code and Inference code are released
  • Jun 26, 2025: "One Look is Enough" is accepted to ICCV 2025
  • Mar 28, 2025: This repository is created

Tested Environment

  • OS: Ubuntu 20.04
  • Python: 3.8
  • PyTorch: 2.1.2
  • CUDA: 12.1
  • GPU: RTX 4090

Environment setup

conda env create -n pro --file environment.yml
conda activate pro

NOTE

Before running the code, please first run:

export PYTHONPATH="${PYTHONPATH}:/path/to/the/folder/One-Look-is-Enough"
export PYTHONPATH="${PYTHONPATH}:/path/to/the/folder/One-Look-is-Enough/external"

Make sure that you have exported the external folder which stores codes from other repos (ZoeDepth, Depth-Anything V1, V2, etc.)

Pretrained Models

Pre-trained models need to be placed in the ./pretrained/ directory.

  • PRO.pth: Trained on the UnrealStereo4K dataset. (Download)
  • depth_anything_v2_vitl.pth: Pre-trained Depth-Anything-V2-Large checkpoint. (Download)

File Structure:

One-Look-is-Enough_private/
└── pretrained/
    ├── Depth-Anything-V2/
    │   └── depth_anything_v2_vitl.pth   # Depth-Anything V2 model
    └── PRO/
        └── PRO.pth                     # PRO model (trained on UnrealStereo4K)

Running

To execute user inference, use the following command:

python tools/test_disp.py configs/test/test_general.py --cfg-option general_dataloader.dataset.rgb_image_dir='<img-directory>' [--save] [-save-residual] --work-dir <output-path> --test-type general --patch-split-num [h, w]

Arguments Explanation:

  • --cfg-option: Specify the input image directory. Maintain the prefix as it indexes the configuration. (To learn more about this, please refer to MMEngine. Basically, we use MMEngine to organize the configurations of this repo).
  • --save: Enable the saving of output files to the specified --work-dir directory (Make sure using it, otherwise there will be nothing saved).
  • --save-residual: Enable the saving of residual outputs, but this option only works if --save is enabled. It saves the residual data alongside the regular output.
  • --work-dir: Directory where the output files will be stored.
  • --patch-split-num: Define how the input image is divided into smaller patches for processing. You can specify any patch size,(h,w), where h is the height and w is the width. This helps control the granularity of image processing during inference. Default: (4 4).

User Training

Please refer to train for more details.

Results

Please visit our project page for more experimental results.

Citation

If the content is useful, please cite our paper:

@misc{kwon2025onelook,
      title={One Look is Enough: Seamless Patchwise Refinement for Zero-Shot Monocular Depth Estimation on High-Resolution Images}, 
      author={Byeongjun Kwon and Munchurl Kim},
      year={2025},
      eprint={2503.22351},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2503.22351}, 
}

License

The source codes including the checkpoint can be freely used for research and education only. Any commercial use should get formal permission from the principal investigator (Prof. Munchurl Kim, [email protected]).

Acknowledgement

This repository is built upon FMA-Net, C-DiffSET, and PatchFusion. We gratefully thank the PatchFusion authors for open-sourcing their code, which made our implementation and experiments much easier.

About

[ICCV 2025] Official repository of "One Look is Enough"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published