Underwater Light Field Retention : Neural Rendering for Underwater Imaging (UWNR) (Accepted by CVPR Workshop2022 NTIRE)
Authors: Tian Ye†, Sixiang Chen†, Yun Liu, Erkang Chen*, Yi Ye, Yuche Li
- † represents equal contributions.
- * represents corresponding author.
Abstract: Underwater Image Rendering aims to generate a true-tolife underwater image from a given clean one, which could be applied to various practical applications such as underwater image enhancement, camera filter, and virtual gaming. We explore two less-touched but challenging problems in underwater image rendering, namely, i) how to render diverse underwater scenes by a single neural network? ii) how to adaptively learn the underwater light fields from natural exemplars, i,e., realistic underwater images? To this end, we propose a neural rendering method for underwater imaging, dubbed UWNR (Underwater Neural Rendering). Specifically, UWNR is a data-driven neural network that implicitly learns the natural degenerated model from authentic underwater images, avoiding introducing erroneous biases by hand-craft imaging models.
Compared with existing underwater image generation methods, UWNR utilizes the natural light field to simulate the main characteristics ofthe underwater scene. Thus, it is able to synthesize a wide variety ofunderwater images from one clean image with various realistic underwater images.
Extensive experiments demonstrate that our approach achieves better visual effects and quantitative metrics over previous methods. Moreover, we adopt UWNR to build an open Large Neural Rendering Underwater Dataset containing various types ofwater quality, dubbed LNRUD.
- python3
- Pytorch 1.9.0
- Numpy 1.19.5
- Opencv 4.5.5.62
- NVDIA 2080TI GPU + CUDA 11.4
- NVIDIA Apex 0.1
- tensorboardX(optional)
The LNRUD generated by our Neural Rendering architecture can be downloaded from LNRUD Password:djhh or Google Drive, which contains 50000 clean images and 50000 underwater images synthesized from 5000 real underwater scene images.
All datasets can be downloaded, including UIEB, NYU, RESIDE and SUID
Train with the DDP mode under Apex 0.1 and Pytorch1.9.0
Put clean images in clean_img_path.
Put depth images in depth_img_path.
Put real underwater images as training ground-truth in underwater_path.
Put real underwater images as FID_gt in fid_gt_path.
Run the following commands:
python3 -m torch.distributed.launch --master_port 42563 --nproc_per_node 2 train_ddp.py --resume=True --clean_img_path clean_img_path --depth_img_path depth_img_path --underwater_path underwater_path --fid_gt_path fid_gt_path --model_name UWNR
You can download pre-trained model from Pre-trained model Password:42w9 or Google Drive and save it in model_path. The Depth Net refers to MegaDepth and we use the depth pre-trained model Password:mzqa or Google Drive from them.
Run the following commands:
python3 test.py --clean_img_path clean_img_path --depth_img_path depth_img_path --underwater_path underwater_path --fid_gt_path fid_gt_path --model_path model_path
The rusults are saved in ./out/
The computation and inferencing runtime of rendering is 138.13GMac/0.026s when the image size is 1024×1024.
@article{ye2022underwater,
title={Underwater Light Field Retention: Neural Rendering for Underwater Imaging},
author={Ye, Tian and Chen, Sixiang and Liu, Yun and Chen, Erkang and Ye, Yi and Li, Yuche},
journal={arXiv preprint arXiv:2203.11006},
year={2022}
}
If you have any questions, please contact the email [email protected] or [email protected]