Skip to content

Latest commit

 

History

History
112 lines (88 loc) · 5.2 KB

README.md

File metadata and controls

112 lines (88 loc) · 5.2 KB

Hint-based-Image-Colorization-Using-MSAU-Net

2022-1, Gachon University, Department of Software, Computer Vision Term Project
This Project code ranked 1st place in Colorization Challenge competitions in ComputerVision lecture ( From CVIP Lab in Gachon Univ. ).

image


Hint-based Image Colorization

Hint-based colorization reconstructs complete colorful images by propagating the color hint given by users. For the hint-based colorization task, each solution should first convert the provided RGB image to LAB image. The inputs of the model should be an ”L” image (i.e. grayscale image or intensity image) and a color hint map. The output will be a predicted ”AB” image that will be concatenated with the input L image to produce the final colorized image. An overview is shown below.

flow example

[example : Real-Time User-Guided Image Colorization with Learned Deep Priors](https://richzhang.github.io/InteractiveColorization)




MSAU-Net Model Architecture

  • Multi-scale Attention(Spatial & Channel) U-Net

network_overview

  • Multi Scale Convolution Block

image

  • Skip-connection with Spatial & Channel Attention

image

Loss Function

image



Training dataset

  • Dataset Size : 10000 / 2000
  • Using Place365 and ImageNet datasets
  • Image size : 256 x 256

skip-connection



Using Hyper-parameter

  • Batch size : 16
  • Optimizer : Adam (betas = (0.5, 0.999))
  • Learning rate : 5e-4 , Decrease 5e-4/70 every time the validation loss falls from PSNR 33
  • Develop Environment RTX Titan in CUDA 10.1


Experiment & Result

  • Model Robust Study
  • Validation is carried out with additional datasets to assess the possibility of overfitting specific data
  • Conduct performance evaluation with 2017 coco test dataset
  • Colorization was performed based on the hint ratio 0.5 %
Model PSNR SSIM
Ours 34.0958 0.9679

  • Ablation Study
  • It was experimentally confirmed how much the performance of additional components affected
Model PSNR SSIM
Unet 30.9374 0.9647
Attention Unet 31.7498 0.9683
Ours 31.7905 0.9685

Test dataset Result

ex1 ex2



How to use model

You can download the optimal weight pth from the link below and run this model.

https://drive.google.com/file/d/1IowH8LEjdWut9chW8cxtM8d3oHGtfJ1d/view?usp=sharing



Reference paper

  1. Ronneberger, Olaf, Philipp Fischer, and Thomas Brox. "U-net: Convolutional networks for biomedical image segmentation." International Conference on Medical image computing and computer-assisted intervention. Springer, Cham, 2015.
  2. Zhao, Hang, et al. "Loss functions for image restoration with neural networks." IEEE Transactions on computational imaging 3.1 (2016): 47-57
  3. Zhang, Richard, Phillip Isola, and Alexei A. Efros. "Colorful image colorization." European conference on computer vision. Springer, Cham, 2016.
  4. Zhang, Richard, et al. "Real-time user-guided image colorization with learned deep priors." arXiv preprint arXiv:1705.02999 (2017)
  5. Woo, Sanghyun, et al. "Cbam: Convolutional block attention module." Proceedings of the European conference on computer vision (ECCV). 2018
  6. Xiao, Yi, et al. "Interactive deep colorization using simultaneous global and local inputs." ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019.
  7. Khanh, Trinh Le Ba, et al. "Enhancing U-Net with spatial-channel attention gate for abnormal tissue segmentation in medical imaging." Applied Sciences 10.17 (2020): 5729.
  8. Su, Run, et al. "MSU-Net: Multi-scale U-Net for 2D medical image segmentation." Frontiers in Genetics 12 (2021): 140.



License

This project belongs to computer vision Team 4, Gachon Univ. (2022-1)