From Monochrome to Color: Efficient Techniques for Realistic Video Colorization
Main Article Content
Abstract
Colorizing grayscale videos is a challenging task that involves adding colors to monochromatic videos to make them appear as natural and realistic as possible, despite the absence of information about the original color distribution. This task is crucial for applications such as restoring historical videos and creating media content requiring realistic and accurate colorization. However, existing methods often face issues such as poor temporal stability, color distortions, and the need for extensive post processing under certain conditions. To address these challenges, this study proposes a novel approach comprising a pre-processing network and a source-reference network trained in an end-to-end manner. The pre-processing network employs an encoder-decoder architecture enhanced with temporal convolutions and skip connections, enabling it to improve video quality, adapt to resolution changes, and leverage batch normalization (BN) and exponential linear unit (ELU). The source-reference network incorporates an encoder for reference image processing and a fusion module with residual blocks to combine feature maps through a source-reference attention mechanism. The final output is generated in the Lab color space and converted to the RGB format, ensuring high-quality video colorization with enhanced temporal stability. Experimental evaluations demonstrate that the proposed model achieves significant improvements in color realism and temporal stability, with a PSNR of 37.89 and an SSIM of 0.999982—which surpasses those of most state-of-the-art methods. These results confirm the effectiveness and applicability of the proposed method for video colorization tasks, making it a robust solution in the domain.
Article Details
Issue
Section
This work is licensed under a Creative Commons Attribution 4.0 International License.
How to Cite
References
T. Wijesinghe, C. Abeysinghe, C. Wijayakoon, L. Jayathilake, and U. Thayasivam, "FlowChroma - A Deep Recurrent Neural Network for Video Colorization," in Image Analysis and Recognition. ICIAR 2020. Lecture Notes in Computer Science, vol. 12131, A. Campilho, F. Karray, and Z. Wang, Eds., Cham: Springer, 2020, pp. 16–29, doi: 10.1007/978-3-030-50347-5_2.
S. Huang, X. Jin, Q. Jiang, and L. Liu, "Deep Learning for Image Colorization: Current and Future Prospects," Engineering Applications of Artificial Intelligence, vol. 114, pp. 105006, 2022.
C. Lei and Q. Chen, "Fully Automatic Video Colorization with Self-Regularization and Diversity," in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, 2019, pp. 3753–3761.
E. Saleem and N. K. El Abbadi, "Auto Colorization of A Gray-Scale Image Using YCbCr Color Space," Iraqi J. Sci., vol. 61, no. 12, pp. 3379–3386, 2020.
M. E. A-Monem and T. Z. Hammood, "Video Colorization Methods: a Survey," Iraqi J. Sci., vol. 61, no. 3, pp. 675–686, 2020.
A. Alsajri, H. A. Salman, and A. Steiti, “Generative Models in Natural Language Processing: A Comparative Study of ChatGPT and Gemini”, Babylonian Journal of Artificial Intelligence, vol. 2024, pp. 134–145, Nov. 2024.
K. Tejashwini, M. Soumya, P. Kalyani, G. Deepika, and C. Sujatha, "Colorization of Black and White Images and Video Using CNN," in Advances in Computational Intelligence and Its Applications, CRC Press, 2024, pp. 296–306.
G. Zhang, X. Hong, Y. Liu, Y. Qian, and X. Cai, "Video Colorization Based on Variational Autoencoder," Electronics, vol. 13, no. 12, 2024.
Y. Yang, J. Jin, Y. Huang, K. Guo, and X. Xu, "Reference-based Video Colorization with AB Chrominance Point and Temporal Propagation," in Proc. 16th Int. Conf. Machine Learning and Computing (ICMLC '24), 2024, pp. 340–346.
S. Chen et al., "Exemplar-based video colorization with long-term spatiotemporal dependency," Knowledge-Based Systems, vol. 284, pp. 111240, 2024.
O. M. Hammad, I. Smaoui, A. Fakhfakh, and M. M. Hashim, “Recent advances in digital image masking techniques Future challenges and trends: a review”, SHIFRA, vol. 2024, pp. 67–73, May 2024, doi: 10.70470/SHIFRA/2024/008.
F. Dogan and I. Turkoglu, "Comparison of Deep Learning Models in Terms of Multiple Object Detection on Satellite Images," J. Eng. Res., vol. 10, no. 3A, 2022.
S. Park and Y. G. Shin, "Generative residual block for image generation," Applied Intelligence, vol. 52, pp. 7808–7817, 2022, doi: 10.1007/s10489-021-02858-6.
V. Thakkar et al., "Batch Normalization in Convolutional Neural Networks — A comparative study with CIFAR-10 data," in 2018 5th Int. Conf. Emerging Applications of Information Technology (EAIT), 2018, pp. 1–5.
A. K. Dubey and V. Jain, "Comparative Study of Convolution Neural Network’s Relu and Leaky-Relu Activation Functions," in Applications of Computing, Automation and Wireless Systems in Electrical Engineering. Lecture Notes in Electrical Engineering, vol. 553, S. Mishra, Y. Sood, and A. Tomar, Eds., Springer, 2019, doi: 10.1007/978-981-13-6772-4_76.
C. Ma, X. Mu, and D. Sha, "Multilayers Feature Fusion of Convolutional Neural Network for the Scene Classification of Remote Sensing," IEEE Access, vol. 7, pp. 121685–121694, 2019.
D. A. Clevert, T. Unterthiner, and S. Hochreiter, "Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs)," arXiv preprint, arXiv:1511.07289, 2016.
D. Wang, J. Wang, W. Li, and P. Guan, "T-CNN: Trilinear Convolutional Neural Networks Model for Visual Detection of Plant Diseases," Comput. Electron. Agric., vol. 190, pp. 106468, 2021.
A. Radford et al., "Learning Transferable Visual Models from Natural Language Supervision," in Proc. Int. Conf. Machine Learning, PMLR, 2021, pp. 8748–8763.
H. Tan and M. Bansal, "Lxmert: Learning Cross-Modality Encoder Representations from Transformers," arXiv preprint, arXiv:1908.07490, 2019.
N. K. El Abbadi and E. Saleem, "Gray Image Colorization Based on General Singular Value Decomposition and YCbCr Color Space," Kuwait J. Sci., vol. 46, no. 4, 2019.
A. S. Ketab and N. K. El Abbadi, "LipPrint: Using Lip Movements as a Silent Password," Mesopotamian J. Cybersecurity, vol. 4, no. 2, pp. 74–87, 2024.
B. Xie, S. Yang, Y. Li, Y. Xu, and G. Wang, "Color Transfer Based on Cartoon Decomposition and Efficient Illuminant Estimation," in Int. Conf. Signal Processing and Communication Technology (SPCT 2022), SPIE, 2023, p. 1261502.
Z. Sheng, H.-L. Shen, B. Yao, and H. Zhang, "Guided Colorization Using Mono-Color Image Pairs," IEEE Trans. Image Process., vol. 32, pp. 905–920, 2023.
Z. Ke, Y. Liu, L. Zhu, N. Zhao, and R. W. H. Lau, "Neural Preset for Color Style Transfer," in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, 2023, pp. 14173–14182.
V. U and G. V, "Superpixel based Image Colorization with Automated Reference Image Selection," in 2023 IEEE Int. Students' Conf. Electrical, Electronics and Computer Science (SCEECS), Bhopal, India, 2023, pp. 1–6, doi: 10.1109/SCEECS57921.2023.10061822.
X. Zhu and E. Hu, "A Local Color Transfer Method Based on Optimal Transmission," in 2023 5th Int. Conf. Artificial Intelligence and Computer Science (AICS 2023), 2023, pp. 483–492.
C. Lv, D. Zhang, S. Geng, Z. Wu, and H. Huang, "Color Transfer for Images: A Survey," ACM Trans. Multimedia Comput. Commun. Appl., vol. 20, 2023, doi: 10.1145/3635152.
C. Lv and D. Zhang, "Palette-based Color Transfer between Images," arXiv preprint, arXiv:2405.08263, 2024.
S. Bao, Y. Zhao, Y. Ji, et al., "Color Transfer Method Based on Saliency Features for Color Images," *Opt. Rev.*, vol. 31, pp. 369–382, 2024, doi: 10.1007/s10043-024-00888-2.