Previous language-guided colorization methods have obvious flaws, like color spillover and inconsistent themes. The root lies in poor info transfer and feature processing in models. Redundant data in transfer lowers efficiency and accuracy of feature representation. Existing feature-fusion modules also can’t integrate diverse features well, harming colorization. We propose a new approach with two key modules. The Reconstruction Convolution Module (RCM) cuts costs by removing redundant info and strengthens feature representation. The Cross-Modal Color Aligner Module (CMCAM) uses multi-scale features to align color and grayscale features precisely, improving semantic understanding. Experiments show our method outperforms state-of-the-art, achieving better results and robustness in high-quality color image generation.
Language-Guided Colorization: Reconvolution and Cross-Modal Align
Yutong Gao,Hao Liu,Xuan Liu,Zheng Liu,Chaomurilige,Shan Jiang
Published 2025 in 2025 IEEE/CIC International Conference on Communications in China (ICCC)
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
2025 IEEE/CIC International Conference on Communications in China (ICCC)
- Publication date
2025-08-10
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-36 of 36 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1