Abstract
Balancing the consistency of style and the integrity of content is the main challenge in arbitrary style transfer domain. Currently, local style details can be effectively captured by attention mechanism but easily produce distorted style patterns and inconsistent content structure. In this paper, we propose a Content Affinity Preserving Arbitrary Style Transfer (CAPAST) framework to ensure style features can be stably integrated into the content structure. Considering the local feature learning ability of CNN and the global feature representation advantage of transformer, a dual encoder is proposed to capture local and global features of images with the combination between transformer and CNN. In addition, a channel and spatially aligned attention (CSAA) is introduced to generate high-quality results by stably fusing style features and content features. In experiments, we demonstrated the superior performance of our method in preventing content structure distortion and maintaining consistency between style and content. Codes are available at https://github.com/miaopashi-zxy/CAPAST.
| Original language | English |
|---|---|
| Journal | ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings |
| Early online date | 2025 |
| DOIs | |
| Publication status | Published - 2025 |
| Event | 2025 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2025 - Hyderabad, India Duration: 6 Apr 2025 → 11 Apr 2025 |
Fingerprint
Dive into the research topics of 'CAPAST: Content Affinity Preserved Arbitrary Style Transfer'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver