Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 146 tok/s
Gemini 2.5 Pro 43 tok/s Pro
GPT-5 Medium 32 tok/s Pro
GPT-5 High 26 tok/s Pro
GPT-4o 37 tok/s Pro
Kimi K2 189 tok/s Pro
GPT OSS 120B 434 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

SEM-CS: Semantic CLIPStyler for Text-Based Image Style Transfer (2303.06334v1)

Published 11 Mar 2023 in cs.CV

Abstract: CLIPStyler demonstrated image style transfer with realistic textures using only the style text description (instead of requiring a reference style image). However, the ground semantics of objects in style transfer output is lost due to style spillover on salient and background objects (content mismatch) or over-stylization. To solve this, we propose Semantic CLIPStyler (Sem-CS) that performs semantic style transfer. Sem-CS first segments the content image into salient and non-salient objects and then transfers artistic style based on a given style text description. The semantic style transfer is achieved using global foreground loss (for salient objects) and global background loss (for non-salient objects). Our empirical results, including DISTS, NIMA and user study scores, show that our proposed framework yields superior qualitative and quantitative performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. “Artistic style transfer with internal-external learning and contrastive learning,” in Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, Eds. 2021, vol. 34, pp. 26561–26573, Curran Associates, Inc.
  2. “Universal style transfer via feature transforms,” Advances in neural information processing systems, vol. 30, 2017.
  3. “Arbitrary style transfer with style-attentional networks,” in proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 5880–5888.
  4. “Learning linear transformations for fast arbitrary style transfer,” ArXiv, vol. abs/1808.04537, 2018.
  5. “The contextual loss for image transformation with non-aligned data,” in Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part XIV. 2018, vol. 11218 of Lecture Notes in Computer Science, pp. 800–815, Springer.
  6. “Cross-modal style transfer,” in 2018 25th IEEE International Conference on Image Processing (ICIP), 2018, pp. 2157–2161.
  7. “Photorealistic style transfer with screened poisson equation,” in Proceedings of the British Machine Vision Conference (BMVC), Gabriel Brostow Tae-Kyun Kim, Stefanos Zafeiriou and Krystian Mikolajczyk, Eds. September 2017, pp. 153.1–153.12, BMVA Press.
  8. “Style transfer using optimal transport via wasserstein distance,” in 2022 IEEE International Conference on Image Processing (ICIP), 2022, pp. 2681–2685.
  9. “Image style transfer using convolutional neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2414–2423.
  10. “A patch-based approach for artistic style transfer via constrained multi-scale image matching,” in 2022 IEEE International Conference on Image Processing (ICIP). IEEE, 2022, pp. 3490–3494.
  11. “Clipstyler: Image style transfer with a single text condition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2022, pp. 18062–18071.
  12. “Deep photo style transfer,” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6997–7005, 2017.
  13. “Deepobjstyle: Deep object-based photo style transfer,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2021, pp. 711–719.
  14. “Generative artisan: A semantic-aware and controllable clipstyler,” .
  15. “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  16. “Deep spectral methods: A surprisingly strong baseline for unsupervised semantic segmentation and localization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 8364–8375.
  17. “Image quality assessment: Unifying structure and texture similarity,” IEEE transactions on pattern analysis and machine intelligence, vol. 44, no. 5, pp. 2567–2581, 2020.
  18. “Nima: Neural image assessment,” IEEE transactions on image processing, vol. 27, no. 8, pp. 3998–4011, 2018.
  19. “Rethinking and improving the robustness of image style transfer,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 124–133.
  20. “Stylegan-nada: Clip-guided domain adaptation of image generators,” ACM Transactions on Graphics (TOG), vol. 41, no. 4, pp. 1–13, 2022.
  21. “Styleclip: Text-driven manipulation of stylegan imagery,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2085–2094.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.