EXPLAINABLE AI FOR STYLE INTERPRETATION IN CONTEMPORARY ART USING VISION TRANSFORMERS

Authors

  • Nikil Tiwari Assistant Professor, School of Fine Arts, AAFT University of Media and Arts, Raipur, Chhattisgarh-492001, India
  • Naman Soni Assistant Professor, School of Fine Arts & Design, Noida International University, Noida, Uttar Pradesh, India
  • Rahul Anantrao Padgilwar Assistant Professor, Department of DESH, Vishwakarma Institute of Technology, Pune, Maharashtra-411037, India
  • Dr. Preeti Pandurang Kale Assistant Professor, Department of Electronics and Computer Engineering, CSMSS Chh. Shahu College of Engineering, Chhatrapati Sambhajinagar, Maharashtra, India
  • Dr. Mandeep Kaur Department of Computer Science and Engineering, CT University, Ludhiana, Punjab, India
  • Dr. Vinay Nagalkar Department of Electronics and Telecommunication Engineering, Ajeenkya DY Patil School of Engineering, Pune, Maharashtra, India

DOI:

https://doi.org/10.29121/shodhkosh.v6.i4s.2025.6941

Keywords:

Explainable AI, Vision Transformers, Style Interpretation, Contemporary Art, Attention Maps

Abstract [English]

This study looks into how Explainable AI (XAI) can be used to figure out style in modern art by using Vision Transformers (ViTs). As the need for interpretability in AI-driven art research has grown, models have had to be made that not only work well but also give clear, understandable reasons for the choices they make. We use ViTs, a cutting-edge deep learning system that is known for being very good at classifying images, to look at and figure out what the style aspects of modern art mean. The study aims to find a balance between the need for high-performance AI models and the need for openness in the art world. It will do this by showing how certain aspects of artworks, like colour schemes, structural structures, and brushstroke patterns, affect the overall style. We present a mixed framework that blends the power of Vision Transformers with techniques for explaining things like Grad-CAM and focus maps. This framework helps you see and understand how the model's predictions work. The results show that the model can correctly spot important creative traits and give visual descriptions, which helps people understand different types of art. Additionally, the suggested method is tried on a wide range of modern artworks, showing that it can be used with various types of art. This work has effects beyond just analysing art; it gives managers, artists, and students a useful tool for working with AI systems in a more open way. It also adds to the field of explainable AI by using these methods to study art analysis, which is very biassed and hard to explain.

References

Ahmed, S., Nielsen, I. E., Tripathi, A., Siddiqui, S., Ramachandran, R. P., and Rasool, G. (2023). Transformers in Time-Series Analysis: A Tutorial. Circuits, Systems, and Signal Processing, 42, 7433–7466. https://doi.org/10.1007/s00034-023-02454-8 DOI: https://doi.org/10.1007/s00034-023-02454-8

Alayrac, J.-B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., Lenc, K., Mensch, A., Millican, K., Reynolds, M., and others. (2022). Flamingo: A Visual Language Model for Few-Shot Learning. Advances in Neural Information Processing Systems, 35, 23716–23736.

Burkart, N., and Huber, M. F. (2021). A Survey on the Explainability of Supervised Machine Learning. Journal of Artificial Intelligence Research, 70, 245–317. https://doi.org/10.1613/jair.1.12228 DOI: https://doi.org/10.1613/jair.1.12228

Chan, A., Schneider, M., and Körner, M. (2023). XAI for Early Crop Classification. In Proceedings of the 2023 IEEE International Geoscience and Remote Sensing Symposium (IGARSS) (2657–2660). IEEE. https://doi.org/10.1109/IGARSS52108.2023.10281498 DOI: https://doi.org/10.1109/IGARSS52108.2023.10281498

Colliot, O. (2023). Machine Learning for Brain Disorders. Springer Nature. https://doi.org/10.1007/978-1-0716-3195-9 DOI: https://doi.org/10.1007/978-1-0716-3195-9

Gemini Team, Google. (2023). Gemini: A Family of Highly Capable Multimodal Models (Technical Report). Google.

Jakobsen, T. S. T., Cabello, L., and Søgaard, A. (2023). Being Right for Whose Right Reasons? In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (1033–1054). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.acl-long.59 DOI: https://doi.org/10.18653/v1/2023.acl-long.59

Madan, B. S., Zade, N. J., Lanke, N. P., Pathan, S. S., Ajani, S. N., and Khobragade, P. (2024). Self-Supervised Transformer Networks: Unlocking New Possibilities for Label-Free Data. Panamerican Mathematical Journal, 34(4), 194–210. https://doi.org/10.52783/pmj.v34.i4.1878 DOI: https://doi.org/10.52783/pmj.v34.i4.1878

Marcinkevičs, R., and Vogt, J. E. (2023). Interpretable and Explainable Machine Learning: A Methods-Centric Overview with Concrete Examples. WIREs Data Mining and Knowledge Discovery, 13, e1493. https://doi.org/10.1002/widm.1493 DOI: https://doi.org/10.1002/widm.1493

Thampi, A. (2022). Interpretable AI: Building Explainable Machine Learning Systems. Simon and Schuster.

Vijayakumar, S. (2022). Interpretability in Activation Space Analysis of Transformers: A Focused Survey. In Proceedings of the CIKM 2022 Workshops Co-Located with the 31st ACM International Conference on Information and Knowledge Management.

Yang, Y., Jiao, L., Liu, F., Liu, X., Li, L., Chen, P., and Yang, S. (2023). An Explainable Spatial-Frequency Multiscale Transformer for Remote Sensing Scene Classification. IEEE Transactions on Geoscience and Remote Sensing, 61, 1–15. https://doi.org/10.1109/TGRS.2023.3265361 DOI: https://doi.org/10.1109/TGRS.2023.3265361

Yu, L., and Xiang, W. (2023). X-Pruner: Explainable Pruning for Vision Transformers. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (24355–24363). IEEE/CVF. https://doi.org/10.1109/CVPR52729.2023.02333 DOI: https://doi.org/10.1109/CVPR52729.2023.02333

Zini, J. E., and Awad, M. (2022). On the Explainability of Natural Language Processing Deep Models. ACM Computing Surveys, 55, 1–31. https://doi.org/10.1145/3529755 DOI: https://doi.org/10.1145/3529755

Downloads

Published

2025-12-25

How to Cite

Tiwari, N., Soni, N., Padgilwar, R. A., Kale, P. P., Kaur, M., & Nagalkar, V. (2025). EXPLAINABLE AI FOR STYLE INTERPRETATION IN CONTEMPORARY ART USING VISION TRANSFORMERS. ShodhKosh: Journal of Visual and Performing Arts, 6(4s), 615–625. https://doi.org/10.29121/shodhkosh.v6.i4s.2025.6941