|
ShodhKosh: Journal of Visual and Performing ArtsISSN (Online): 2582-7472
AI for Preserving Indigenous Folk Art Patterns Muthukumaran Malarvel 1 1 Department
of Computer Science and Engineering, Aarupadai Veedu Institute of Technology,
Vinayaka Mission’s Research Foundation (DU). Chennai, Tamil Nadu, India 2 Centre
of Research Impact and Outcome, Chitkara University, Rajpura- 140417, Punjab,
India 3 Assistant Professor, School of Business Management, Noida International
University, India 4 Department of Instrumentation and Control Engineering Vishwakarma
Institute of Technology, Pune, Maharashtra, 411037, India 5 Professor, Department of Computer Science and Engineering, Sathyabama
Institute of Science and Technology, Chennai, Tamil Nadu, India 6 Assistant Professor, Department of Computer Science and Information
Technology, Siksha 'O' Anusandhan (Deemed to be University), Bhubaneswar,
Odisha, India
1. INTRODUCTION Some of the earliest forms of visual language still extant in human civilization, indigenous folk art is seen as an expression of social beliefs, mythology, knowledge of craft, and collective memory of various communities. Every motif, stroke, color scheme and geometric composition has underlying strata of encoded cultural significance that have been passed on through generations of oral teaching, apprenticeship and practice. Nevertheless, globalization, industrial production, depopulation and slow erosion of skilled crafts have been challenging many of these traditions to unprecedented levels. With younger generations moving on to urban work and commercial forces of design homogenizing aesthetics, various folk art possibilities are in danger of disappearance, at least in part. This cultural susceptibility should be addressed with a system of digital preservation initiatives that have the potential to preserve, interpret, and rejuvenate indigenous motifs with complete faithfulness and contextual acuity. The use of Artificial Intelligence (AI) as the visual recognition tool, pattern analysis, and generative modeling tool has become a potent force that can be used to revolutionize the work of cultural heritage preservation. It is now possible to teach AI systems complex relationships between colors, textures, brushstroke patterns, symbolic structures, and regional styles with the advent of deep learning architectures, including Convolutional Neural Networks (CNNs), Vision Transformers (ViTs), and multimodal generative models Khan (2024). Such abilities allow automatic recognition of the folk art patterns and style features grouping, as well as restoration of the damaged or fragmented paintings. In addition, AI can support digital archiving, as it helps museums, researchers, and cultural institutions to catalogue and search large collections of images, thus saving on manual labour and enhancing the consistency of metadata Leshkevich and Motozhanets (2022). In spite of them, the use of AI in preserving folk art has not been fully investigated in comparison to other fields like natural image classification, medical imaging, and commercial design. Folk art poses special problems: there is a lot of stylistic heterogeneity, little or unbalanced dataset, differences in artisan styles, and annotating such works culturally. Motifs in turn represent non-Western visual grammars which mainstream computer vision models are not optimized to. Also, any technological intervention should be informed by ethical issues like proper respect of community ownership, cultural misappropriation and meaningful involvement of artisans Marchello et al. (2023). Thus, AI-based conservation of indigenous folk art patterns needs not just technical quality but also systems that are based on cultural sensitivity and heritage custodianship. The study fills these gaps by suggesting an all-encompassing AI-powered approach to the capture, analysis, and reconstruction of the patterns of indigenous folk art. The research starts with the creation of the culturally sound dataset that is gathered in museums, digital archives, communities of artisans, field surveys, and heritage facilities Ghaith and Hutson (2024). The annotation protocols are well created in order to record the geometry of motifs, their chromatic style, and the shape of strokes, references to materials, and identities of regions. The metadata standards are used to guarantee interoperability with larger databases within the cultural heritage. The high-level feature extraction methods, which use CNNs, ViTs, and style-embedding networks, permit the description of motifs on the levels of texture, symbolism, and artistic variation Zhong et al. (2021). Generative models aid in the reconstruction of motifs and pattern reproduction without losing the stylistic authenticity. 2. Literature Review 2.1. Traditional approaches to documenting folk art motifs Historical, traditional documentation of the indigenous folk art motifs have traditionally been implemented by use of manual, community-based and ethnographic methods. Field visits were usually made by researchers, anthropologists, and art historians to artisan communities, where it was documented in the form of sketches, photographs, interviews and handwritten catalogues. These methods focused more on the richness of situations - to capture not only motifs but also tales, ritual meanings, methods of production, and the process of teaching across generations Song (2023). The physical objects of folk art, together with the metadata describing them have frequently been held by museums and cultural archives. Much of this documentation is however basically scattered, not complete, or not in standard formats, which complicates large-scale comparative analysis. The artisans themselves have been very fundamental in passing of the motifs in terms of apprenticeship models, where knowledge is incarnated and not in writings Xu et al. (2024). This verbal and practical conveyance maintains the originality of motifs yet it is susceptible to interferences by urban migration, diminishing interest in young generation and disappearance of traditional means of livelihood. There have been print repositories including community books, regional art catalogues, and pattern manuals, which have tried to standardize motifs, but many are lost or not available to the general public or are of a low-resolution image Ye (2022). But on the whole, traditional forms of documentation have a rich and cultural depth, but cannot meet the modern demands of scalability, international accessibility and digital storage especially as factors of cultural homogenization and socio-economic transformation continue to threaten folk art traditions. 2.2. Computer Vision Applications in Cultural Artifact Recognition Computer vision has also been used in a growing body of applications to the recognition, classification, and retrieval of cultural objects, and has been used to perform automated analysis of historical textiles, pottery, manuscripts, architecture, and sculptural motifs. Often-used techniques, such as feature detection (SIFT, HOG), texture analysis, and shape modeling, have been heavily used to identify patterns, categorize objects by area or time and identify the similarities in styles between collections Chen and García (2022). CNN-based models like ResNet, Inception and EfficientNet have (with the introduction of deep learning) significantly advanced the accuracy of recognition through learning hierarchical visual feature representations, including strokes, symmetry structures and decoration, which are typical of traditional art forms. Later, multimodal embeddings and contrastive learning models (e.g., CLIP) alongside Vision Transformers (ViTs) have increased the abilities to make comparisons between images based on stylistic domains and match motifs on the text description basis. It has been used in the detection of forgeries, reconstruction of damaged artifacts, colorization of historical images, and motifs extracted out of a complicated background. A number of digital humanities projects have embraced the use of these tools in heritage conservation, including automatic transcription of manuscripts, search of textile designs, and categorization of archaeological shapes Amany et al. (2022). Nonetheless, the area of folk art recognition is underresearched because of the lack of data, large intra-class stylistic differences, and annotations that require cultural context. Altogether, computer vision provides good opportunities to scale-up and objectively analyze cultural artifact, yet the application to the native folk art traditions needs adaptation Morlotti et al. (2024). 2.3. AI-Driven Cultural Preservation Frameworks Globally Artificial intelligence-based cultural protection efforts have been on the rise around the globe with governments, museums, and research centers exploring solutions to protect intangible and tangible heritage on a large scale. Google Arts and Culture is one project using deep learning to digitize high-resolution artworks, tag artworks automatically, and explore artworks with an interactive visual interface. Programs funded by UNESCO have investigated machine learning applications in preserving linguistic, retrieving threats to music, and storytelling on the internet Bosco et al. (2021). Japan In Japan, AIs have been used to categorize Ukiyo-e woodblock prints, whereas in China, the deep learning models helped with restoring ancient murals and analysing historical calligraphy. CNNs and GANs are applied in the cultural heritage digitization process in Europe to restore damaged objects, as well as recreate lost motifs and improve archival images. The learning and generation of culturally inspired artworks, re-creation of lost patterns, and the provision of educational interfaces with the introduction of heritage motifs to younger audiences have also been implemented using generative models, including GANs and diffusion networks Janas et al. (2022). Multimodal AI systems combine visual, written, and historical metadata to have a more thorough culture with the ability to find patterns and compare styles across centuries. Along with these improvements, there are still these issues as cultural appropriation, imbalanced access to technology, and a lack of engagement with indigenous communities Trigona et al. (2022). Ethical constructs focus on participative design, rights to own cultural data, data sovereignty, and visible AI usage. Table 1 is a summary of AI-based solutions towards the preservation of cultural heritage and folk art. Together, these initiatives around the world are illustrative of the cultural preservation power of AI, but the necessity to have community-based technologies with cultural foundations. Table 1
3. Dataset Development and Collection 3.1. Sourcing images from museums, archives, artisans, and field surveys The process of developing the dataset starts with the massive sourcing of indigenous folk art images in the various cultural repositories. Museums and heritage archives offer scans of preserved works of high-quality, which are controlled in lighting and provenance. The digitization teams in the archives coordinate to access image collections, exhibition documentation papers, and indexed collections. In parallel, the work with local craftsmen allows getting the most up-to-date motifs in the form of high-resolution photos of works under way, murals, and textile designs. In Figure 1, the flow of acquiring multisource images in indigenous folk art data sets is illustrated. The field surveys in the rural and tribal areas help to note the regional differences in colors, styles, and symbols which are not always represented in institutional sources. Figure 1 |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Table 2 Comparative Evaluation of AI Models for Folk Art Motif Recognition |
||||
|
Model Type |
Accuracy (%) |
Precision (%) |
Recall (%) |
F1-Score (%) |
|
ResNet50 (CNN) |
90.3 |
91.2 |
89.6 |
90.4 |
|
EfficientNet-B4 |
92.1 |
93.3 |
91.4 |
92.3 |
|
Vision Transformer (ViT-B16) |
94.8 |
95.5 |
94.1 |
94.7 |
|
CNN–ViT Hybrid |
96.2 |
97 |
95.8 |
96.3 |
Table 2 introduces the comparative analysis of different AI structures used in the area of indigenous folk art motif recognition. The findings demonstrate that deep learning models have high performance in all the measured metrics, and they gain significantly in performance as the architecture changes continuously in the way of conventional CNNs to hybrid systems with transformers. Figure 3 presents model performance analysis using major metrics of evaluation. The ResNet50 architecture demonstrated 90.3% accuracy, which had a good performance with simple geometric motifs but had minor weakness with the multi-layered and complex patterns.
Figure 3

Figure 3 Analysis of Deep Models Across Key Performance
Metrics
The recognition rate of EfficientNet-B4 rose to 92.1, which is better than that of competitors and shows better precision and recall with feature scaling and efficient use of parameters. This resulted in a great jump of 94.8% accuracy, which is able to capture long-range dependencies and contextual relationships between complex motifs through the use of self-attention mechanism (ViT-B16). Figure 4 presents the visualization of the accuracy improvement with the growing model architecture.
Figure 4

Figure 4 Visualization of Accuracy Improvements Across Model
Architectures
The CNNViT hybrid model achieved the best score of 96.2, as it was able to merge the sensitivity of CNN features to local cases and the global contextual diversity of transformers. In general, the hybrid method is characterized by strong generalization, greater motif separation, and better flexibility to culturally diverse folk art system datasets, and is therefore the most effective design to use in preservation and recognition of patterns.
5.2. Quality of Reconstructed and Generated Folk Art Patterns
The reconstructs created by AI demonstrated the extraordinary ability to maintain the stylistic consistency and cultural values. Structural similarity Index (SSIM = 0.935) and Fréchet Inception Distance (FID = 21.6) quantitative assessments demonstrated that the quality of reproduction was of high quality and similar to artisan originals. GAN-based models as well as diffusion models were able to recover missing motifs, reproduce faded pigments, and create variations of the image that did not alter the sayings of the symbolism in cultures. The rating of the reconstructed outputs was 92 and 89 percent respectively by the expert reviewers when it comes to visual and cultural authenticity respectively. The texture and colour gradients were very close to the transition of natural pigments and geometric harmony was preserved. These results show that AI can not just imitate but also rejuvenate endangered visual languages to promote their preservation, online learning, and long-term inheritance of the heritage of indigenous folk art.
Table 3
|
Table 3 Quantitative Evaluation of Reconstructed and Generated Folk Art Patterns |
|||||
|
Model Type |
SSIM (%) |
PSNR (dB) |
FID Score (↓) |
Color Fidelity (%) |
Texture Consistency (%) |
|
GAN-Based Reconstruction |
92.1 |
28.6 |
25.4 |
91.2 |
88.6 |
|
Diffusion Model |
93.5 |
30.8 |
21.6 |
93.7 |
92.3 |
|
CNN Autoencoder |
90.3 |
27.1 |
29.2 |
89.6 |
87.1 |
|
ViT–GAN Hybrid |
94.7 |
31.5 |
19.8 |
95.4 |
94.2 |
Table 3 shows the quantitative analysis of AI models that could be applied to reconstruct and generate indigenous patterns in folk art, with reference to visual fidelity and structural accuracy. The ViTGAN hybrid model was the best performing in the comparison with an SSIM of 94.7, PSNR of 31.5 dB, and the lowest FID of 19.8 meaning high reconstruction realism and low perceptual difference with original artworks. Fig. 5 depicts visualization of the image reconstruction quality at models.
Figure 5

Figure 5
Visualization of
Image Reconstruction Quality Across Models
It was able to retain tiny-grained details like pigment transitions, uniformity of texture, and motif symmetry. Diffusion model came next, and the results had a higher quality and more tonal gradient and continuity of colors (SSIM 93.5%). GAN-based reconstruction was also successful (SSIM 92.1%), and had subtle textual discrepancies in complicated motifs. Figure 6 presents the improvement of SSIM in various reconstructions of images with different models.
Figure 6

Figure 6 Analysis of SSIM Improvements Across Reconstruction
Models
On the other hand, the CNN autoencoder, despite being computationally efficient, produced relatively lower perceptual scores, because of the lack of contextual skill in the interpretation of symbolic patterns. On the whole, transformer-integrated generative systems were more effective than other ones, guaranteeing better authenticity, color fidelity and structural integrity- showing how AI can digitally conserve and recover and re-create traditional folk art motifs in a culturally competent manner.
6. Conclusion
The analysis illustrates that the ethical integration of the Artificial Intelligence into the cultural-directed practice can be used as a revolutionary tool to safeguard the indigenous folk art patterns. The combination of the fields of computer vision, cultural analytics, and heritage studies helps to bring the proposed framework a step higher and enhance the digital protection of artistic traditions which are otherwise on the verge of extinction. The systematic gathering of images of museums, archives, artisans and field surveys provides a culturally based dataset rich in annotations of motifs, materials, and local styles, which are produced through the research. Vision transformers and Vision CNNs can sufficiently be trained to grasp visual hierarchies and symbolic interrelations of indigenous designs resulting in high accuracy in motif recognition and pattern classification. Equally important, the addition of style embeddings and symbolic motif encoding adds an additional ability of AI to detect patterns as well as interpret cultural patterns. The system is not only able to reconstruct damaged or incomplete works of art with an amazing level of fidelity but it is also able to produce new designs that are adherent to the old aesthetic rules. Technical stability is measured by the performance indicators like SSIM and FID and cultural appropriateness is enhanced and artistic integrity is emphasized through the assessment of experts. In addition to the computational performance, the study highlights the ethical obligations, that is, providing community engagement, data stewardship, and artisanal ownership. It imagines AI as a creative partner of human in a co-creation process and does not displace human art. This research has the potential to fill a replicable model of heritage conservation in the world, by reconciling cultural heritage and digital innovation. Finally, AI-enhanced maintenance of native folk art provides an opportunity to preserve creative legacies so that the future generations could experience indigenous motifs, colors, and symbols that characterize the overall cultural image of humanity.
CONFLICT OF INTERESTS
None.
ACKNOWLEDGMENTS
None.
REFERENCES
Amany, M. K., Gehad, G. M., Niveen, K. F., and Ammar, W. B. (2022). Conservation of an Oil Painting from the Beginning of 20th Century. Редакционная Коллегия, 27–31, 315.
Arafin, P., Billah, A. M., and Issa, A. (2024). Deep Learning-Based Concrete Defects Classification and Detection Using Semantic Segmentation. Structural Health Monitoring, 23, 383–409. https://doi.org/10.1177/14759217231168212
Bosco, E., Suiker, A. S. J., and Fleck, N. A. (2021). Moisture-Induced Cracking in a Flexural Bilayer with Application to Historical Paintings. Theoretical and Applied Fracture Mechanics, 112, 102779. https://doi.org/10.1016/j.tafmec.2020.102779
Chen, Y., and García, F. L. D. B. (2022). Análisis Constructivo y Reconstrucción Digital 3D de Las Ruinas del Antiguo Palacio de Verano de Pekín (Yuanmingyuan): El Pabellón de la Paz Universal (Wanfanganhe). Virtual Archaeology Review, 13, 1–16. https://doi.org/10.4995/var.2022.16523
Ghaith, K., and Hutson, J. (2024). A Qualitative Study on the Integration of Artificial Intelligence in Cultural Heritage Conservation. Metaverse, 5, 2654. https://doi.org/10.54517/m.v5i2.2654
Hojjati, H., Ho, T. K. K., and Armanfard, N. (2024). Self-Supervised Anomaly Detection in Computer Vision and Beyond: A Survey and Outlook. Neural Networks, 172, 106106. https://doi.org/10.1016/j.neunet.2024.106106
Janas, A., Mecklenburg, M. F., Fuster-López, L., Kozłowski, R., Kékicheff, P., Favier, D., Andersen, C. K., Scharff, M., and Bratasz, Ł. (2022). Shrinkage and Mechanical Properties of Drying Oil Paints. Heritage Science, 10, 181. https://doi.org/10.1186/s40494-022-00814-2
Khan, Z. (2024). AI and Cultural Heritage Preservation in India. International Journal of Cultural Studies and Social Sciences, 20, 131–138.
Leshkevich, T., and Motozhanets, A. (2022). Social Perception of Artificial Intelligence and Digitization of Cultural Heritage: Russian Context. Applied Sciences, 12, 2712. https://doi.org/10.3390/app12052712
Marchello, G., Giovanelli, R., Fontana, E., Cannella, F., and Traviglia, A. (2023). Cultural Heritage Digital Preservation Through AI-Driven Robotics. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 48, 995–1000. https://doi.org/10.5194/isprs-archives-XLVIII-M-2-2023-995-2023
Morlotti, M., Forlani, F., Saccani, I., and Sansonetti, A. (2024). Evaluation of Enzyme Agarose Gels for Cleaning Complex Substrates in Cultural Heritage. Gels, 10, 14. https://doi.org/10.3390/gels10010014
Shah, N., Froment, E., and Seymour, K. (2023). The Identification, Approaches to Cleaning and Removal of a Lead-Rich Salt Crust from the Surface of an 18th Century Oil Painting. Heritage Science, 11, 80. https://doi.org/10.1186/s40494-023-00925-4
Song, S. (2023). New Era for Dunhuang Culture Unleashed by Digital Technology. International Core Journal of Engineering, 9, 1–14.
Trigona, C., Costa, E., Politi, G., and Gueli, A. M. (2022). IoT-Based Microclimate and Vibration Monitoring of a Painted Canvas on a Wooden Support in the Monastero of Santa Caterina (Palermo, Italy). Sensors, 22, 5097. https://doi.org/10.3390/s22145097
Xu, Z., Yang, Y., Fang, Q., Chen, W., Xu, T., Liu, J., and Wang, Z. (2024). A Comprehensive Dataset for Digital Restoration of Dunhuang Murals. Scientific Data, 11, 1–17. https://doi.org/10.1038/s41597-024-03785-0
Yang, X., Zheng, L., Chen, Y., Feng, J., and Zheng, J. (2023). Recognition of Damage Types of Chinese Gray-Brick Ancient Buildings Based on Machine Learning: Taking the Macau World Heritage Buffer Zone as an Example. Atmosphere, 14, 346. https://doi.org/10.3390/atmos14020346
Ye, J. (2022). The Application of Artificial Intelligence Technologies in Digital Humanities: Applying to Dunhuang Culture Inheritance, Development, and Innovation. Journal of Computer Science and Technology Studies, 4, 31–38. https://doi.org/10.32996/jcsts.2022.4.2.5
Zhong, H., Wang, L., and Zhang, H. (2021). The Application of Virtual Reality Technology in the Digital Preservation of Cultural Heritage. Computer Science and Information Systems, 18, 535–551. https://doi.org/10.2298/CSIS200208009Z
|
|
This work is licensed under a: Creative Commons Attribution 4.0 International License
© ShodhKosh 2025. All Rights Reserved.