|
ShodhKosh: Journal of Visual and Performing ArtsISSN (Online): 2582-7472
Artificial Intelligence as a Creative Collaborator in Contemporary Visual and Performing Arts Pushpalatha P 1 1 Department
of Computer Science, Meenakshi College of Arts and Science, Meenakshi Academy
of Higher Education and Research, India 2 Professor,
Department of English, Meenakshi College of Arts and Science, Meenakshi Academy
of Higher Education and Research, India 3 Department of Computer Science, Meenakshi College of Arts and
Science, Meenakshi Academy of Higher Education and Research, India
4 Central Research Laboratory, Meenakshi Medical College Hospital
& Research Institute, Meenakshi Academy of Higher Education and Research,
India
5 Assistant Professor, Department of Pharmacology, Meenakshi Ammal
Dental College and Hospital, Meenakshi Academy of Higher Education and Research,
India
6 Faculty of
Management, Shinawatra University, Thailand Research Fellow, INTI International
University, Malaysia
1. INTRODUCTION The modern artistic practice is being affected by the development of the artificial intelligence system more and more, reshaping the conceptions, creation, and experience of visual and performing arts. The mathematical models that produce images, music, choreography, and multimedia performances have relegated the status of digital technologies as a passive tool to those that do creative work Ali et al. (2021). Recent AI designs include deep neural networks, generative adversarial networks and diffusion-based models that have shown the capability to analyze large amounts of art and generate new aesthetics outputs that mimic human creativity. This has led to the emergence of new opportunities to the artists, designers, and performers who want to experiment with hybrid forms of creativity through the interaction of human imaginative resources and machine intelligence in dynamic and iterative processes Ali et al. (2021). Figure 1 |
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Gap Analysis |
|||
|
Approach |
Strength |
Limitation |
Research
Gap |
|
GAN-based
Art Generation Black (2018) |
Produces
highly realistic images and paintings |
Limited
interpretability and artistic control |
Need
for interactive generation frameworks |
|
Diffusion-Based
Image Models Blanco (2024), Colton and Wiggins (2012) |
High-quality
visual synthesis |
Computationally
intensive and less interactive |
Real-time
collaborative creativity required |
|
AI
Music and Dance Generation Boden (2004) |
Capable
of generating structured sequences |
Often
operates autonomously without human feedback |
Integration
of performer–AI interaction systems |
|
Rule-Based
Computational Creativity Bowen and Giannini (2019), Phurisikul and Phuakkong
(2025) |
Predictable
and interpretable outputs |
Limited
creativity and stylistic diversity |
Hybrid
human–AI creativity frameworks needed |
3. Framework of AI–Human Creative Collaboration
Human-computer-creative collaboration is a new paradigm of computational systems that are involved in the artistic process together with human artists. In contrast to the old-fashioned digital tools that only help to execute them, modern AI systems are able to produce innovative artistic works, interpret the stylistic patterns and adapt to human input. With these capabilities it is possible to create a co-creative space where artistic concepts are developed through generative interaction between human intuition and machine generative intelligence. In this system of collaboration, artificial intelligence also serves as a dynamic co-worker who can enhance creativity in humans but not to eliminate it. Jadhav (2027)
Figure 2

Figure 2
AI-Augmented
Artistic Co-Creation Workflow
The theoretical background of the creative cooperation between human beings and AI may be explained in terms of a number of co-creation models which explain various stages of interaction between artists and intelligent systems. The former is commonly known as AI-assisted creativity, and AI is depicted as a collaborative system to facilitate certain elements of the artistic process. In this setup, the human artist retains the complete conceptual control and AI systems are available to become image generators, style transfer, sound generators, or automated editors. The innovative idea is also supplied by the human creator, and AI enhances the production process mostly. The second model focuses on the AI-enhanced creativity where human creators and AI systems collaboratively affect a creative product. In this case, artists can offer conceptual guidance, aesthetic limitations or training inputs, with AI models offering variations or surprising design opportunities. With the help of repetitive feedback loops, the artist reviews the outputs produced by machines and also refines parameters, giving rise to a co-creative process, which is increasingly evolutional, involving human judgment and algorithmic experimentation. The most developed model is the symbiotic human-AI co-creation, whereby one engages in real-time interaction in a process of artistic creation. In this type of systems, the artificial intelligence system is constantly adjusted to the actions of the performers, the situations of the environment, or according to the reaction of the audience. Examples are AI-based choreography systems that react to dancer motions or computer-generated music systems that dynamically modify music in the course of live performances. This partnership model makes AI one of the players in the creative ecosystem. Suri et al. (2025)
4. AI Techniques Used in Contemporary Art Creation
Techniques in artificial intelligence have quickly changed the artistic practice by means of machines learning aesthetic patterns and creating other form of visual and performing art. The contemporary creative AI applications are founded on deep learning architectures that can process complex data sets that include images, audio signals, motion sequences, and performance recordings. The techniques enable computational systems to create works of art with a sense of stylistic consistency, structural composition, and expressiveness comparable to that of an artist created by human beings. The Generative Adversarial Network (GAN) can be considered one of the most powerful approaches employed in creative AI systems. GAN architecture is a set of two neural networks which work in a competition: a generator which creates artificial data and a discriminator which rates the similarity of the generated content with the real artistic samples. The generator is trained by progressively learning to create visual output of high quality that looks like an authentic work of art. GAN-based models have been massively applied to produce paintings, portraits, digital images, and stylized visual images. In modern visual art, GANs are able to allow artists to test style synthesis, abstract visual metamorphosis, and algorithmically produced aesthetics. Vasanthan et al. (2019)
Figure 3

Figure 3
AI Techniques Used
in Contemporary Art Creation
The other notable trend is the rise of diffusion-based generative models that have recorded impressive results in terms of generating high-resolution pictures and intricate artistic designs. Diffusion models are the models that involve creating images by means of progressive refinement, where random noise is slowly converted into structure into a visual image. The model can also be used to produce much more stable models of intricate textures, lighting conditions, and stylistic artistic effects, unlike previous generative methods. Computer-generated illustrations, cinematic concept design, animation production and interactive visual installations are also becoming made with diffusion-based systems. Transformer-based architectures are used in sequence-oriented creative tasks, e.g. music composition, choreography generation and storytelling. Transformers have attention modules through which models are able to establish long-term connections in a sequential data. Transformer models are used in music generation to learn rhythmic patterns, harmonic structures and melodic transitions based on massive musical sets. Otherwise, motion-based transformers have the ability to produce dance moves or performance sequences and learn temporal features of motion capture data. The application of reinforcement learning also has an important role in the adaptive artistic environments when the systems can dynamically interact with the human performers or the audience. Reinforcement learning agents are learning to acquire the best creative strategies aiming at feedback in response to their environment and hence they adapt artistic outputs depending on user feedback or performance scenario. In the interactive installations, generative music performances, and responsive stage environments such adaptive systems are especially helpful.
5. System Architecture for AI-Assisted Creative Collaboration
Creative systems involving AI need a well-organized structure involving human-to-human interaction, data processing, machine learning models, and artistic rendering modules. Defining a layered architecture structure not only guarantees that artistic contribution, computational intelligence and generation of output are running in an environment of coordination, but also enables the production of output in an intelligent system that can provide a coordinated environment between artists and intelligent systems to work collaboratively in creative endeavors. Data layer is the backbone of the architecture and it is in charge of gathering and structuring multimodal artistic data. This layer combines various inputs which include visual images, audio records, motion capture records and textual prompts given by artists. Data pre processing modules do normalization, feature extraction and metadata tagging to get the datasets ready to be subjected to machine learning. Rawandale and Kolte (2021)

Live information feeds of sensors, microphones, cameras, or motion capture devices can be included in interactive artistic settings, too. The AI model layer, where computational creativity is applied based on the deep learning frameworks, is placed above the data layer. It is based on generative models including GANs, diffusion networks and architectures based on transformers that learn artistic patterns on the basis of training data. The training, inference and optimization are handled by model orchestration components. In certain systems, the reinforcement learning modules support the adaptability when it comes to responding to actions by the performer or the interactions with the audience. The interface layer facilitates the connection between human designers and AI. Artists deal with the system via graphical interface, digital design tools or through interactive performance platforms. This layer enables artists to give prompting, adjust parameters, and judge generated outputs and direct the creative operation with recurrent feedback. The ability to interact in real time is especially crucial in the performing arts implementation where AI systems can react to human behaviors in real time. Lastly, the output layer is where the artistic artifacts created in the course of the collaborative workflow are created. Visualization tools and rendering engines convert AI-generated data into one of the interpretable artistic forms that can be used in an exhibition or performance or that can be distributed online. Karwande et al. (2024)
6. Experimental Demonstration / Case Study
Creative collaboration aided by AI was experimentally validated by a case study illustrating how generative artificial intelligence can be incorporated in a visual art production process. The objective of the experiment was to investigate how AI systems could help artists to create new visual compositions without losing control over artistic decision-making to human actors. The process of the experiment was composed of dataset preparation, training of the AI model, interactive generation, and qualitative analysis of generated artworks. The data in this experiment was a curated set of about 10,000 digital works drawn out of the available art repositories on the Internet and open creative sources in the open creative data. The dataset covered paintings, digital illustrations, abstract art and concept artwork that represent various representations of visual styles which include impressionism, surrealism and contemporary computerized art. Before training, the images had been standardized to take uniform input sizes and perceptual quality. Preprocessing processes were done on image resizing, normalization and feature extraction. Table 1 summarizes the composition and preprocessing activities performed on the dataset in order to train it.
Table 1
|
Table 1 Experimental Dataset Characteristics |
||||
|
Dataset
Category |
Number
of Images |
Resolution |
Source
Type |
Preprocessing
Applied |
|
Classical
Paintings |
3,500 |
512×512 |
Open
Art Datasets |
Normalization,
Color Balancing |
|
Digital
Illustrations |
4,200 |
512×512 |
Online
Art Repositories |
Resizing,
Feature Extraction |
|
Abstract
Art |
1,800 |
512×512 |
Creative
Commons Collections |
Noise
Reduction |
|
Concept
Art |
500 |
512×512 |
Digital
Artist Archives |
Style
Tagging |
|
Total
Dataset |
10,000 |
512×512 |
Mixed
Sources |
Standardized
Pipeline |
Following the preparation of the datasets, a diffusion-based generative model was followed to produce high-resolution artistic images. The choice of diffusion models was explained by the fact that they are stable when trained and can generate expansive visual images with intricate textures and color gradients. The deep learning framework based on Python and trained on a cloud-based computing platform was used to train the model. The main parameters to be considered in the training of the model are shown in Table 2 where the details about the architecture set up and the environment are provided.
Table 2
|
Table 2 AI Model Training Configuration |
|||
|
Parameter |
Value |
Description |
Implementation
Tool |
|
Model
Type |
Diffusion
Model |
Image
generation architecture |
PyTorch |
|
Training
Epochs |
120 |
Total
training iterations |
Python |
|
Batch
Size |
32 |
Images
per training cycle |
CUDA
GPU |
|
Learning
Rate |
0.0002 |
Optimization
step size |
Adam
Optimizer |
|
Training
Time |
14
Hours |
Total
model training duration |
Cloud
GPU (NVIDIA A100) |
The artists were able to define the visual themes, style choices or composition limits, and the AI model used to produce several candidate artworks. These outputs were then evaluated by the artist and preferred variations were picked upon to further refine it. It enabled artists to experiment with an extensive variety of visual ideas and at the same time retain control over the end artistic product through a series of successive adjustments to the prompts. In order to consider the efficiency of AI-assisted creativity, the process of the traditional human-only artwork creation and the AI-assisted artistic workflow were compared. Some of the measures that were incorporated in the evaluation were time of creating artwork, change in style, and visual quality as perceived. A panel of digital artists and visual designers was engaged in the assessment and scored generated artworks on a set of standardised scoring criteria. The findings of this comparative assessment are represented in Table 3.
Table 3
|
Table 3 AI-Assisted vs Human-Only Artwork Generation Performance |
|||
|
Evaluation
Metric |
Human-Only
Creation |
AI-Assisted
Creation |
Improvement
(%) |
|
Average
Artwork Creation Time |
6.5
hours |
1.8
hours |
72%
Faster |
|
Style
Variation Generated |
3
variants |
18
variants |
500%
Increase |
|
Artist
Satisfaction Score (1–10) |
7.1 |
8.6 |
21% |
|
Visual
Quality Rating (1–10) |
7.4 |
8.3 |
12% |
|
Concept
Exploration Speed |
Medium |
Very
High |
— |
The findings indicate that AI-enhanced creative systems provide a much faster creative exploration stage of visual arts creation as well as boost stylistic variety and experimentation. This collaborative workflow enabled artists to quickly come up with a number of visual options and develop their concepts more effectively compared to manual processes. These findings show how artificial intelligence can become a creative partner that can be used to improve the creative output of artwork without stealing the human author or creative intent.
7. Results and Performance Analysis
The review of the AI creative workflow was performed through the analysis of the numerical data, provided in Table 3, which compares the conventional approach toward artistic production solely done by humans and the one facilitated by AI. Three main areas of creative performance are examined in the analysis: efficiency of production, artistic exploration diversity and perceived artistic quality. All these metrics give the understanding of how effective artificial intelligence can be used as a partner in creative work. There is also a significant decrease in time spent on the creation of an artwork as the time spent was decreased by the AI-assisted workflow, and the average time spent on creating an artwork was shortened to 1.8 hours, which is a significant decrease of the time spent on artwork creation, about 72 percent. This enhancement indicates that generative AI systems are much faster in initial phases of artistic creation, especially in exploring the concept and in prototyping a composition. The conventional creative processes involve artists taking a lot of time in trying out sketches, color combinations and structural layouts before finalizing on an idea. The use of AI systems is able to quickly create various candidate works of art, allowing artists to potentially review and edit ideas significantly more effectively. The other significant change is found in the style variation generated metric. On the human-onlyed mode, the average number of conceptual variations generated was three, on AI- assisted systems, about eighteen visual variants were generated using one prompt. The corresponding growth in conceptual diversity of almost 500 percent points to the ability of generative models to increase the creative search space. Artists can also use algorithmic generation of a wide range of stylistic options, rather than manually generating a number of drafts, which can enable them to find a promising artistic direction more rapidly. The outcomes also show the increase in the satisfaction of artists and the perceived visual quality. The average score of satisfaction among the participants regarding AI-assisted output was 8.6 out of 10, and 7.1 in case of traditional methods. Likewise, the visual quality rating was raised by 7.4 to 8.3, which means that the artists felt that AI-generated works were aesthetically worthy and creatively helpful. These advancements imply that AI tools not only make the process more efficient but also more creative as they create new stylistic opportunities.
Table 4
|
Table 4 Performance Metrics Comparison (Human vs AI-Assisted Workflow) |
|||
|
Performance
Metric |
Human-Only
Workflow |
AI-Assisted
Workflow |
Improvement |
|
Average
Creation Time |
6.5
hours |
1.8
hours |
72%
Faster |
|
Style
Variants Generated |
3 |
18 |
500% |
|
Artist
Satisfaction Score |
7.1 |
8.6 |
21% |
|
Visual
Quality Rating |
7.4 |
8.3 |
12% |
The comparative analysis provided in Table 4 proves the fact that AI-contributing creative pipelines contribute to both efficiency and creative diversity to a large extent. It can be used to devote more time to conceptual development than time spent in manual experimentation and the time spent creating is reduced. Simultaneously, the number of stylistic variants created increases, which enlarges the compositional possibilities that may be experimented in the process of design. Patil (2025)
Table 5
|
Table 5 Artist Evaluation Results (User Study) |
|||
|
Evaluation
Criterion |
Mean
Score (1–10) |
Standard
Deviation |
Interpretation |
|
Visual
Aesthetic Quality |
8.3 |
0.7 |
High
quality outputs |
|
Creativity
and Novelty |
8.5 |
0.6 |
Strong
originality |
|
Ease
of Collaboration with AI |
8.1 |
0.8 |
Effective
interaction |
|
Overall
Satisfaction |
8.6 |
0.5 |
Positive
user experience |
Findings of the user assessment as presented in Table 5 also suggest that the artists saw the AI system as a useful partner instead of a human creativity substitute. The scores of creativity and novelty are high, indicating that generative models bring new visual concepts to the artistic experimentation. The high score on ease of collaboration also demonstrates that the interactive interface is effective in supporting creative workflows that are iterative.
8. Analysis of Interpretation
Experimental findings support the fact that AI-based creative systems may achieve a significant level of increase in artistic production efficiency and diversity without sacrificing the fundamental significance of human creativity. The speed of the decrease in art creation time was observed, which means that the conceptual development stage of visual art can be extremely fast with the help of generative AI models. The system allows the artist to experiment with a variety of stylistic options quickly, so that even in the initial stages of sketching and preliminary design, less effort is necessary than with manual methods.
Figure 4

Figure 4
Comparison of
Artwork Creation Time Between Human-Only and AI-Assisted Workflows
The next crucial finding is the growth in the variety of stylistic tools that have been obtained due to AI-based workflows. Conventional artistic methods are known to give rise to few conceptual options due to the fact that the artist has to hand draw his or her design options. Generative models, in contrast, are capable of generating many stylistic variations of a single prompt, e.g. diffusion networks. It is an option that increases the creative exploration space and enables artists to consider more visual possibilities and then choose the most promising artistic course.
Figure 5

Figure 5
Performance Comparison
Across Key Creative Metrics
It is also possible to note the improvement in perceived quality of artwork and artist satisfaction as shown by the evaluation results. The increased quality scores of AI-assisted output indicate that generative models may acquire complicated visual attributes like composition of texture, color equilibrium, and spatial equilibrium. Such abilities offer an opportunity of inspiration and creative enhancement to artists instead of weighing upon the authorship of humans. By implementing timely adjustment and feedback, artists are in control of the creative process and use computational intelligence to produce alternative visual ideas. The presented interaction model shows that AI technologies can act as the collaborative partners which can further the creative potential of artists and enable new modes of digital artistic experimentation.
9. Conclusion
Artificial intelligence as a concept that is being introduced into modern visual art signifies a paradigm shift in the manner in which visual arts are thought out and created. This study explored the role of AI as a creative partner that could assist artists by their generative modeling, interactive creative process, and creative exploration through adaptive modeling. The suggested framework integrates information preparation, generative models founded on deep learning, human-AI interaction, and rendering to create a creative pipeline in collaboration. Empirical testing with a curated collection of digital artworks proved that the creative process with the help of AI leads to a significant increase in productivity and exploration of ideas. Its findings show that generative AI systems are capable of shortening the time taken to create artwork, improving the level of stylistic diversity, or improving the satisfaction of artists, without preventing human artistic control. By enabling the interaction on a basis of promptness and refinement through iteration, artists can be the focus of the creative process but take advantage of the computational power of the modern AI systems. The results point to the possibilities of AI technologies to broaden the scope of digital creativity and make it possible to create new ways of interaction in art. With the further development of the generative models, human-AI collaboration is probable to have a more significant role in the visual and performing art future. The possibilities of future work include incorporating multimodal generative models, immersive virtual environments and real time collaborative interfaces that in turn enhance the synergy of human creativity and artificial intelligence.
CONFLICT OF INTERESTS
None.
ACKNOWLEDGMENTS
None.
REFERENCES
Ali, S., DiPaola, D., Lee, I., Hong, J., and Breazeal, C. (2021). Exploring Generative Models with Middle School Students. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems ( 1–13). https://doi.org/10.1145/3411764.3445226
Ali, S., DiPaola, D., Lee, I., Sindato, V., Kim, G., Blumofe, R., and Breazeal, C. (2021). Children as Creators, Thinkers and Citizens in an AI-Driven Future. Computers and Education: Artificial Intelligence, 2, Article 100040. https://doi.org/10.1016/j.caeai.2021.100040
Ali, S., Park, H. W., and Breazeal, C. (2020). Can cHildren Emulate a Robotic Non-Player Character’s Figural Creativity? In Proceedings of the Annual Symposium on Computer-Human Interaction in Play (CHI PLAY ’20) ( 499–509). https://doi.org/10.1145/3410404.3414251
Ambrosio, C. (2019). Unsettling Robots and the Future of Art. Science, 365(6448), 38–39. https://doi.org/10.1126/science.aay1956
Anadol, R. (2022). Space in the Mind of a Machine: Immersive Narratives. Architectural Design, 92(3), 28–37. https://doi.org/10.1002/ad.2810
Anadol, R., and Kivrak, P. (2023). Machines That Dream: How AI-Human Collaborations in Art Deepen Audience Engagement. Management and Business Review, 3, 101–107. https://doi.org/10.1177/2694105820230301018
Avlonitou, Z. (2018). Collector and Collection, an Indivisible Unity: The Case of the Collector and Collection of Giorgos Kostakis (Doctoral Dissertation, University of Ioannina).
Bellaiche, L., et al. (2023). Humans Versus AI: Whether and Why we Prefer Human-Created Compared to AI-Created Artwork. Cognitive Research: Principles and Implications, 8, Article 42. https://doi.org/10.1186/s41235-023-00499-6
Benjamin, W. (2008). The Work of Art in the Age of Its Technological Reproducibility, and Other Writings on Media (M. W. Jennings, B. Doherty, and T. Y. Levin, Eds.). Belknap Press. https://doi.org/10.2307/j.ctv1nzfgns
Bhullar, R. (2024). Creative Artificial Intelligence: Exploring the Qualities of Popular AI Art Tools to Determine Effectiveness. NHSJS.
Black, G. (2018). Meeting the Audience Challenge in the “Age of Participation.” Museum Management and Curatorship, 33(4), 302–319. https://doi.org/10.1080/09647775.2018.1469097
Blanco, A. D., et al. (2024, November 11). Exploring the Neural Impact of AI-Generated Art at MoMA: An EEG Study on Refik Anadol’s Unsupervised. OSF Preprints.
Boden, M. A. (2004). The Creative Mind: Myths and Mechanisms (2nd ed.). Routledge. https://doi.org/10.4324/9780203508527
Bowen, J. P., and Giannini, T. (2019). The Digital Future for Museums. In T. Giannini and J. P. Bowen (Eds.), Museums and Digital Culture. Springer. https://doi.org/10.1007/978-3-319-97457-6_28
Cao, Y., et al. (2023). A Comprehensive Survey of AI-Generated Content (AIGC): A History of Generative AI from GAN to ChatGPT. arXiv.
Colton, S., and Wiggins, G. A. (2012). Computational Creativity: The Final Frontier? In Proceedings of the 20th European Conference on Artificial Intelligence (ECAI 2012) ( 21–26).
Jadhav, K. D. (2027). Integrating IoT, Sensors, and Machine Learning for Enhancing Crop Yield and Irrigation Efficiency Systems. In R. R. Janghel, R. Doriya, J. Lachure, and Y. K. Rathore (Eds.), Sustainable Agriculture Production Using Blockchain Technology. https://doi.org/10.1002/9781394248711.ch11
Karwande, V. S., Pawar, U. B., and Pattnaik, O. (2024). Leveraging Speech-Driven Patterns Multimodal Machine Learning Framework for Accurate Early-Stage Parkinson’s Disease Prediction: A Survey. In Proceedings of the 2nd International Conference on Advanced Computing and Communication Technologies (ICACCTech 2024) ( 525–532). https://doi.org/10.1109/ICACCTech65084.2024.00091
Patil, R. V., et al. (2025). Decentralized Autonomous Organizations as Emerging Economic Entities in Accounting and Governance Frameworks. International Journal of Accounting and Economics Studies, 12(4), 166–177. https://doi.org/10.14419/1sy2j677
Phurisikul, P., and Phuakkong, C. (2025). The Elderly Welfare Management of Local Administrative Government Organizations in Thailand. International Journal of Research and Development in Management Research, 14(1), 21–28. https://doi.org/10.65521/ijrdmr.v14i1.102
Rawandale, U. S., and Kolte, M. T. (2021). Design, Development and Analysis of Variable Bandwidth Filter Bank for Enhancing the Performance of Hearing Aid System. International Journal of Intelligent Engineering Systems, 14(6), 391–401. https://doi.org/10.22266/ijies2021.1231.35
Suri, S., Lakshman, K., Goyal, E., Goyal, G., Sood, G., Mirajkar, G. S., and Anerao, P. (2025). Emotion Modeling in Sculpture Design Using Neural Networks. ShodhKosh: Journal of Visual and Performing Arts, 3(3s), 31–40. https://doi.org/10.29121/shodhkosh.v6.i3s.2025.6756
Vasanthan, R., Balaji, P., and Dzuvichu, K. (2019). A Study on Kinaesthetic Style Preferences and Second Language Learning Requirements at the Primary Level. Journal of Advanced Research in Dynamical and Control Systems, 11(11 Special Issue), 861–869. https://doi.org/10.5373/JARDCS/V11SP11/20193108
|
|
This work is licensed under a: Creative Commons Attribution 4.0 International License
© ShodhKosh 2026. All Rights Reserved.