REINFORCEMENT LEARNING IN CREATIVE SKILL DEVELOPMENT

Authors

  • Madhur Taneja Centre of Research Impact and Outcome, Chitkara University, Rajpura- 140417, Punjab, India
  • Harsh Tomer Assistant Professor, Department of Journalism & Mass Communication, Vivekananda Global University, Jaipur, India
  • Akhilesh Kumar Khan Greater Noida, Uttar Pradesh 201306, India
  • Sunil Thakur Professor,School,of,Engineering,&,Technology,,Noida,international,University,203201
  • Manish Nagpal Chitkara Centre for Research and Development, Chitkara University, Himachal Pradesh, Solan, 174103, India
  • Dr. Anita Walia Associate Professor, Department of Management Studies, JAIN (Deemed-to-be University), Bengaluru, Karnataka, India
  • Shailesh Kulkarni Department of E&TC Engineering Vishwakarma Institute of Technology, Pune, Maharashtra, 411037 India

DOI:

https://doi.org/10.29121/shodhkosh.v6.i3s.2025.6791

Keywords:

Reinforcement Learning, Computational Creativity, Creative Skill Development, Reward Modeling, Generative AI, Adaptive Clothing

Abstract [English]

Reinforcement Learning (RL) is a very effective computational model that has been used to model adaptive decision-making but not exploited yet in advancing development in creative skills. This paper explores the application of RL in developing creativity in areas of visual art, music composition, design, and writing. We can place RL as a natural process of directing agents to new and valuable outcomes by theorizing creativity as a process that can be learned, and improved through exploration, evaluation and refinement. The study incorporates the knowledge in the areas of cognitive science, computational creativity, and available applications of RL to develop a methodology (environment simulation, creative dataset, and reward-based learning algorithm) to achieve environment simulation. We have an RL-based creative agent, which can interact with environment-specific domains via the feedback loop and dynamically determined reward functions that encourage originality, coherence and aesthetic or usability. The model structure focuses on multi-modal input representation, hierarchical acquisition of policy and adapting the reward modulation in order to stimulate cognitive diversity and intentional exploration. Testing is based on quantitative measures of creativity (novelty support, distributional distance, and statistical surprise) and qualitative measures of creativity, as determined by human subjects. Findings show that creative heuristics can be gradually learned by the agents of RL and that more and more original artifacts can be produced by the agents and that the agents update their strategies based on the evaluative feedback.

References

Ashton, M. C., Lee, K., and De Vries, R. E. (2020). The HEXACO Model of Personality Structure and the Importance of Agreeableness. European Journal of Personality, 34(1), 3–19. https://doi.org/10.1002/per.2242 DOI: https://doi.org/10.1002/per.2242

DeYoung, C. G., and Krueger, R. F. (2021). Understanding Personality Through Biological and Genetic Bases. Annual Review of Psychology, 72, 555–580.

Dong, S., Wang, P., and Abbas, K. (2021). A Survey on Deep Learning and its Applications. Computer Science Review, 40, Article 100379. https://doi.org/10.1016/j.cosrev.2021.100379 DOI: https://doi.org/10.1016/j.cosrev.2021.100379

Janiesch, C., Zschech, P., and Heinrich, K. (2021). Machine Learning and Deep Learning. Electronic Markets, 31, 685–695. https://doi.org/10.1007/s12525-021-00475-2 DOI: https://doi.org/10.1007/s12525-021-00475-2

Khalid, U., Naeem, M., Stasolla, F., Syed, M., Abbas, M., and Coronato, A. (2024). Impact of Ai-Powered Solutions in Rehabilitation Process: Recent Improvements and Future Trends. International Journal of General Medicine, 17, 943–969. https://doi.org/10.2147/IJGM.S453903 DOI: https://doi.org/10.2147/IJGM.S453903

Li, Y., Xiong, H., Kong, L., Zhang, R., Xu, F., Chen, G., and Li, M. (2023). MHRR: MOOCs Recommender Service with Meta Hierarchical Reinforced Ranking. IEEE Transactions on Services Computing, 16, 4467–4480. https://doi.org/10.1109/TSC.2023.3325302 DOI: https://doi.org/10.1109/TSC.2023.3325302

Liu, S., and Rizzo, P. (2021). Personality-Aware Virtual Agents: Advances and Challenges. IEEE Transactions on Affective Computing, 12, 1012–1027.

Mageira, K., Pittou, D., Papasalouros, A., Kotis, K., Zangogianni, P., and Daradoumis, A. (2022). Educational AI Chatbots for Content and Language Integrated Learning. Applied Sciences, 12(7), Article 3239. https://doi.org/10.3390/app12073239 DOI: https://doi.org/10.3390/app12073239

Shakya, A. K., Pillai, G., and Chakrabarty, S. (2023). Reinforcement Learning Algorithms: A Brief Survey. Expert Systems with Applications, 231, Article 120495. https://doi.org/10.1016/j.eswa.2023.120495 DOI: https://doi.org/10.1016/j.eswa.2023.120495

Song, Y., Suganthan, P. N., Pedrycz, W., Ou, J., He, Y., Chen, Y., and Wu, Y. (2023). Ensemble Reinforcement Learning: A Survey. Applied Soft Computing, 149, Article 110975. https://doi.org/10.1016/j.asoc.2023.110975 DOI: https://doi.org/10.1016/j.asoc.2023.110975

Tran, K. A., Kondrashova, O., Bradley, A., Williams, E. D., Pearson, J. V., and Waddell, N. (2021). Deep Learning in Cancer Diagnosis, Prognosis and Treatment Selection. Genome Medicine, 13, Article 152. https://doi.org/10.1186/s13073-021-00968-x DOI: https://doi.org/10.1186/s13073-021-00968-x

Wells, L., and Bednarz, T. (2021). Explainable AI and Reinforcement Learning: A Systematic Review of Current Approaches and Trends. Frontiers in Artificial Intelligence, 4, Article 550030. https://doi.org/10.3389/frai.2021.550030 DOI: https://doi.org/10.3389/frai.2021.550030

Zhang, Y., Liu, Y., Kang, W., and Tao, R. (2024). VSS-Net: Visual Semantic Self-Mining Network for Video Summarization. IEEE Transactions on Circuits and Systems for Video Technology, 34, 2775–2788. https://doi.org/10.1109/TCSVT.2023.3312325 DOI: https://doi.org/10.1109/TCSVT.2023.3312325

Zhang, Y., Wang, S., Zhang, Y., and Yu, P. (2025). Asymmetric Light-Aware Progressive Decoding Network for Rgb-Thermal Salient Object Detection. Journal of Electronic Imaging, 34(1), Article 013005. https://doi.org/10.1117/1.JEI.34.1.013005 DOI: https://doi.org/10.1117/1.JEI.34.1.013005

Zhang, Y., Wu, C., Guo, W., Zhang, T., and Li, W. (2023). CFANet: Efficient Detection of UAV Image Based on Cross-Layer Feature Aggregation. IEEE Transactions on Geoscience and Remote Sensing, 61, 1–11. https://doi.org/10.1109/TGRS.2023.3273314 DOI: https://doi.org/10.1109/TGRS.2023.3273314

Zhang, Y., Wu, C., Zhang, T., and Zheng, Y. (2024). Full-Scale Feature Aggregation and Grouping Feature Reconstruction-Based UAV Image Target Detection. IEEE Transactions on Geoscience and Remote Sensing, 62, 1–11. https://doi.org/10.1109/TGRS.2024.3392794 DOI: https://doi.org/10.1109/TGRS.2024.3392794

Zhang, Y., Zhang, T., Wang, S., and Yu, P. (2025). An Efficient Perceptual Video Compression Scheme Based on Deep Learning-Assisted Video Saliency and Just Noticeable Distortion. Engineering Applications of Artificial Intelligence, 141, Article 109806. https://doi.org/10.1016/j.engappai.2024.109806 DOI: https://doi.org/10.1016/j.engappai.2024.109806

Zhang, Y., Zhen, J., Liu, T., Yang, Y., and Cheng, Y. (2025). Adaptive Differentiation Siamese Fusion Network for Remote Sensing Change Detection. IEEE Geoscience and Remote Sensing Letters, 22, 1–5. https://doi.org/10.1109/LGRS.2024.3516775 DOI: https://doi.org/10.1109/LGRS.2024.3516775

Downloads

Published

2025-12-20

How to Cite

Taneja, M., Tomer, H., Khan, A. K., Thakur, S., Nagpal, M., Walia, A., & Kulkarni, S. (2025). REINFORCEMENT LEARNING IN CREATIVE SKILL DEVELOPMENT. ShodhKosh: Journal of Visual and Performing Arts, 6(3s), 282–292. https://doi.org/10.29121/shodhkosh.v6.i3s.2025.6791