DEEP LEARNING FOR PHOTOREALISTIC RENDERING IN ART EDUCATION

Authors

  • Priyadarshani Singh Associate, Professor, School of Business Management, Noida international University 203201
  • Abhishek Singla Chitkara Centre for Research and Development, Chitkara University, Himachal Pradesh, Solan, 174103, India
  • Pooja Sharma Centre of Research Impact and Outcome, Chitkara University, Rajpura- 140417, Punjab, India
  • Trilochan Tarai Assistant Professor, Department of Computer Science and Engineering, Institute of Technical Education and Research, Siksha 'O' Anusandhan (Deemed to be University) Bhubaneswar, Odisha, India
  • Samrat Bandyopadhyay Assistant Professor, Department of Computer Science & IT, ARKA JAIN University Jamshedpur, Jharkhand, India
  • Dr. Pravin .A Professor, Department of Computer Science and Engineering, Sathyabama Institute of Science and Technology, Chennai, Tamil Nadu, India

DOI:

https://doi.org/10.29121/shodhkosh.v6.i2s.2025.6733

Keywords:

Deep Learning, Photorealistic Rendering, Art Education, Generative Adversarial Networks (GANs), Convolutional Neural Networks (CNNs), Neural Style Transfer

Abstract [English]

It has altered the environment of art education, as deep learning techniques provide a three-dimensional visual experience and creative opportunities of a scale never seen before. It is a study that dwells on the application of convolutional neural networks (CNNs) and generative adversarial networks (GANs) and diffusion models in generating artistic scenes, high fidelity and photorealistic images and simulations to teach. The system that employs neural style transfer, the perceptual losses and the Volumetric rendering equations are used to reproduce the dynamism in the light transportation and texture with precision. The rendering process is mathematically formulated as a perceptual quality maximization by adversarial loss to make the results realistic by progressive refinement of the discriminator networks. It is the framework of art education that enable students imagine the conceptual compositions in different situations with different lighting conditions, material characteristics, which lead to the development of the awareness of spatial aesthetic, composition and the realism. Deep learning models make it more democratic for less expensive access to professional quality rendering tools without having to rely on more expensive ray tracing techniques. The introduction of these technologies into the teaching of art facilitates the learning by experience and creative experimentation but it is in line with sustainable and open digital art practices. Also, neural rendering allows for real-time feedback, which makes it possible to set up adaptive learning environments, where pupils can improve their work using AI-provided feedback. This paper focuses on the pedagogical, technologies, and aesthetic consequences of deep learning-based photorealism in art education and outlines a system by which intelligent rendering systems can be developed in the future that combine artistic creativity and computational accuracy.

References

Bennett, C., and Robinson, C. (2023). Non-Photorealistic Rendering for Interactive Graphics Using Deep Neural Networks. Computer Graphics Forum, 42(6), 1610–1622.

Boubekeur, T., and Kato, H. (2023). Artistic Style Transfer with High-Level Feature Loss for Non-Photorealistic Rendering. Computer Graphics Forum, 42(1), 59–72.

Chen, W., Liu, Y., and Zhang, X. (2022). Dual-Cycle Consistency for Detail-Preserving Style Transfer in CycleGAN. IEEE Transactions on Neural Networks and Learning Systems, 33(9), 4236–4249.

González, A., and Pérez, P. (2023). Adversarial Training for Artistic Image Synthesis: A GAN Approach. Journal of Visual Communication and Image Representation, 88, 103365.

Huang, J., Li, L., and He, J. (2021). Improved CNN-Based Style Transfer Using Perceptual Similarity Metrics. Pattern Recognition, 108, 107539.

Kim, H., Choi, K., and Lee, J. (2023). Hybrid GAN Models with Attention Mechanisms for Style Transfer and Detail Preservation. IEEE Transactions on Image Processing, 32, 184–195.

Schulz, M., and Kähler, C. (2023). Style-Preserving GANs for Artistic Rendering in 3D Modeling. Computer Graphics Forum, 42(2), 75–88.

Zhang, H., and Liu, Y. (2022). Fine-Tuning CycleGAN for Style Consistency and Texture Preservation. Computer Vision and Image Understanding, 214, 103347.

Zhang, L., Zhang, Y., and Zhang, D. (2020). Adaptive Neural Texture Synthesis for Style Transfer. IEEE Transactions on Image Processing, 29, 4512–4524. https://doi.org/10.1109/TIP.2020.2969081 DOI: https://doi.org/10.1109/TIP.2020.2969081

Zhao, J., Liu, X., and Huang, T. (2021). Improved Detail Preservation in GAN-Based Style Transfer. Journal of Visual Communication and Image Representation, 75, 103070.

Zhou, W., and Cang, M. (2024). A Deep Learning-Based Non-Photorealistic Rendering (NPR) Generation Method. In Proceedings of the 4th International Symposium on Artificial Intelligence and Intelligent Manufacturing (AIIM 2024) (981–984). IEEE. https://doi.org/10.1109/AIIM64537.2024.10934487 DOI: https://doi.org/10.1109/AIIM64537.2024.10934487

Zhu, J., Xu, H., and Wang, Y. (2020). Self-Attention-Based Methods for Style Transfer. International Journal Of Computer Vision, 128(5), 1232–1243.

Downloads

Published

2025-12-16

How to Cite

Singh, P., Singla, A., Sharma, P., Tarai, T., Bandyopadhyay, S., & Pravin .A. (2025). DEEP LEARNING FOR PHOTOREALISTIC RENDERING IN ART EDUCATION. ShodhKosh: Journal of Visual and Performing Arts, 6(2s), 408–416. https://doi.org/10.29121/shodhkosh.v6.i2s.2025.6733