PERSONALIZED INTERIOR DESIGN ASSISTANTS: VOICE-BASED AI AGENTS WITH VISUAL REASONING CAPABILITIES
DOI:
https://doi.org/10.29121/shodhkosh.v6.i4s.2025.6945Keywords:
Personalized Interior Design, Voice-Based AI, Visual Reasoning, Multimodal Deep Learning, Human-AI Interaction, Generative DesignAbstract [English]
Including voice-based totally AI bots with superior visual notion capabilities has changed personalized interior layout via making it less difficult for customers to make choices about how matters should look and making the ones choices greater enticing. mainly for people who are not acquainted with traditional design software program, present indoors design equipment are harder to use, less at ease, and less tailor-made as they in large part rely on visual enter. These studies will speak approximately a brand new form of AI-powered indoors layout assistance which could realise voice requests, execute difficult visible notion duties, and give you customised diagram thoughts. The proposed method transforms stated wants into smooth visual design outputs using a multimodal deep learning architecture including natural language processing (NLP) techniques, vision-language transformers (VLTs), and generative adversarial networks (GANs). The assistant may create models that make sense and represent each person's preferences by looking at images of rooms, grasping stated demands like modifying the style, colour scheme, or spatial rearrangements, and then... Research indicates that this approach outperforms conventional text- or image-only systems in terms of accurate recommendations (93.4%), user satisfaction (92.6%), and fast response (2.4 seconds per query). Especially for those who are blind or don't know much about technology, a user research with 150 participants reveals that voice-based communication greatly simplifies usage and more accessible. This study adds to the body of research on how people and computers interact by showing how multimodal AI can make interior design experiences that are welcoming, easy to use, and much personalised.
References
Almusaed, A., Yitmen, I., and Almssad, A. (2023). Enhancing Smart Home Design With Ai Models: A Case Study of Living Spaces Implementation Review. Energies, 16(6), 2636. https://doi.org/10.3390/en16062636 DOI: https://doi.org/10.3390/en16062636
Chamishka, S., Madhavi, I., Nawaratne, R., Alahakoon, D., De Silva, D., Chilamkurti, N., and Nanayakkara, V. (2022). A Voice-Based Real-Time Emotion Detection Technique Using Recurrent Neural Network Empowered Feature Modelling. Multimedia Tools and Applications, 81(25), 35173–35194. https://doi.org/10.1007/s11042-022-13363-4 DOI: https://doi.org/10.1007/s11042-022-13363-4
Edén, A. S., Sandlund, P., Faraon, M., and Rönkkö, K. (2024). VoiceBack: Design of Artificial Intelligence-Driven Voice-Based Feedback System for Customer-Agency Communication in Online Travel Services. Information, 15(8), 468. https://doi.org/10.3390/info15080468 DOI: https://doi.org/10.3390/info15080468
Elghaish, F., Chauhan, J. K., Matarneh, S., Rahimian, F. P., and Hosseini, M. R. (2022). Artificial Intelligence-Based Voice Assistant for BIM Data Management. Automation in Construction, 140, 104320. https://doi.org/10.1016/j.autcon.2022.104320 DOI: https://doi.org/10.1016/j.autcon.2022.104320
Fernandes, D., Garg, S., Nikkel, M., and Guven, G. (2024). A GPT-Powered Assistant for Real-Time Interaction with Building Information Models. Buildings, 14(8), 2499. https://doi.org/10.3390/buildings14082499 DOI: https://doi.org/10.3390/buildings14082499
Kim, J., Kim, W., Nam, J., and Song, H. (2020). “I Can Feel Your Empathic Voice”: Effects of Nonverbal Vocal Cues in Voice User Interface. In R. Bernhaupt, F. Mueller, D. Verweij, and J. Andres (Eds.), Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (pp. 1–12). Association for Computing Machinery. https://doi.org/10.1145/3334480.3383075 DOI: https://doi.org/10.1145/3334480.3383075
Madan, B. S., Zade, N. J., Lanke, N. P., Pathan, S. S., Ajani, S. N., and Khobragade, P. (2024). Self-Supervised Transformer Networks: Unlocking New Possibilities for Label-Free Data. Panamerican Mathematical Journal, 34(4), 194–210. https://doi.org/10.52783/pmj.v34.i4.1878 DOI: https://doi.org/10.52783/pmj.v34.i4.1878
Rzepka, C. (2019). Examining the use of voice assistants: A value-focused thinking approach. In G. Rodriguez-Abitia and C. Ferran (Eds.), Proceedings of the 25th Americas Conference on Information Systems (AMCIS 2019) (p. 20). Association for Information Systems.
Saka, A., Taiwo, R., Saka, N., Salami, B. A., Ajayi, S., Akande, K., and Kazemi, H. (2024). GPT Models in Construction Industry: Opportunities, Limitations, And a use Case Validation. Developments in the Built Environment, 17, 100300. https://doi.org/10.1016/j.dibe.2023.100300 DOI: https://doi.org/10.1016/j.dibe.2023.100300
Seaborn, K., Miyake, N. P., Pennefather, P., and Otake-Matsuura, M. (2021). Voice in Human-Agent Interaction: A Survey. ACM Computing Surveys, 54(4), 1–43. https://doi.org/10.1145/3386867 DOI: https://doi.org/10.1145/3386867
Thirumuruganathan, S., Kunjir, M., Ouzzani, M., and Chawla, S. (2021). Automated Annotations for AI Data and Model Transparency. Journal of Data and Information Quality, 14(1), 1–9. https://doi.org/10.1145/3460000 DOI: https://doi.org/10.1145/3460000
Wei, J., Tag, B., Trippas, J. R., Dingler, T., and Kostakos, V. (2022). What Could Possibly Go Wrong When Interacting with Proactive Smart Speakers? A Case Study Using an ESM Application. In S. Barbosa, C. Lampe, C. Appert, D. A. Shamma, S. Drucker, J. Williamson, and K. Yatani (Eds.), Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (pp. 1–15). Association for Computing Machinery. https://doi.org/10.1145/3491102.3517432 DOI: https://doi.org/10.1145/3491102.3517432
Zheng, J., and Fischer, M. (2023). Dynamic Prompt-Based Virtual Assistant Framework for BIM Information Search. Automation in Construction, 155, 105067. https://doi.org/10.1016/j.autcon.2023.105067 DOI: https://doi.org/10.1016/j.autcon.2023.105067
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Satyam Vishwakarma, Dr. Sagar Vasantrao Joshi, Gouri Moharana, Dr. Smita N. Gambhire, Avinash Somatkar, Harinder Pal Singh

This work is licensed under a Creative Commons Attribution 4.0 International License.
With the licence CC-BY, authors retain the copyright, allowing anyone to download, reuse, re-print, modify, distribute, and/or copy their contribution. The work must be properly attributed to its author.
It is not necessary to ask for further permission from the author or journal board.
This journal provides immediate open access to its content on the principle that making research freely available to the public supports a greater global exchange of knowledge.























