This paper presents the development of a technological solution based on image processing for the immersive digitization of the Murocomba Protected Forest, located in the province of Los Ríos, Ecuador. The proposal integrates the use of a 360° camera to capture high-quality spherical images, which were digitally processed to optimize their sharpness, correct imperfections, and facilitate their integration into an interactive virtual environment. As part of the system, an artificial intelligence model with YOLOv8 architecture was implemented, trained to detect plant species present in the forest trails. These detections were incorporated into a web application that allows users to explore the trail remotely, interact with the content, and identify the flora of the site without the need for physical presence. The project seeks to offer an alternative for the conservation of the ecosystem through digital access to its biodiversity, reducing direct human impact and promoting environmental education. The results demonstrate the effectiveness of combining immersive technologies and sensing models to represent natural environments with high ecological value in an accessible and sustainable way.

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
References
Cho, Y., & Park, K. S. (2023). Designing Immersive Virtual Reality Simulation for Environmental Science Education. Electronics 2023, 12(2), 315. https://doi.org/10.3390/ELECTRONICS12020315
Guo, K., Fan, A., Lehto, X., & Day, J. (2021). Immersive Digital Tourism: The Role of Multisensory Cues in Digital Museum Experiences. Journal of Hospitality & Tourism Research, 47(6), 1017–1039. https://doi.org/10.1177/10963480211030319
Schöne, B., Kisker, J., Lange, L., Gruber, T., Sylvester, S., & Osinsky, R. (2023). The reality of virtual reality. Frontiers in Psychology, 14. https://doi.org/10.3389/FPSYG.2023.1093014
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep Residual Learning for Image Recognition (pp. 770–778). http://image-net.org/challenges/LSVRC/2015/
Nazare, A.-K., Moldoveanu, A., & Moldoveanu, F. (2024). Virtual Journeys, Real Engagement: Analyzing User Experience on a Virtual Travel Social Platform. Information, 15(7), 396. https://doi.org/10.3390/info15070396
Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You Only Look Once: Unified, Real-Time Object Detection (pp. 779–788). http://pjreddie.com/yolo/
Rodríguez-Lira, D.-C., Córdova-Esparza, D.-M., Álvarez-Alvarado, J. M., Romero-González, J.-A., Terven, J., & Rodríguez-Reséndiz, J. (2024). Comparative Analysis of YOLO Models for Bean Leaf Disease Detection in Natural Environments. AgriEngineering, 6(4), 4585-4603. https://doi.org/10.3390/agriengineering6040262
Shorten, C., & Khoshgoftaar, T. M. (2019). A survey on Image Data Augmentation for Deep Learning. Journal of Big Data, 6(1), 1–48. https://doi.org/10.1186/S40537-019-0197-0/FIGURES/33
Vongpaxa, V. (2024). Application of WebVR Technology for 360-Panoramic Heritage Exploration: A Case Study of Wat Phou Heritage, Laos. Engineering and Technology Journal, 09(11). https://doi.org/10.47191/ETJ/V9I11.01