Recent advances in computer vision and Convolutional Neural Networks (CNNs) have facilitated the use of street view imagery (SVI) for the automatic assessment of physical and perceived attributes of walkable areas. However, these methods still overlook the broader urban context and fail to capture and communicate to the user the qualitative factors influencing the assessed walkability score. This paper addresses these challenges by leveraging a Multimodal Large Language Model (MLLM) to provide a holistic assessment of walkability, consisting of both quantitative scores and linguistic qualitative insights. This approach offers a more comprehensive understanding of the factors contributing to the walkability score attributed to the image and enhances the interpretability and practical applicability of the assessments for urban planners and policymakers. Preliminary experiments demonstrate that a MLLM-based methodology can effectively capture a diverse range of factors of walkability, suggesting a promising direction for future developments of evaluation tools aimed at supporting the design of pedestrian-friendly environments.

Enhancing Urban Walkability Assessment with Multimodal Large Language Models / Blečić, Ivan; Saiu, Valeria; Trunfio, Giuseppe A.. - 14819 LNCS:(2024), pp. 394-411. ( 24th International Conference on Computational Science and Its Applications, ICCSA 2024 vnm 2024) [10.1007/978-3-031-65282-0_26].

Enhancing Urban Walkability Assessment with Multimodal Large Language Models

Trunfio, Giuseppe A.
2024-01-01

Abstract

Recent advances in computer vision and Convolutional Neural Networks (CNNs) have facilitated the use of street view imagery (SVI) for the automatic assessment of physical and perceived attributes of walkable areas. However, these methods still overlook the broader urban context and fail to capture and communicate to the user the qualitative factors influencing the assessed walkability score. This paper addresses these challenges by leveraging a Multimodal Large Language Model (MLLM) to provide a holistic assessment of walkability, consisting of both quantitative scores and linguistic qualitative insights. This approach offers a more comprehensive understanding of the factors contributing to the walkability score attributed to the image and enhances the interpretability and practical applicability of the assessments for urban planners and policymakers. Preliminary experiments demonstrate that a MLLM-based methodology can effectively capture a diverse range of factors of walkability, suggesting a promising direction for future developments of evaluation tools aimed at supporting the design of pedestrian-friendly environments.
2024
9783031652813
9783031652820
Enhancing Urban Walkability Assessment with Multimodal Large Language Models / Blečić, Ivan; Saiu, Valeria; Trunfio, Giuseppe A.. - 14819 LNCS:(2024), pp. 394-411. ( 24th International Conference on Computational Science and Its Applications, ICCSA 2024 vnm 2024) [10.1007/978-3-031-65282-0_26].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11388/367989
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 7
  • ???jsp.display-item.citation.isi??? 7
social impact