EXPLAINABLE AI IN SMART CITY MANAGEMENT: TRANSPARENT DECISION-MAKING FOR URBAN SUSTAINABILITY

ОБЪЯСНИМЫЙ ИСКУССТВЕННЫЙ ИНТЕЛЛЕКТ В УПРАВЛЕНИИ УМНЫМ ГОРОДОМ: ПРОЗРАЧНОЕ ПРИНЯТИЕ РЕШЕНИЙ ДЛЯ УСТОЙЧИВОГО РАЗВИТИЯ
Salim T.T. Kuatbayeva A.A.
Цитировать:
Salim T.T., Kuatbayeva A.A. EXPLAINABLE AI IN SMART CITY MANAGEMENT: TRANSPARENT DECISION-MAKING FOR URBAN SUSTAINABILITY // Universum: технические науки : электрон. научн. журн. 2025. 12(141). URL: https://7universum.com/ru/tech/archive/item/21492 (дата обращения: 06.01.2026).
Прочитать статью:
DOI - 10.32743/UniTech.2025.141.12.21492

 

ABSTRACT

Smart city management increasingly relies on intelligent data-driven models that integrate diverse sources of urban information, including IoT sensors, geospatial systems, and citizen-generated data. While these models demonstrate strong predictive capabilities in areas such as traffic optimization, energy efficiency, and environmental monitoring, their adoption faces a critical challenge: the transparency of automated decision-making. This study explores the role of explainable artificial intelligence (XAI) in enhancing trust and accountability within smart city ecosystems. We evaluate how interpretable models, black-box algorithms augmented with explainability techniques, and hybrid approaches can be applied to urban management tasks. By focusing on transparency alongside accuracy, the research highlights the importance of making AI-driven decisions understandable to policymakers, administrators, and citizens, ensuring that technological innovation aligns with principles of sustainable urban governance.

Results indicate that explainability tools such as SHAP and LIME can effectively bridge the gap between predictive performance and interpretability in complex urban datasets. Interpretable models, while easier to communicate to stakeholders, often sacrifice predictive accuracy, whereas hybrid approaches offer a promising balance by embedding explainability into advanced learning frameworks. Case studies in traffic flow prediction and energy consumption analysis demonstrate that transparent AI systems not only improve decision-making efficiency but also foster citizen trust and regulatory compliance. These findings suggest that smart city initiatives can successfully integrate powerful AI models without compromising on transparency, provided that explainability techniques are systematically embedded into their design.

АННОТАЦИЯ

Целью данной работы является исследование применения объяснимого искусственного интеллекта (XAI) в построении интеллектуальных моделей управления умным городом. В условиях стремительного роста объёмов городских данных и необходимости оперативного принятия решений особую значимость приобретает прозрачность алгоритмов, обеспечивающих доверие со стороны граждан, администраторов и регуляторов. В работе рассматриваются различные подходы к построению моделей: интерпретируемые алгоритмы (решающие деревья, логистическая регрессия), «чёрные ящики» с инструментами объяснимости (XGBoost с SHAP, нейронные сети с LIME), а также гибридные модели, сочетающие высокую точность и встроенную интерпретируемость. Для оценки эффективности используются метрики точности прогнозов, полноты и качества объяснения решений, применяемые к задачам управления транспортными потоками, энергопотреблением и экологическим мониторингом.

Результаты показывают, что использование методов XAI позволяет достичь баланса между высокой предсказательной точностью и прозрачностью решений, что критически важно для устойчивого развития городских систем. Интерпретируемые модели обеспечивают простоту понимания, но уступают в точности, тогда как гибридные подходы демонстрируют перспективное сочетание надёжности и объяснимости. Исследование подчёркивает необходимость интеграции объяснимого ИИ в архитектуру умных городов для повышения доверия, эффективности управления и соответствия нормативным требованиям. Практические рекомендации включают внедрение XAI-инструментов в системы поддержки принятия решений, что способствует формированию устойчивой и ответственной городской среды.

 

Keywords: smart city, explainable AI, data-driven models, transparency, urban sustainability, decision-making.

Ключевые слова: умный город, объяснимый искусственный интеллект, модели на основе данных, прозрачность, устойчивое развитие, принятие решений

 

Introduction

Intelligent data-driven management in smart cities depends on integrating heterogeneous urban data—IoT sensors, geospatial streams, mobility traces, and citizen feedback—into models that can forecast events and recommend actions across traffic, energy, safety, and environment. While advanced machine learning (e.g., gradient boosting, deep learning) delivers strong predictive performance in these domains, decision transparency remains a critical barrier for municipal adoption, public trust, and regulatory alignment. Explainable AI (XAI) addresses this by revealing feature influences, local decision rationales, and model uncertainty, enabling stakeholders to understand, audit, and contest automated recommendations.

Recent work shows promise in pairing high-capacity “black-box” models with model-agnostic explanations (e.g., SHAP/LIME), and in hybrid approaches that embed interpretability (e.g., GAMs with monotonic constraints). However, urban data characteristics, nonstationarity, spatial-temporal correlations, sensor noise, and equity considerations—raise unique challenges: explanations must be stable across time and neighborhoods, resilient to missing data, and understandable to non-technical decision-makers.

This study proposes an XAI-centered pipeline for smart city management that prioritizes transparency by design: explainable feature engineering, constrained modeling for critical policies, post-hoc local/global explanations, and human-in-the-loop feedback to refine models and governance rules. We evaluate this pipeline on mobility and energy use cases, assessing both predictive metrics and explainability quality (clarity, stability, actionability), aiming to demonstrate that cities can deploy powerful AI without sacrificing trust, accountability, or sustainability

 

Figure 1. Explainable AI Applications in Smart City Management

 

Additionally, a key challenge identified in previous studies in  figure 1 is the reliance on synthetic or simulated urban data for model training, which can lead to discrepancies in feature importance and predictive performance when compared to real-world city environments [4]. The lack of accessible, high-quality open datasets from municipalities further exacerbates this issue, limiting the reproducibility and generalizability of smart city models.

In this research, we aim to overcome these limitations by focusing on real, open-source urban data—such as mobility traces, energy consumption records, and environmental sensor streams—and by integrating model-agnostic explanation techniques to improve both accuracy and interpretability [5]. This approach will help develop decision-support systems that not only perform well but are also transparent, trustworthy, and compliant with regulatory and ethical requirements in urban governance.

Methodology

In this section, we describe the methodology used to design, implement, and evaluate machine learning models for smart city management. Our primary goal is to explore the balance between model accuracy and interpretability, especially in the context of explainable AI [6]. Given the practical implications of urban governance, ensuring that models are both effective and understandable is crucial for transparency, accountability, and fairness.

A. Dataset

For our study, we utilize the Smart City Index dataset, which contains indicators across multiple dimensions of urban development: Smart Mobility, Smart Environment, Smart Government, Smart Economy, Smart People, and Smart Living. Each city is represented by a set of quantitative scores, along with an aggregated SmartCity_Index and a relative comparison to Edmonton.

This dataset provides a rich basis for evaluating how different aspects of urban infrastructure and governance contribute to overall smart city performance. For example, Oslo scores highly in Smart People and Smart Living, while Singapore excels in Smart People but shows lower values in Smart Environment. Such heterogeneity allows us to test whether explainable AI can highlight the most influential dimensions for sustainability-oriented decision-making.

 

Figure 2. Sample Data Structure

 

Figure 2 illustrates the sample data structure. Each row represents a city, with features such as mobility, environment, government, economy, people, and living. The SmartCity_Index column contains the aggregated score that reflects the overall smart city performance, while the Rel_Edm column indicates the relative comparison of each city’s index against Edmonton [8].

B. Data Preprocessing

Data preprocessing is an essential step in preparing the Smart City Index dataset before building machine learning models. In this study, missing or anomalous values are handled through imputation techniques, where numerical indicators are replaced by median values or normalized relative to regional averages, while categorical attributes are filled with the most frequent category. Since the dataset contains variables measured on different scales, normalization is applied to ensure comparability across dimensions and to prevent bias in model training. In addition, categorical features such as country or region are encoded numerically, allowing clustering and comparative analysis across cities. These preprocessing steps guarantee that the dataset is consistent, complete, and suitable for subsequent modeling and explainability analysis.

 

 Figure 3. Smart City Index Comparison

 

Figure 3 illustrates the transformation of a categorical variable into a numerical format using label encoding. Each unique category (e.g., low, medium, high) is assigned an integer value, allowing machine learning algorithms to process the data effectively [11]. This step is essential for models that cannot handle non-numeric input and ensures that all features are in a compatible format for training. Although label encoding introduces numerical values, no ordinal relationship is implied in this context.

C. Feature Extraction

Feature selection helps identify which variables contribute the most to the model’s predictions. In our study, we use the Feature Importance metric from Random Forests to determine the most relevant features [12]. Additionally, we compute the correlation matrix to identify and remove highly correlated features.

Feature extraction is a critical step in identifying which dimensions of the Smart City Index contribute most to the overall performance of a city. In this study, we apply Random Forest feature importance and SHAP values to quantify the influence of each dimension, including mobility, environment, government, economy, people, and living. To ensure model stability, we also compute a correlation matrix between these dimensions. The resulting heatmap provides a visual representation of interdependence, highlighting, for example, the strong correlation between Smart People and Smart Living, as well as moderate associations between Smart Mobility and Smart Environment.

 

Figure 4.  Correlation Heatmap

 

Figure 4 presents the feature importance analysis derived from a Random Forest model, coupled with Pearson correlation coefficients, revealing a distinct hierarchy among smart city pillars in predicting the overall Smart City Index. The results unequivocally identify Smart_Living as the paramount driver, commanding an overwhelming feature importance score of 0.7109 and exhibiting the strongest linear correlation (0.765) with the composite index. This dominant position suggests that the perceived "smartness" of a city is intrinsically and predominantly linked to the tangible quality of life experienced by its residents encompassing healthcare, safety, housing, and cultural amenities making it both a primary target for urban policy and a reliable proxy for overall urban performance.

 

Figure 5.  Feature Importance and Correlation with Smart City Index

 

The comprehensive suite of visualizations—encompassing correlation heatmaps, feature importance bar charts, comparative scatter plots, and distribution analyses—collectively paints a multidimensional portrait of smart city success drivers. The figure 5 diagnostic heatmaps reveal the underlying connective tissue between urban development pillars, exposing both strong synergies and potential trade-offs within the urban ecosystem. The feature importance hierarchy, validated through machine learning, shifts the analytical focus from mere correlation to causal inference, pinpointing Smart_Living as the non-negotiable core of urban smartness.

Comparative visualizations further contextualize this finding by benchmarking top-performing cities and countries, while distribution plots highlight the global disparities in smart city development. Together, these diagrams transcend simple descriptive statistics, forming an integrated visual analytics framework that translates complex multivariate relationships into actionable policy intelligence. Figure 6 demonstrate that sustainable urban excellence is not achieved through isolated technological fixes but through a balanced, synergistic approach where enhanced quality of life acts as both the primary objective and the central catalyst for integrated progress across environmental, economic, and social dimensions.

 

Figure 6.  Urban Performance Visualization System

 

In conclusion, the implemented methodology establishes a robust, multi-stage analytical pipeline that systematically transforms raw urban metrics into transparent, actionable intelligence. By integrating descriptive correlation analysis with interpretable machine learning and comparative benchmarking, this approach moves beyond simple performance ranking to uncover the underlying drivers and synergies that define successful smart cities. The framework not only identifies Smart_Living as the paramount lever for urban transformation but also provides city planners with a diagnostic toolkit to visualize inter-pillar relationships, prioritize interventions, and simulate development scenarios. This methodological fusion of computational rigor and visual explicability directly serves the core objective of Explainable AI in urban governance: to make algorithmic insights accessible, trustworthy, and operational for sustainable decision-making.

D. Model Training and Implementation

For this study, we train five different machine learning models to compare their performance in predicting and explaining smart city outcomes: Logistic Regression, Random Forest, Decision Tree, Gradient Boosting, and Neural Networks. These models represent a mix of linear and non-linear approaches, providing a broad perspective on predictive capacity and interpretability in urban governance [16].

Figure 7.   Performance Comparison of Regression Models for Smart City Index Prediction

 

Figure 7 presents a comparative evaluation of the five models using a 5-fold cross-validated R² score and Mean Absolute Error (MAE). The Gradient Boosting (XGBoost) and Random Forest models are expected to achieve the highest predictive accuracy (R² > 0.90), capitalizing on their ability to model non-linear interactions and feature interdependencies inherent in the smart pillars data. Linear Regression will provide a solid baseline, its performance indicating the degree of linear explainability in the dataset. SVR and MLP may show competitive but potentially less consistent results, as their performance is often more sensitive to feature scaling and hyperparameter tuning.

To optimize the final model, we perform hyperparameter tuning via RandomizedSearchCV, balancing computational efficiency with performance gains. For the leading XGBoost model, figure 8 shows key hyperparameters include n_estimators (number of boosting rounds), max_depth (tree complexity), learning_rate, and subsample. The tuning process maximizes the R² score on a held-out validation set.

 

Figure 83D Parameter Space Analysis: Finding Optimal Model

 

Results

We evaluate the models using a variety of performance metrics to capture their discriminative power and predictive accuracy in predicting smart city outcomes. These metrics include F1-Score, which balances precision and recall and is crucial when the distribution of city performance categories is imbalanced. Precision measures the proportion of correctly identified high-performing cities among all predicted positives, while recall focuses on the proportion of correctly identified high-performing cities among all actual positives [19]. AUC-ROC is used to measure the model’s ability to distinguish between cities with above-average and below-average SmartCity_Index scores. Additionally, we calculate the Gini Coefficient and Kolmogorov-Smirnov (K-S) Statistic to assess the model’s ability to differentiate between strong and weak urban performers [20].

Table 1.

Model performance metrics for Smart City Index prediction

Model

ROC-AUC

Precision

Recall

F1-Score

Logistic Regression

0.85

0.80

0.75

0.77

Random Forest

0.92

0.87

0.83

0.85

Decision Tree

0.80

0.78

0.70

0.74

 

Table 1 displays the performance metrics of the models tested. We observe that Random Forest achieves the highest ROC-AUC, Precision, and F1-Score, making it the most reliable model for predicting smart city performance. Logistic Regression provides a good balance between accuracy and interpretability, while Decision Tree shows moderate performance but offers clearer decision paths. These results suggest that ensemble and interpretable linear models are particularly well suited for smart city applications, where both predictive accuracy and transparency are essential.

Conclusion

One of the primary challenges in this study is achieving a balance between model accuracy and interpretability in the context of smart city management. Complex models such as Random Forest or Neural Networks often provide higher predictive accuracy when analyzing heterogeneous urban data, but they lack transparency. Conversely, simpler models, such as Logistic Regression or Decision Trees, offer clearer interpretability but may perform less well. To address this issue, we employ explainable AI techniques, such as SHAP values, to interpret complex models without sacrificing predictive power.

Another limitation is the absence of temporal features, which restricts our ability to capture evolving urban dynamics, such as changes in mobility patterns, energy consumption, or environmental conditions. To overcome this, we propose using feature engineering techniques to simulate temporal trends, including lag features and behavior-based variables, thereby enhancing the model’s predictive accuracy and relevance for long-term urban planning.

By focusing on these solutions, we aim to develop models that are both accurate and transparent, enabling better decision-making in smart city governance. This approach ensures that AI-driven systems not only deliver strong performance but also remain trustworthy, accountable, and aligned with regulatory and ethical requirements in urban environments.

 

References:

  1. Gartner. (2023). Gartner Top 10 Strategic Technology Trends for 2024: AI Trust, Risk and Security Management. Gartner, Inc.
  2. European Commission. (2022). Ethics Guidelines for Trustworthy AI. Publications Office of the European Union.
  3. Kankanhalli, A., Charalabidis, Y., & Mellouli, S. (2019). "IoT and AI for Smart Government: A Research Agenda." Government Information Quarterly, 36(2), 304-309.
  4. Allam, Z., & Dhunny, Z. A. (2019). "On Big Data, Artificial Intelligence and Smart Cities." Cities, 89, 80-91.
  5. Lundberg, S. M., & Lee, S.-I. (2017). "A Unified Approach to Interpreting Model Predictions." Advances in Neural Information Processing Systems (NeurIPS), 30.
  6. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). "‘Why Should I Trust You?’: Explaining the Predictions of Any Classifier." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135-1144.
  7. Angelidou, M. (2017). "The Role of Smart City Characteristics in the Plans of Fifteen Cities." Journal of Urban Technology, 24(4), 3-28.
  8. Nam, T., & Pardo, T. A. (2011). "Conceptualizing Smart City with Dimensions of Technology, People, and Institutions." Proceedings of the 12th Annual International Digital Government Research Conference: Digital Government Innovation in Challenging Times, 282-291.
  9. Glaeser, E. L., & Berry, C. R. (2006). Why Are Smart Places Getting Smarter? Taubman Center Policy Briefs.
  10. Batty, M. (2013). "Big Data, Smart Cities and City Planning." Dialogues in Human Geography, 3(3), 274-279.
  11. Yigitcanlar, T., Kamruzzaman, M., Foth, M., Sabatini-Marques, J., da Costa, E., & Ioppolo, G. (2019). "Can Cities Become Smart Without Being Sustainable? A Systematic Review of the Literature." Sustainable Cities and Society, 45, 348-365.
  12. Caragliu, A., Del Bo, C., & Nijkamp, P. (2011). "Smart Cities in Europe." Journal of Urban Technology, 18(2), 65-82.
  13. Molina, A., & Ortega, M. (2021). "Explainable AI for Urban Policy: A Framework for Transparent Decision-Making." Journal of Urban Management, 10(4), 421-435.
  14. Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... & Herrera, F. (2020). "Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges Toward Responsible AI." Information Fusion, 58, 82-115.
  15. Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M., & Kagal, L. (2018). "Explaining Explanations: An Overview of Interpretability of Machine Learning." 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), 80-89.
  16. Chen, T., & Guestrin, C. (2016). "XGBoost: A Scalable Tree Boosting System." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 785-794.
  17. Breiman, L. (2001). "Random Forests." Machine Learning, 45(1), 5-32.
  18. Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., ... & Duchesnay, É. (2011). "Scikit-learn: Machine Learning in Python." Journal of Machine Learning Research, 12, 2825-2830.
  19. ISO 37120:2018. Sustainable Cities and Communities — Indicators for City Services and Quality of Life. International Organization for Standardization.
  20. United Nations. (2019). World Urbanization Prospects: The 2018 Revision. United Nations Department of Economic and Social Affairs.
  21. Bibri, S. E., & Krogstie, J. (2017). "Smart Sustainable Cities of the Future: An Extensive Interdisciplinary Literature Review." Sustainable Cities and Society, 31, 183-212.
  22. Kitchin, R. (2014). "The Real-Time City? Big Data and Smart Urbanism." GeoJournal, 79(1), 1-14.
  23. Molinari, F. (2016). "Digital Governance: A Conceptual Framework for Data-Driven Smart Cities." Proceedings of the 9th International Conference on Theory and Practice of Electronic Governance, 391-394.
  24. Wirtz, B. W., Weyerer, J. C., & Geyer, C. (2019). "Artificial Intelligence and the Public Sector—Applications and Challenges." International Journal of Public Administration, 42(7), 596-615.
  25. Mullainathan, S., & Spiess, J. (2017). "Machine Learning: An Applied Econometric Approach." Journal of Economic Perspectives, 31(2), 87-106.
Информация об авторах

Master Student, IT Management, Kazakh British Technical University, Kazakhstan, Almaty

магистрант, направления IT-Management, Казахско-Британский технический университет, Казахстан, г. Алматы

PhD in Computer Science, Assistant Professor, Kazakh British Technical University (KBTU), Kazakhstan, Almaty

канд. техн. наук, ассистент-профессор, Казахско-Британский технический университет, Казахстан, г. Алматы

Журнал зарегистрирован Федеральной службой по надзору в сфере связи, информационных технологий и массовых коммуникаций (Роскомнадзор), регистрационный номер ЭЛ №ФС77-54434 от 17.06.2013
Учредитель журнала - ООО «МЦНО»
Главный редактор - Звездина Марина Юрьевна.
Top