This project focuses on predicting crime patterns using machine learning models and improving interpretability using Explainable Artificial Intelligence (XAI) techniques. The aim is not only to achieve accurate predictions but also to understand why those predictions are made.
- Predict crime outcomes using machine learning models
- Compare performance of multiple models
- Apply SHAP and LIME for model explainability
- Perform feature selection using XAI techniques
- Analyze how explainability affects model performance
-
Dataset 1 – Crimes Against Women
- Structured dataset with features like rape, kidnapping, domestic violence, etc.
-
Dataset 2 – Crime Dataset India
- Real-world dataset with case-level details (city, crime type, victim info, etc.)
- K-Nearest Neighbors (KNN)
- Logistic Regression
- Random Forest
- Gaussian Naive Bayes (GNB)
- Multi-Layer Perceptron (MLP)
-
SHAP (SHapley Additive Explanations)
- Provides global feature importance
- Used for feature selection
-
LIME (Local Interpretable Model-Agnostic Explanations)
- Explains individual predictions
- Used for local interpretability
- Data preprocessing and feature engineering
- Train baseline models (KNN, Logistic Regression, Random Forest)
- Train main models (GNB, MLP)
- Evaluate performance (Accuracy, Precision, Recall, F1-score)
- Apply SHAP and LIME
- Select top features
- Retrain models using selected features
- Compare results before and after XAI
- Dataset 1 achieved high accuracy due to structured data
- Dataset 2 showed lower accuracy due to complexity and noise
- MLP consistently performed better than simpler models
- SHAP improved performance, especially for MLP
- LIME provided local explanations but showed less stability
- SHAP summary plots
- Feature importance bar plots
- Model comparison graphs
- SHAP force plots
- Python
- Scikit-learn
- SHAP
- LIME
- Pandas, NumPy, Matplotlib
- Data quality has a major impact on model performance
- Complex models (MLP) benefit more from feature selection
- SHAP is more effective for global interpretability
- LIME is useful for explaining individual predictions
- Use advanced models like XGBoost and deep learning
- Build real-time crime prediction systems
- Apply geospatial analysis for crime hotspot detection
- Combine SHAP and LIME for hybrid explainability
- Anoushka Deb
- Shalini Shree
- Triyanjana Paul
- Srija Adhya
This project was developed as part of an academic research initiative on Explainable AI and crime prediction.