@inbook{6425a41f3bf041af8a4e631be42e3b0c,
title = "Insurance Risk Prediction Using Machine Learning",
abstract = "Underwriting decisions by insurance companies make a significant contribution to their profitability. Machine Learning (ML) techniques in underwriting decision making have saved time and improved operational efficiencies. A user-friendly cause-and-effect explanation of model{\textquoteright}s predictions is useful to stakeholders, financial institutions and regulators. This research performed comparative analysis between tree-based classifiers such as Decision Tree, Random Forest and XGBoost. The study focused on enhancing risk assessment capabilities for life insurance companies using predictive analytics by classifying the insurance risk based on the historical data and propose the appropriate model to assess risk. Its purpose also included incorporating mechanisms that can aid in user friendly interpretation of ML models. Of all the models created as part of this research, the XGBoost classifier performed the best when compared to other classifiers, with an AUC value of 0.86 and F1-score above 0.56 on the validation set. The Random Forest classifier got AUC value of.84 and f1 score of.53 on the validation dataset. The results indicate the importance and advantages of tree -based models. These models i.e., XGBoost, decision tree and random forest are one of the best alternate techniques after the advent and popularity of the new age techniques in the machine learning such as neural networks, deep learning etc. The research also provides an insight on the interpretability of these conventional techniques by way of {\textquoteleft}SHAP{\textquoteright} or shapley values and {\textquoteleft}Feature Importance{\textquoteright} or {\textquoteleft}Variable Importance{\textquoteright}. SHAP was used on complex models such as XGBoost and neural networks whereas Feature Importance is used in supervised learning methods such as Logistic Regression and tree- based models such as Decision Tree and Random Forest. Overall, the study was able to propose XGBoost as the most accurate model for Insurance risk classification and predictions.",
keywords = "Insurance, Machine learning, Neural networks, Random forest, Risk prediction, Support vector machine",
author = "Rahul Sahai and Ali Al-Ataby and Sulaf Assi and Manoj Jayabalan and Panagiotis Liatsis and Loy, {Chong Kim} and Abdullah Al-Hamid and Sahar Al-Sudani and Maitham Alamran and Hoshang Kolivand",
note = "Publisher Copyright: {\textcopyright} 2023, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.",
year = "2023",
doi = "10.1007/978-981-99-0741-0_30",
language = "British English",
series = "Lecture Notes on Data Engineering and Communications Technologies",
publisher = "Springer Science and Business Media Deutschland GmbH",
pages = "419--433",
booktitle = "Lecture Notes on Data Engineering and Communications Technologies",
address = "Germany",
}