Skip to main content

Research Repository

Advanced Search

Embedding transparency in artificial intelligence machine learning models: managerial implications on predicting and explaining employee turnover

Chowdhury, Soumyadeb; Joel-Edgar, Sian; Dey, Prasanta; Bhattacharya, Sudeshna; Kharlamov, Alexander

Embedding transparency in artificial intelligence machine learning models: managerial implications on predicting and explaining employee turnover Thumbnail


Authors

Soumyadeb Chowdhury

Sian Joel-Edgar

Prasanta Dey

SUDESHNA BHATTACHARYA S.Bhattacharya@nottingham.ac.uk
Assistant Professor in Organisational Behaviour and Human Resource Management

Alexander Kharlamov



Abstract

Employee turnover (ET) is a major issue faced by firms in all business sectors. Artificial intelligence (AI) machine learning (ML) prediction models can help to classify the likelihood of employees voluntarily departing from employment using historical employee datasets. However, output responses generated by these AI-based ML models lack transparency and interpretability, making it difficult for HR managers to understand the rationale behind the AI predictions. If managers do not understand how and why responses are generated by AI models based on the input datasets, it is unlikely to augment data-driven decision-making and bring value to the organisations. The main purpose of this article is to demonstrate the capability of Local Interpretable Model-Agnostic Explanations (LIME) technique to intuitively explain the ET predictions generated by AI-based ML models for a given employee dataset to HR managers. From a theoretical perspective, we contribute to the International Human Resource Management literature by presenting a conceptual review of AI algorithmic transparency and then discussing its significance to sustain competitive advantage by using the principles of resource-based view theory. We also offer a transparent AI implementation framework using LIME which will provide a useful guide for HR managers to increase the explainability of the AI-based ML models, and therefore mitigate trust issues in data-driven decision-making.

Citation

Chowdhury, S., Joel-Edgar, S., Dey, P., Bhattacharya, S., & Kharlamov, A. (2023). Embedding transparency in artificial intelligence machine learning models: managerial implications on predicting and explaining employee turnover. International Journal of Human Resource Management, 34(14), 2732-2764. https://doi.org/10.1080/09585192.2022.2066981

Journal Article Type Article
Acceptance Date Jan 10, 2022
Online Publication Date Apr 27, 2022
Publication Date Aug 6, 2023
Deposit Date Jan 26, 2024
Publicly Available Date Jan 31, 2024
Journal International Journal of Human Resource Management
Print ISSN 0958-5192
Electronic ISSN 1466-4399
Publisher Routledge
Peer Reviewed Peer Reviewed
Volume 34
Issue 14
Pages 2732-2764
DOI https://doi.org/10.1080/09585192.2022.2066981
Keywords Artificial Intelligence; Machine Learning; Employee Turnover; AI Transparency; Local Interpretation; Model Explainability; Human Intelligence
Public URL https://nottingham-repository.worktribe.com/output/30150474
Publisher URL https://www.tandfonline.com/doi/full/10.1080/09585192.2022.2066981

Files




You might also like



Downloadable Citations