Soumyadeb Chowdhury
Embedding transparency in artificial intelligence machine learning models: managerial implications on predicting and explaining employee turnover
Chowdhury, Soumyadeb; Joel-Edgar, Sian; Dey, Prasanta; Bhattacharya, Sudeshna; Kharlamov, Alexander
Authors
Sian Joel-Edgar
Prasanta Dey
SUDESHNA BHATTACHARYA S.Bhattacharya@nottingham.ac.uk
Assistant Professor in Organisational Behaviour and Human Resource Management
Alexander Kharlamov
Abstract
Employee turnover (ET) is a major issue faced by firms in all business sectors. Artificial intelligence (AI) machine learning (ML) prediction models can help to classify the likelihood of employees voluntarily departing from employment using historical employee datasets. However, output responses generated by these AI-based ML models lack transparency and interpretability, making it difficult for HR managers to understand the rationale behind the AI predictions. If managers do not understand how and why responses are generated by AI models based on the input datasets, it is unlikely to augment data-driven decision-making and bring value to the organisations. The main purpose of this article is to demonstrate the capability of Local Interpretable Model-Agnostic Explanations (LIME) technique to intuitively explain the ET predictions generated by AI-based ML models for a given employee dataset to HR managers. From a theoretical perspective, we contribute to the International Human Resource Management literature by presenting a conceptual review of AI algorithmic transparency and then discussing its significance to sustain competitive advantage by using the principles of resource-based view theory. We also offer a transparent AI implementation framework using LIME which will provide a useful guide for HR managers to increase the explainability of the AI-based ML models, and therefore mitigate trust issues in data-driven decision-making.
Citation
Chowdhury, S., Joel-Edgar, S., Dey, P., Bhattacharya, S., & Kharlamov, A. (2023). Embedding transparency in artificial intelligence machine learning models: managerial implications on predicting and explaining employee turnover. International Journal of Human Resource Management, 34(14), 2732-2764. https://doi.org/10.1080/09585192.2022.2066981
Journal Article Type | Article |
---|---|
Acceptance Date | Jan 10, 2022 |
Online Publication Date | Apr 27, 2022 |
Publication Date | Aug 6, 2023 |
Deposit Date | Jan 26, 2024 |
Publicly Available Date | Jan 31, 2024 |
Journal | International Journal of Human Resource Management |
Print ISSN | 0958-5192 |
Electronic ISSN | 1466-4399 |
Publisher | Routledge |
Peer Reviewed | Peer Reviewed |
Volume | 34 |
Issue | 14 |
Pages | 2732-2764 |
DOI | https://doi.org/10.1080/09585192.2022.2066981 |
Keywords | Artificial Intelligence; Machine Learning; Employee Turnover; AI Transparency; Local Interpretation; Model Explainability; Human Intelligence |
Public URL | https://nottingham-repository.worktribe.com/output/30150474 |
Publisher URL | https://www.tandfonline.com/doi/full/10.1080/09585192.2022.2066981 |
Files
19367
(963 Kb)
PDF
You might also like
Downloadable Citations
About Repository@Nottingham
Administrator e-mail: discovery-access-systems@nottingham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search