Explainable paper classification system using topic modeling and SHAP
- Abstract
- The exponential growth of academic papers necessitates sophisticated classification systems to effectively manage and navigate vast information repositories. Despite the proliferation of such systems, traditional approaches often rely on embeddings that do not allow for easy interpretation of classification decisions, creating a gap in transparency and understanding. To address these challenges, we propose an innovative explainable paper classification system that combines Latent Semantic Analysis (LSA) for topic modeling with explainable artificial intelligence (XAI) techniques. Our objective is to identify which topics significantly influence the classification outcomes, incorporating Shapley additive explanations (SHAP) as a key XAI technique. Our system extracts topic assignments and word assignments from paper abstracts using LSA topic modeling. Topic assignments are then employed as embeddings in a multilayer perceptron (MLP) classification model, with the word assignments further utilized alongside SHAP for interpreting the classification results at the corpus, document, and word levels, enhancing interpretability and providing a clear rationale for each classification decision. We applied our model to a dataset from the Web of Science, specifically focusing on the field of nanomaterials. Our model demons
- Author(s)
- 정호현; 신나경; 김준희; 이율희; 문희성
- Issued Date
- 2025-05-01
- Type
- Article
- Keyword
- 인공지능
- DOI
- 10.3233/IDA-240075
- URI
- http://repository.sungshin.ac.kr/handle/2025.oak/8756
- Publisher
- SAGE PUBLICATIONS INC
- ISSN
- 1088-467X
-
Appears in Collections:
- 수리통계데이터사이언스학부 > 학술논문
- 공개 및 라이선스
-
- 파일 목록
-
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.