Umfassende Service-Einschränkungen im Bereich Ausleihe ab 17. März!

Treffer: Enhancing concept alignment with explanatory interactive disentangled representation learning.

Title:
Enhancing concept alignment with explanatory interactive disentangled representation learning.
Authors:
Meng X; College of Computer Science and Technology, Zhejiang University, Hangzhou, China. Electronic address: mengxiyu@zju.edu.cn., Lin Y; College of Computer Science and Technology, Zhejiang University, Hangzhou, China. Electronic address: linyilong@zju.edu.cn., Wu Y; College of Computer Science and Technology, Zhejiang University, Hangzhou, China. Electronic address: wuyuhan@zju.edu.cn., Ying L; College of Computer Science and Technology, Zhejiang University, Hangzhou, China. Electronic address: yingluu@zju.edu.cn.
Source:
Neural networks : the official journal of the International Neural Network Society [Neural Netw] 2026 Apr; Vol. 196, pp. 108346. Date of Electronic Publication: 2025 Nov 22.
Publication Type:
Journal Article
Language:
English
Journal Info:
Publisher: Pergamon Press Country of Publication: United States NLM ID: 8805018 Publication Model: Print-Electronic Cited Medium: Internet ISSN: 1879-2782 (Electronic) Linking ISSN: 08936080 NLM ISO Abbreviation: Neural Netw Subsets: MEDLINE
Imprint Name(s):
Original Publication: New York : Pergamon Press, [c1988-
Contributed Indexing:
Keywords: Concept alignment; Contrastive learning; Explainable machine learning; Explanatory interactive learning; Visual analytics
Entry Date(s):
Date Created: 20251127 Date Completed: 20260203 Latest Revision: 20260203
Update Code:
20260203
DOI:
10.1016/j.neunet.2025.108346
PMID:
41308263
Database:
MEDLINE

Weitere Informationen

Deep learning aims to learn "good" representations that are effective for machine learning tasks, but these representations often lack interpretability due to the black-box nature of neural networks. Disentangled representation learning opts to separate the representations with regard to independent human-defined concepts, which paves the path to model explainability. However, traditional supervised learning approaches require extensive manual concept labeling, which is impractical for large-scale datasets. In this paper, we propose a XIDRL framework (eXplanatory Interactive Disentangled Representation Learning) that enables efficient collaboration between our proposed state-of-the-art representation disentangling technique, supervised contrastive learning with invariant risk minimization (SCL+IRM), and human experts. The introduced SCL+IRM algorithm can provide improved alignment capabilities, further enhancing concept alignment. Based on the framework, we design and develop a visual analytics system to assist machine learning experts in exploring concept alignments, comprehending model behaviors, and refining concepts. Additionally, we incorporate the w-BiLRP algorithm to enhance model interpretability. The insights derived from these endeavors are utilized to update the model and align the data with the human concepts. Besides, we present two case studies that demonstrate how our prototype system facilitates the creation of interpretable and human-controllable disentangled representations. Code, data, and model checkpoints will be released after the review period.
(Copyright © 2025. Published by Elsevier Ltd.)

Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.