Chrome Extension
WeChat Mini Program
Use on ChatGLM

Conditional Mutual Information Constrained Deep Learning for Classification

IEEE transactions on neural networks and learning systems(2025)

Cited 0|Views17
Abstract
The concepts of conditional mutual information (CMI) and normalized conditional mutual information (NCMI) are introduced to measure the concentration and separation performance of a classification deep neural network (DNN) in the output probability distribution space of the DNN, where CMI and the ratio between CMI and NCMI represent the intra-class concentration and inter-class separation of the DNN, respectively. By using NCMI to evaluate popular DNNs pretrained over ImageNet in the literature, it is shown that their validation accuracies over ImageNet validation data set are more or less inversely proportional to their NCMI values. Based on this observation, the standard deep learning (DL) framework is further modified to minimize the standard cross entropy function subject to an NCMI constraint, yielding CMI constrained deep learning (CMIC-DL). A novel alternating learning algorithm is proposed to solve such a constrained optimization problem. Extensive experiment results show that DNNs trained within CMIC-DL outperform the state-of-the-art models trained within the standard DL and other loss functions in the literature in terms of both accuracy and robustness against adversarial attacks. In addition, visualizing the evolution of learning process through the lens of CMI and NCMI is also advocated.
More
Translated text
Key words
Alternating minimization,concentration and separation,conditional mutual information (CMI),cross entropy (CE),deep learning (DL)
PDF
Bibtex
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper

要点】:论文提出了一种基于条件互信息约束的深度学习框架(CMIC-DL),通过引入条件互信息(CMI)和归一化条件互信息(NCMI)衡量深度神经网络在输出概率分布空间的分类性能,实现了更高的准确性和鲁棒性。

方法】:使用CMI和NCMI评估DNN在输出概率分布空间的分类性能,并根据NCMI值修改标准深度学习框架,通过约束优化问题提出CMIC-DL,并采用交替学习算法解决该问题。

实验】:在ImageNet验证数据集上进行的广泛实验表明,基于CMIC-DL训练的DNN在准确性和对抗攻击的鲁棒性方面均优于标准深度学习框架和其他文献中的损失函数方法。