[2603.21056] Semi-Supervised Learning with Balanced Deep Representation Distributions
About this article
Abstract page for arXiv paper 2603.21056: Semi-Supervised Learning with Balanced Deep Representation Distributions
Computer Science > Machine Learning arXiv:2603.21056 (cs) [Submitted on 22 Mar 2026] Title:Semi-Supervised Learning with Balanced Deep Representation Distributions Authors:Changchun Li, Ximing Li, Bingjie Zhang, Wenting Wang, Jihong Ouyang View a PDF of the paper titled Semi-Supervised Learning with Balanced Deep Representation Distributions, by Changchun Li and 4 other authors View PDF HTML (experimental) Abstract:Semi-Supervised Text Classification (SSTC) mainly works under the spirit of self-training. They initialize the deep classifier by training over labeled texts; and then alternatively predict unlabeled texts as their pseudo-labels and train the deep classifier over the mixture of labeled and pseudo-labeled texts. Naturally, their performance is largely affected by the accuracy of pseudo-labels for unlabeled texts. Unfortunately, they often suffer from low accuracy because of the margin bias problem caused by the large difference between representation distributions of labels in SSTC. To alleviate this problem, we apply the angular margin loss, and perform several Gaussian linear transformations to achieve balanced label angle variances, i.e., the variance of label angles of texts within the same label. More accuracy of predicted pseudo-labels can be achieved by constraining all label angle variances balanced, where they are estimated over both labeled and pseudo-labeled texts during self-training loops. With this insight, we propose a novel SSTC method, namely Sem...