Search papers, labs, and topics across Lattice.
This paper tackles feature collision in class-incremental learning (CIL) by framing the problem through a causal lens, identifying intra- and inter-task spurious correlations as the root cause. They propose a Probability of Necessity and Sufficiency (PNS)-based regularization method, termed CPNS, to guide feature expansion by quantifying causal completeness and inter-task separability. A dual-scope counterfactual generator based on twin networks is introduced to minimize intra- and inter-task PNS risk, leading to more robust and separable features.
Stop CIL models from catastrophically forgetting by explicitly minimizing causal incompleteness within tasks and maximizing separability between tasks.
Current expansion-based methods for Class Incremental Learning (CIL) effectively mitigate catastrophic forgetting by freezing old features. However, such task-specific features learned from the new task may collide with the old features. From a causal perspective, spurious feature correlations are the main cause of this collision, manifesting in two scopes: (i) guided by empirical risk minimization (ERM), intra-task spurious correlations cause task-specific features to rely on shortcut features. These non-robust features are vulnerable to interference, inevitably drifting into the feature space of other tasks; (ii) inter-task spurious correlations induce semantic confusion between visually similar classes across tasks. To address this, we propose a Probability of Necessity and Sufficiency (PNS)-based regularization method to guide feature expansion in CIL. Specifically, we first extend the definition of PNS to expansion-based CIL, termed CPNS, which quantifies both the causal completeness of intra-task representations and the separability of inter-task representations. We then introduce a dual-scope counterfactual generator based on twin networks to ensure the measurement of CPNS, which simultaneously generates: (i) intra-task counterfactual features to minimize intra-task PNS risk and ensure causal completeness of task-specific features, and (ii) inter-task interfering features to minimize inter-task PNS risk, ensuring the separability of inter-task representations. Theoretical analyses confirm its reliability. The regularization is a plug-and-play method for expansion-based CIL to mitigate feature collision. Extensive experiments demonstrate the effectiveness of the proposed method.