Learning Tight Rejection Boundaries without Negatives for Strict One-Class Audio Deepfake Detection
Abstract
The rapid evolution of audio deepfakes requires robust detection capable of generalizing to unseen attacks. One-class learning offers inherent robustness for this task by characterizing real speech distributions to detect anomalies. However, establishing a compact decision boundary without spoof supervision remains a fundamental challenge. Existing relaxed approaches often compromise this strictness by introducing auxiliary negative samples, which biases the boundary toward seen artifacts and degrades generalization to unseen attacks. To address this, we propose CA-SOADD, a framework that refines the acceptance region by constructing off-manifold boundary probes. Our proposed centroid-anchored tri-objective learning paradigm simultaneously enforces centroid compactness and a centroid-referenced margin against these probes, thereby explicitly tightening the acceptance region without treating them as an explicit negative class. We further extend the framework to heterogeneous settings through domain-conditioned centroids. Experiments on ASVSpoof and MLAAD benchmarks demonstrate that our strict real-only method consistently outperforms strong baselines under unseen attack types and domain shifts, with its effectiveness further validated through extensive ablation studies.