CODiff: One-Step Diffusion Model for Camouflaged Object Detection
Abstract
Diffusion-based camouflaged object detection (COD) has recently shown great potential. In contrast to existing approaches that rely on multiple sample steps to refine the predicted masks, we propose CODiff, which reformulates the diffusion process to enable one-step mask prediction while maintaining competitive accuracy. Specifically, we first establish the theoretical feasibility of one-step sampling for COD. Based on this, we design a dedicated network for one-step inference with a global semantic guidance mechanism to guide the denoising process globally and hierarchical condition integration blocks to provide fine-grained structural semantics. In addition, we design a straight-forward regularization to learn better intermediate features by bridging the representation gap between the condition backbone and the diffusion model. Extensive experiments demonstrate that CODiff achieves state-of-the-art performance across multiple benchmarks, improving MAE by over 22\% on the challenging COD10K dataset. The code will be released upon publication.