基于扩散模型的情绪识别对抗样本生成方法
首发时间:2026-03-02
摘要:现有针对人脸情绪识别(Facial Expression Recognition, FER)系统的对抗样本生成方法普遍存在视觉失真严重、表情不自然等问题,难以在实际场景中兼顾高攻击性与高隐蔽性。为解决传统像素级扰动导致图像质量低下的问题,本文提出一种基于扩散模型的对抗样本生成方法。该方法将扰动优化迁移至高级语义空间,通过在扩散自编码器提取的语义编码上施加受控扰动,并结合情绪攻击和动作单元(AU)合理性等多目标损失函数,引导情绪表征向目标类别演化,并生成高保真、生理合理的对抗样本。最终通过仿真实验,证明本文所提方法在攻击成功率、视觉质量等多个关键指标上均显著优于现有先进基线方法,生成的对抗样本能够有效欺骗主流FER模型,同时在视觉上保持高度自然,为人脸情绪隐私的主动防护提供了强有力的技术支持。
For information in English, please click here
Generating adversarial samples for emotion recognition based on diffusion model
Abstract:Existing adversarial example generation methods for Facial Expression Recognition (FER) systems generally suffer from severe visual distortion and unnatural expressions, making it difficult to achieve both high attackability and high concealment in real-world scenarios. To address the issue of low image quality caused by traditional pixel-level perturbations, this paper proposes an adversarial example generation method based on a diffusion model. This method transfers perturbation optimization to the high-level semantic space, applies controlled perturbations on the semantic encoding extracted by a diffusion autoencoder, and combines multi-objective loss functions such as emotional attack and action unit (AU) rationality to guide the evolution of emotional representations towards the target category, generating high-fidelity and physiologically plausible adversarial examples. Finally, simulation experiments demonstrate that the proposed method significantly outperforms existing advanced baseline methods in multiple key metrics, including attack success rate and visual quality. The generated adversarial examples can effectively deceive mainstream FER models while maintaining high visual naturalness, providing strong technical support for the active protection of facial expression privacy.
Keywords: Emotion recognition adversarial example generation diffusion model
基金:
引用

No.****
同行评议
勘误表
基于扩散模型的情绪识别对抗样本生成方法
评论
全部评论