论文标题

FAMLP:用于域概括的频率感知MLP架构

FAMLP: A Frequency-Aware MLP-Like Architecture For Domain Generalization

论文作者

Zheng, Kecheng, Cao, Yang, Zhu, Kai, Zhao, Ruijing, Zha, Zheng-Jun

论文摘要

最近重新审视了完全基于多层感知器建立的MLP型模型,表现出与变压器的可比性能。由于网络能力与大规模识别任务中的网络能力与效率之间的良好权衡,这是最有前途的架构之一。但是,由于域信息的广泛保留,其对异质任务的概括性能不如其他体系结构(例如CNN和变形金刚)。为了解决这个问题,我们提出了一种新颖的频率感知MLP体系结构,其中特定于域特异性的特征在转换的频域中被过滤,从而增强了标签预测的不变描述符。具体而言,我们设计了一个自适应傅里叶滤波器层,其中利用可学习的频率过滤器来通过优化真实和虚构零件来调整幅度分布。进一步提出了一个低级增强模块,以通过添加来自SVD分解的低频组件来纠正过滤的特征。最后,利用动量更新策略来稳定优化模型参数和通过加权历史状态的输出蒸馏来波动。据我们所知,我们是第一个提出类似MLP的主链来进行域泛化的人。在三个基准上进行的广泛实验表明,概括性的表现显着,表现分别优于最先进的方法,分别为3%,4%和9%。

MLP-like models built entirely upon multi-layer perceptrons have recently been revisited, exhibiting the comparable performance with transformers. It is one of most promising architectures due to the excellent trade-off between network capability and efficiency in the large-scale recognition tasks. However, its generalization performance to heterogeneous tasks is inferior to other architectures (e.g., CNNs and transformers) due to the extensive retention of domain information. To address this problem, we propose a novel frequency-aware MLP architecture, in which the domain-specific features are filtered out in the transformed frequency domain, augmenting the invariant descriptor for label prediction. Specifically, we design an adaptive Fourier filter layer, in which a learnable frequency filter is utilized to adjust the amplitude distribution by optimizing both the real and imaginary parts. A low-rank enhancement module is further proposed to rectify the filtered features by adding the low-frequency components from SVD decomposition. Finally, a momentum update strategy is utilized to stabilize the optimization to fluctuation of model parameters and inputs by the output distillation with weighted historical states. To our best knowledge, we are the first to propose a MLP-like backbone for domain generalization. Extensive experiments on three benchmarks demonstrate significant generalization performance, outperforming the state-of-the-art methods by a margin of 3%, 4% and 9%, respectively.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源