数据受限条件下基于混合源域与目标域方法的生成对抗网络训练算法

Data-Constrained Generative Adversarial Networks Based on Mixing Source and Target Domain Methods

  • 摘要: 生成对抗网络(GAN)在人工智能领域持续推动着众多应用的发展。然而,在训练数据受限的情况下,现有迁移算法常难以充分学习目标域的特征,且面临生成结果多样性不足的问题。为此,提出一种混合源域与目标域的训练算法,通过隐式约束策略,在保留源域先验知识的同时,赋予模型必要的灵活性。该算法包含两项核心技术:面向判别器的双重适应性训练(SAT)与面向生成器的潜在分布扩展(ELD)。SAT通过隐式混合特征图,引导判别器的深层网络与预训练特征空间对齐,从而在保留源域判别逻辑的同时,抑制过拟合;ELD则通过潜在空间中的插值,将源域的高多样性分布与从目标域挖掘的分布相融合,以此增强模型的拟合能力并缓解模式崩溃问题。为验证所提方法的性能,在7种图像迁移任务上对比了该算法与多种现有迁移方法的生成质量与多样性。实验结果表明,ELD+SAT方法在所有任务上均取得了最优的FID分数,其性能显著优于MineGAN、FreezeD等主流迁移方法。该研究为数据受限条件下的GAN迁移训练提供了一种无需设计显式损失函数的新视角。

     

    Abstract: Generative adversarial network (GAN) continue to drive advancements in numerous artificial intelligence applications. However, in data-scarce scenarios, existing transfer algorithms often struggle to adequately capture target domain characteristics while suffering from insufficient diversity in generated results. To address these limitations, a training algorithm based on mixing source and target domains is proposed. Through an implicitly constrained strategy, it preserves prior knowledge from the source domain while endowing the model with the necessary flexibility. The algorithm incorporates two core techniques: Swap Adaptable Training (SAT) for the discriminator and Expanded Latent Distribution (ELD) for the generator. SAT implicitly mixes feature maps to align deep discriminator layers with the pre-trained feature space, thereby preserving source domain discrimination logic while mitigating overfitting. ELD enhances model fitting capacity and alleviates mode collapse by combining the high-diversity distribution of the source domain with distributions mined from the target domain through latent space interpolation. The generation quality and diversity of the proposed algorithm were evaluated against several existing transfer methods across seven image transfer tasks. Experimental results demonstrate that the ELD+SAT method achieves optimal Fréchet Inception Distance (FID) scores across all tasks, significantly outperforming mainstream transfer methods such as MineGAN and FreezeD. This study offers a novel perspective for GAN transfer training under data-limited conditions without requiring explicit loss function design.

     

/

返回文章
返回