专栏首页专知【论文推荐】最新八篇生成对抗网络相关论文—BRE、图像合成、多模态图像生成、非配对多域图、注意力、对抗特征增强、深度对抗性训练

【论文推荐】最新八篇生成对抗网络相关论文—BRE、图像合成、多模态图像生成、非配对多域图、注意力、对抗特征增强、深度对抗性训练

【导读】专知内容组整理了最近八篇生成对抗网络(Generative Adversarial Networks )相关文章,为大家进行介绍,欢迎查看!

1.Improving GAN Training via Binarized Representation Entropy (BRE) Regularization(通过二值化表示熵(BRE)正则改进GAN训练)



作者:Yanshuai Cao,Gavin Weiguang Ding,Kry Yik-Chau Lui,Ruitong Huang

Published as a conference paper at the 6th International Conference on Learning Representations (ICLR 2018)

摘要:We propose a novel regularizer to improve the training of Generative Adversarial Networks (GANs). The motivation is that when the discriminator D spreads out its model capacity in the right way, the learning signals given to the generator G are more informative and diverse. These in turn help G to explore better and discover the real data manifold while avoiding large unstable jumps due to the erroneous extrapolation made by D. Our regularizer guides the rectifier discriminator D to better allocate its model capacity, by encouraging the binary activation patterns on selected internal layers of D to have a high joint entropy. Experimental results on both synthetic data and real datasets demonstrate improvements in stability and convergence speed of the GAN training, as well as higher sample quality. The approach also leads to higher classification accuracies in semi-supervised learning.

期刊:arXiv, 2018年5月9日

网址

http://www.zhuanzhi.ai/document/1daa64655b1a4199334be9631bb7dc98

2.MC-GAN: Multi-conditional Generative Adversarial Network for Image Synthesis(MC-GAN:多条件生成对抗网络的图像合成)



作者:Hyojin Park,YoungJoon Yoo,Nojun Kwak

机构:Seoul National University

摘要:In this paper, we introduce a new method for generating an object image from text attributes on a desired location, when the base image is given. One step further to the existing studies on text-to-image generation mainly focusing on the object's appearance, the proposed method aims to generate an object image preserving the given background information, which is the first attempt in this field. To tackle the problem, we propose a multi-conditional GAN (MC-GAN) which controls both the object and background information jointly. As a core component of MC-GAN, we propose a synthesis block which disentangles the object and background information in the training stage. This block enables MC-GAN to generate a realistic object image with the desired background by controlling the amount of the background information from the given base image using the foreground information from the text attributes. From the experiments with Caltech-200 bird and Oxford-102 flower datasets, we show that our model is able to generate photo-realistic images with a resolution of 128 x 128. The source code of MC-GAN is available soon.

期刊:arXiv, 2018年5月8日

网址

http://www.zhuanzhi.ai/document/5c97977b9f8ad4b237d8b6f9ba75497f

3.MEGAN: Mixture of Experts of Generative Adversarial Networks for Multimodal Image Generation(MEGAN: 多模态图像生成对抗网络专家的混合)



作者:David Keetae Park,Seungjoo Yoo,Hyojin Bahng,Jaegul Choo,Noseong Park

27th International Joint Conference on Artificial Intelligence (IJCAI 2018)

机构:Korea University,University of North Carolina at Charlotte

摘要:Recently, generative adversarial networks (GANs) have shown promising performance in generating realistic images. However, they often struggle in learning complex underlying modalities in a given dataset, resulting in poor-quality generated images. To mitigate this problem, we present a novel approach called mixture of experts GAN (MEGAN), an ensemble approach of multiple generator networks. Each generator network in MEGAN specializes in generating images with a particular subset of modalities, e.g., an image class. Instead of incorporating a separate step of handcrafted clustering of multiple modalities, our proposed model is trained through an end-to-end learning of multiple generators via gating networks, which is responsible for choosing the appropriate generator network for a given condition. We adopt the categorical reparameterization trick for a categorical decision to be made in selecting a generator while maintaining the flow of the gradients. We demonstrate that individual generators learn different and salient subparts of the data and achieve a multiscale structural similarity (MS-SSIM) score of 0.2470 for CelebA and a competitive unsupervised inception score of 8.33 in CIFAR-10.

期刊:arXiv, 2018年5月8日

网址

http://www.zhuanzhi.ai/document/4e0326f4e3c2d60536da6874c9c8fc63

4.Unpaired Multi-Domain Image Generation via Regularized Conditional GANs(利用正则化的条件GANs生成非配对的多域图)



作者:Xudong Mao,Qing Li

机构:City University of Chinese Hong Kong

摘要:In this paper, we study the problem of multi-domain image generation, the goal of which is to generate pairs of corresponding images from different domains. With the recent development in generative models, image generation has achieved great progress and has been applied to various computer vision tasks. However, multi-domain image generation may not achieve the desired performance due to the difficulty of learning the correspondence of different domain images, especially when the information of paired samples is not given. To tackle this problem, we propose Regularized Conditional GAN (RegCGAN) which is capable of learning to generate corresponding images in the absence of paired training data. RegCGAN is based on the conditional GAN, and we introduce two regularizers to guide the model to learn the corresponding semantics of different domains. We evaluate the proposed model on several tasks for which paired training data is not given, including the generation of edges and photos, the generation of faces with different attributes, etc. The experimental results show that our model can successfully generate corresponding images for all these tasks, while outperforms the baseline methods. We also introduce an approach of applying RegCGAN to unsupervised domain adaptation.

期刊:arXiv, 2018年5月7日

网址

http://www.zhuanzhi.ai/document/2782f0094d1961fad34e81a96a114c56

5.Attentive Generative Adversarial Network for Raindrop Removal from a Single Image(从单个图像去除雨滴的注意力的生成对抗网络



作者:Rui Qian,Robby T. Tan,Wenhan Yang,Jiajun Su,Jiaying Liu

CVPR2018 Spotlight

机构:Peking University,National University of Singapore

摘要:Raindrops adhered to a glass window or camera lens can severely hamper the visibility of a background scene and degrade an image considerably. In this paper, we address the problem by visually removing raindrops, and thus transforming a raindrop degraded image into a clean one. The problem is intractable, since first the regions occluded by raindrops are not given. Second, the information about the background scene of the occluded regions is completely lost for most part. To resolve the problem, we apply an attentive generative network using adversarial training. Our main idea is to inject visual attention into both the generative and discriminative networks. During the training, our visual attention learns about raindrop regions and their surroundings. Hence, by injecting this information, the generative network will pay more attention to the raindrop regions and the surrounding structures, and the discriminative network will be able to assess the local consistency of the restored regions. This injection of visual attention to both generative and discriminative networks is the main contribution of this paper. Our experiments show the effectiveness of our approach, which outperforms the state of the art methods quantitatively and qualitatively.

期刊:arXiv, 2018年5月6日

网址

http://www.zhuanzhi.ai/document/911b81fc00817b37b029c73318675209

6.Adversarial Feature Augmentation for Unsupervised Domain Adaptation(非监督域适应的对抗特征增强)



作者:Riccardo Volpi,Pietro Morerio,Silvio Savarese,Vittorio Murino

Accepted to CVPR 2018

机构:Universita di Verona,Stanford University

摘要:Recent works showed that Generative Adversarial Networks (GANs) can be successfully applied in unsupervised domain adaptation, where, given a labeled source dataset and an unlabeled target dataset, the goal is to train powerful classifiers for the target samples. In particular, it was shown that a GAN objective function can be used to learn target features indistinguishable from the source ones. In this work, we extend this framework by (i) forcing the learned feature extractor to be domain-invariant, and (ii) training it through data augmentation in the feature space, namely performing feature augmentation. While data augmentation in the image space is a well established technique in deep learning, feature augmentation has not yet received the same level of attention. We accomplish it by means of a feature generator trained by playing the GAN minimax game against source features. Results show that both enforcing domain-invariance and performing feature augmentation lead to superior or comparable performance to state-of-the-art results in several unsupervised domain adaptation benchmarks.

期刊:arXiv, 2018年5月4日

网址

http://www.zhuanzhi.ai/document/74ec78227fe38592dd361ff04f049d4e

7.Boosting Noise Robustness of Acoustic Model via Deep Adversarial Training(通过深度对抗性训练提高声学模型的噪声鲁棒性)



作者:Bin Liu,Shuai Nie,Yaping Zhang,Dengfeng Ke,Shan Liang,Wenju Liu1

机构:University of Chinese Academy of Sciences,Beijing Forestry University

摘要:In realistic environments, speech is usually interfered by various noise and reverberation, which dramatically degrades the performance of automatic speech recognition (ASR) systems. To alleviate this issue, the commonest way is to use a well-designed speech enhancement approach as the front-end of ASR. However, more complex pipelines, more computations and even higher hardware costs (microphone array) are additionally consumed for this kind of methods. In addition, speech enhancement would result in speech distortions and mismatches to training. In this paper, we propose an adversarial training method to directly boost noise robustness of acoustic model. Specifically, a jointly compositional scheme of generative adversarial net (GAN) and neural network-based acoustic model (AM) is used in the training phase. GAN is used to generate clean feature representations from noisy features by the guidance of a discriminator that tries to distinguish between the true clean signals and generated signals. The joint optimization of generator, discriminator and AM concentrates the strengths of both GAN and AM for speech recognition. Systematic experiments on CHiME-4 show that the proposed method significantly improves the noise robustness of AM and achieves the average relative error rate reduction of 23.38% and 11.54% on the development and test set, respectively.

期刊:arXiv, 2018年5月2日

网址

http://www.zhuanzhi.ai/document/9dd23e2b343ed994cf5e6143700df612

8.Controllable Generative Adversarial Network(可控的生成对抗网络)



作者:Minhyeok Lee,Junhee Seok

机构:Korea University

摘要:Recently introduced generative adversarial network (GAN) has been shown numerous promising results to generate realistic samples. The essential task of GAN is to control the features of samples generated from a random distribution. While the current GAN structures, such as conditional GAN, successfully generate samples with desired major features, they often fail to produce detailed features that bring specific differences among samples. To overcome this limitation, here we propose a controllable GAN (ControlGAN) structure. By separating a feature classifier from a discriminator, the generator of ControlGAN is designed to learn generating synthetic samples with the specific detailed features. Evaluated with multiple image datasets, ControlGAN shows a power to generate improved samples with well-controlled features. Furthermore, we demonstrate that ControlGAN can generate intermediate features and opposite features for interpolated and extrapolated input labels that are not used in the training process. It implies that ControlGAN can significantly contribute to the variety of generated samples.

期刊:arXiv, 2018年5月2日

网址

http://www.zhuanzhi.ai/document/a8cf307e26ee5cbbc1af9a940b7bcc7f

-END-

本文分享自微信公众号 - 专知(Quan_Zhuanzhi),作者:专知内容组

原文出处及转载信息见文内详细说明,如有侵权,请联系 yunjia_community@tencent.com 删除。

原始发表时间:2018-05-14

本文参与腾讯云自媒体分享计划,欢迎正在阅读的你也加入,一起分享。

我来说两句

0 条评论
登录 后参与评论

相关文章

  • 深度学习在计算机视觉领域的前沿进展

    用户1737318
  • BAT机器学习面试1000题系列(第76~149题)

    76、看你是搞视觉的,熟悉哪些CV框架,顺带聊聊CV最近五年的发展史如何?深度学习 DL应用 难 原英文:adeshpande3.github.io 作者:Ad...

    用户1332428
  • 2020腾讯犀牛鸟精英人才培养计划课题介绍(八)——视觉及多媒体计算

    ? 精英人才培养计划是一项校企联合人才培养项目,入选学生将受到业界顶尖技术团队与高校导师的联合指导及培养。培养期间,学生将获得3个月以上到访腾讯开展科研访问的...

    腾讯高校合作
  • 超100篇!CVPR 2020最全GAN论文梳理汇总!

    下述论文已分类打包好!共116篇,事实上仍有一些GAN论文未被包含入内,比如笔者发推文时,又看到一篇《Rotate-and-Render: Unsupervis...

    公众号机器学习与生成对抗网络
  • 生成对抗网络也需要注意力机制

    尽管传统的生成对抗网络可以生成相当逼真的图像,但是它们无法捕获到图像中的长距离依赖。这些传统的生成对抗网络在不包含太多的结构和几何信息的图像上效果是不错的(例如...

    机器之心
  • 干货 | 深度学习最强资源推荐:一文看尽 GAN 的前世今生

    AI 科技评论按,生成对抗网络是当前最热门的技术之一,它掀起了一场技术革命,取得了很多重大的突破。不久前,伊利诺伊大学香槟分校的学生 Ajay Uppili A...

    AI研习社
  • 7 Papers | KDD2019最佳论文;AutoML SOTA 综述

    摘要:在本文中,研究者深入探索了真实世界图谱谱密度的核心。他们借用了凝聚态物理学中开发的工具,并添加了新的适应性来处理常见图形的谱特征。他们计算了单个计算节点上...

    机器之心
  • CVPR 2020 | 几篇GAN在low-level vision中的应用论文

    【图像分离、去雨/反射/阴影等】Deep Adversarial Decomposition: A Unified Framework for Separat...

    公众号机器学习与生成对抗网络
  • GANs有嘻哈:一次学完10个GANs明星模型(附视频)

    以“左右互搏”的观念为人所知,GANs从概念开始就让人激动不已。不过从GANs被提出的第一天起,深度学习圈就出了个“大坑”,而且越挖越大、完全没有被填满的趋势。

    大数据文摘
  • ECCV 2020 | 腾讯 AI Lab 16篇入选论文解读

    来自Tencent AI实验室。本文主要介绍 ECCV 2020 中腾讯 AI Lab 16篇入选论文。

    深度学习技术前沿公众号博主
  • 今日 Paper | TRANSFORMER结构;图像到图像翻译;缺失标注目标检测解决;GAN水下图像颜色校正等

    作者:Nikita Kitaev / Lukasz Kaiser / Anselm Levskaya

    AI科技评论
  • 今日 Paper | 动态图像检索;实时场景文本定位;感知场景表示;双重网络等

    论文名称:Sketch Less for More: On-the-Fly Fine-Grained Sketch Based Image Retrieval

    AI科技评论
  • 7 Papers | 周志华等NeuralPS2019接收论文;网易用GAN建立游戏角色

    论文 1:Deep Learning: Computational Aspects

    机器之心
  • 17篇入选CVPR 2020,腾讯优图 9 篇精选论文详解

    全球计算机视觉顶级会议CVPR2020 (IEEE Conference on Computer Vision and Pattern Recognition...

    AI科技评论
  • 资源 | 生成对抗网络及其变体的论文汇总

    选自Deephunt 作者:Avinash Hindupur 参与:黄小天、蒋思源 生成对抗网络(GAN)是近段时间以来最受研究者关注的机器学习方法之一,深度学...

    机器之心
  • 悉尼大学陶大程:遗传对抗生成网络有效解决GAN两大痛点

    悉尼大学教授、澳大利亚科学院院士、优必选人工智能首席科学家陶大程博士在9月20日的AI WORLD 2018世界人工智能峰会上发表《AI破晓:机遇与挑战》的主题...

    数据派THU
  • 7 papers | 3篇综述论文,全面了解机器阅读理解、图表征学习等

    1.标题:Machine Learning Testing: Survey, Landscapes and Horizons

    机器之心
  • 学术资讯|腾讯优图17篇论文入选CVPR 2020,含类比学习、人脸识别、行人重识别等领域

    全球计算机视觉顶级会议CVPR2020  (IEEE Conference on Computer Vision and Pattern Recognition...

    优图实验室
  • 今日 Paper | CVPR 2020 论文推荐:Social-STGCNN;无偏场景图生成;深度人脸识别;4D 关联图等

    CVPR 2020 | Social-STGCNN:一种用于行人轨迹预测的社会时空图卷积神经网络

    AI科技评论

扫码关注云+社区

领取腾讯云代金券