专栏首页CreateAMinddeepmind 做通用人工智能的思路

deepmind 做通用人工智能的思路

Early Visual Concept Learning with Unsupervised Deep Learning

Abstract

Automated discovery of early visual concepts from raw image data is a major open challenge in AI research.

自动学习基础物体概念

Addressing this problem, we propose an unsupervised approach for learning disentangled representations of the underlying factors of variation. 属性分解

We draw inspiration from neuroscience, and show how this can be achieved in an unsupervised generative model by applying the same learning pressures as have been suggested to act in the ventral visual stream in the brain. 视觉脑神经启发

By enforcing redundancy reduction, encouraging statistical independence, and exposure to data with transform continuities analogous to those to which human infants are exposed, we obtain a variational autoencoder (VAE) framework capable of learning disentan gled factors. 婴儿类似的视觉环境

Our approach makes few assumptions and works well across a wide variety of datasets. Furthermore, our solution has useful emergent properties, such as zero-shot inference and an intuitive understanding of “objectness"。高级功能

1 Introduction 简介对实现类人智能的论述非常精彩

State-of-the-art AI approaches still struggle with some scenarios where humans excel [21],

such as knowledge transfer, where faster learning is achieved by reusing learnt representations for numerous tasks (Fig. 1A); or zero-shot inference, where reasoning about new data is enabled by recombining previously learnt factors (Fig. 1B).

属性组合推理快速学习判断

[21] suggest incorporating certain “start-up” abilities into deep models, such as intuitive understanding of physics, to help bootstrap learning in these scenarios.

Elaborating on this idea, we believe that learning basic visual concepts, such as the “objectness” of things in the world, and the ability to reason about objects in terms of the generative factors that specify their properties, is an important step towards building machines that learn and think like people.

物体概念及概念的相关属性。

We believe that this can be achieved by learning a disentangled posterior distribution of the generative factors of the observed sensory input by leveraging the wealth of unsupervised data [4, 21]. 属性分解

We wish to learn a representation where single latent units are sensitive to changes in single generative factors, while being relatively invariant to changes in other factors [4]. 单一变量属性

With a disentangled representation, knowledge about one factor could generalise to many configurations of other factors, thus capturing the “multiple explanatory factors” and “shared factors across tasks” priors suggested by [4].

Unsupervised disentangled factor learning from raw image data is a major open challenge in AI. Most previous attempts require a priori knowledge of the number and/or nature of the data generative factors [16, 25, 35, 34, 13, 20, 8, 33, 17]. This is infeasible in the real world, where the newborn learner may have no a priori knowledge and little to no supervision for discovering the generative factors. So far any purely unsupervised approaches to disentangled factor learning have not scaled well [11, 30, 9, 10].

婴儿最初视觉学习是完全无监督的。

We propose a deep unsupervised generative approach for disentangled factor learning inspired by neuroscience [2, 3, 24, 15]. We apply similar learning constraints to the model as have been suggested to act in the ventral visual stream in the brain [28]: redundancy reduction, an emphasis on learning statistically independent factors, and exposure to data with transform continuities analogous to those human infants are exposed to [2, 3]. We show that the application of such pressures to a deep unsupervised generative model can be realised in the variational autoencoder (VAE)framework [19, 26].

跟婴儿相似的视觉学习环境

Our main contributions are the following: 1) we show the importance of neuroscience inspired constraints (data continuity, redundancy reduction and statistical independence) for learning disentangled representations of continuous visual generative factors; 2) we devise a protocol to quantitatively compare the degree of disentanglement learnt by different models; and 3) we demonstrate how learning disentangled representations enables zero-shot inference and the emergence of basic visual concepts, such as “objectness”.

本文由zdx3578推荐。

本文分享自微信公众号 - CreateAMind(createamind),作者:zdx3578

原文出处及转载信息见文内详细说明,如有侵权,请联系 yunjia_community@tencent.com 删除。

原始发表时间:2016-11-13

本文参与腾讯云自媒体分享计划,欢迎正在阅读的你也加入,一起分享。

我来说两句

0 条评论
登录 后参与评论

相关文章

  • STCN

    https://www.arxiv-vanity.com/papers/1902.06568/

    用户1908973
  • 强化学习框架 IMPALA 介绍

    In this work we aim to solve a large collection of tasks using a single reinforc...

    用户1908973
  • 开源ALNS 自适应大邻域搜索(Adaptive Large Neighborhood Search)

    This package offers a general, well-documented and tested implementation of the ...

    用户1908973
  • AlexNet论文总结

    Introduction Preconditions ImageNet Objects in realistic settings exhibit cons...

    Tyan
  • 部署机器学习以帮助数字人性化:提高OpenStreetMap中的图像注释效率(CS HC)

    在发展中国家的农村地区进行人口定位吸引了人性化绘图项目的关注,因为规划措施对脆弱地区的影响很大。最近的努力已经解决了这个问题,即在空中图像中检测建筑物。但是,像...

    gaowanting
  • 随机供应和灵活消费者的最优动态机制设计(CS GT)

    我们考虑的问题是设计一个预期收益最大化机制,在T个时间步长内将k个品种的多个非易腐商品分配给灵活的消费者。在我们的模型中,每个品种的随机数量的商品在每个时间都可...

    Rosalie
  • How to Select an Object-Relational Mapping Tool for .NET

    The single, most important trap to watch out for when choosing an object-relatio...

    张善友
  • 2017美国数学建模MCM C题(大数据)翻译 “合作和导航”

    Traffic capacity is limited in many regions of the United States due to the numb...

    AI那点小事
  • 基于PfSPZ的疟疾疫苗生产的蚊子取放系统(CS RO)

    疟疾的治疗是一项全球性的健康挑战,而从该疾病的疫苗的广泛引入中受益匪浅。已经开发出一种利用寄生虫恶性疟原虫(Pf)的子孢子(SPZ)制备活体生物疫苗的方法,该子...

    时代在召唤
  • 【论文推荐】最新5篇图像分割相关论文—条件随机场和深度特征学习、移动端网络、长期视觉定位、主动学习、主动轮廓模型、生成对抗性网络

    【导读】专知内容组整理了最近五篇视觉图像分割(Image Segmentation)相关文章,为大家进行介绍,欢迎查看! 1. Conditional Rand...

    WZEARW

扫码关注云+社区

领取腾讯云代金券