前往小程序,Get更优阅读体验!
立即前往
首页
学习
活动
专区
工具
TVP
发布
社区首页 >专栏 >周志华:Learnware 将是机器学习的未来

周志华:Learnware 将是机器学习的未来

作者头像
新智元
发布2018-03-22 14:52:56
1.6K0
发布2018-03-22 14:52:56
举报
文章被收录于专栏:新智元新智元

【新智元导读】《机器学习》作者、南京大学教授周志华在本文中,针对当前机器学习环境适应低、数据共享难等局限,提出新概念 learnware(学件)。Learnware 具有 reusable(可重用)、evolvable(可演进)、comprehensible(可了解) 三大特点,如若能得以实现,强大的机器学习模型也能用小数据训练,数据保密问题也能得以缓解,更多终端用户都能像专家一样使用机器学习技术,并最终形成一个 learnware 市场。

Zhi-Hua Zhou is a professor at the Department of Computer Science and Technology, Nanjing University, China. He is the standing deputy director of the National Key Laboratory for Novel Software Technology, and founding director of LAMDA. He is a fellow of the AAAI, IEEE, IAPR and CCF, and also an ACM Distinguished Scientist. His main research interests are in artificial intelligence, machine learning and data mining.



Frontiers of Computer Science, 2016 DOI 10.1007/s11704-016-6906-3

PERSPECTIVE

Learnware: on the future of machine learning

Zhi-Hua ZHOU

National Key Laboratory for Novel Software Technology, Department of Computer Science &Technology, Nanjing University, Nanjing 210023, China

Current machine learning techniques have achieved great success; however, there are many deficiencies. First, to train a strong model, a large amount of training examples are required, whereas collecting the data, particularly data with labels, is expensive or even difficult in many real tasks. Second, once a model has been trained, if environment changes, which often happens in real tasks,the model can hardly perform well or even become useless. Third, the trained models are usually black-boxes, whereas people usually want to know what have been learned by the models, particularly in real tasks where decision reliability is crucial and rigorous judgments by human beings are critical.

In addition to the above deficiencies, there are several relevant issues requiring attention. First, some data have to be shared in most current machine learning studies if one hopes to pass helpful information from one task to another. The data privacy or data proprietary, however, usually disables public data sharing. Thus, it is hard for people to build their learning tasks based on the results of other people. Second, machine learning is still a kind of magic: Even with sufficient training data, most end users, except machine learning experts, can hardly produce strong models.

Considering the above issues, here we propose learnware. A learnware is a well-performed pre-trained machine learning model with a specification which explains the purpose and/or specialty of the model. The specification can be logic-based descriptions, and/or statistics that reveal the target to which the model aimed, and/or even a few simplified training samples that disclose the scenario for which the model was trained. The owner of a learnware can put it into a market, with little risk of data privacy leakage. As the comic illustrates, when a person is going to tackle a machine learning task, rather than build his model from scratch, he can do it in this way: Figure out his own requirement, and then browse/search the market, identify and take a good learnware whose specification matches his requirement. In some cases he can use the learnware directly, whereas in more cases he may need to use his own data to adapt/polish the learnware. Nevertheless, the whole process can be much less expensive and more efficient than building a model from scratch by himself.

For this purpose, a learnware should have at least three important properties: Reusable, Evolvable, and Comprehensible.

A learnware should be reusable,otherwise it can hardly be useful for other users. In particular, the pre-trained model should be able to be enhanced or adapted, by its new user through a slight modification or refinement using information, such as a small amount of training data, from the task of the new user. This process may be subtle: On one hand, one needs to avoid important learned knowledge being washed out by the refining; on the other hand, the model should have sufficient flexibility to incorporate necessary modification desired by its new user. There are some machine learning studies that can be regarded as preliminary attempt for this purpose, e.g., model adaptation [1], and transfer learning [2].

Evolvable means that the learnware should be able to get accustomed to environment change. If reusable is viewed as the learnware’s ability of passive adaptation driven by the user, evolvable can be viewed as its ability of active adaptation: The learnware should be able to perceive the environment change and do the adaptation by itself. There are at least three reasons for this need. First, the learning task of the new user is usually somewhat different from the original task for which the learnware was constructed, because one can hardly expect a learning task to exactly appear again. Second, the learnware specification and/or the user requirement can hardly be very accurate descriptions, and there may exist some gaps that the learnware must be able to get through. Third, many real environments are non-stationary and changing in nature, e.g., data distribution may change [3], new classes may occur [4,5], features may change [6], etc. These issues have also been emphasized to be tackled on the way toward robust artificial intelligence[7].

A learnware should be comprehensible; that means, the learning models need to be transparent to some extent, at least enabling the writing of specification for the learnware. For example, one needs to know what kind of target the learnware was trained for,how good the learnware performed, what specific problem structure the learnware can be applied, etc. On one hand, black-box models need to be made white; for this there are many efforts, e.g., trying to improve the comprehensibility of black-box models [8], or even produce accurate and comprehensible models [9].On the other hand, one needs to be able to write the specification that well describes a pre-trained machine learning model; unfortunately there is little study about this. Inspiration may be learned from the field of software engineering, where specification has a long history of study and application.

Note that in most cases, the end user may be unable to identify a single learnware which exactly matches his requirement; instead, he may find multiple learnwares and each meets a part. In such cases, ensemble methods [10] that combine multiple models to use may offer some solutions, just like the description of reusable ensemble [10] (pp.184) where reusable components are searched and put together, and only functional components that could not be found need to be constructed.

If learnwares come true, strong machine learning models can be achieved even for tasks with small data, because the models are built upon well-performed learnwares, and only a small amount of data are needed for adaptation or refinement. Data privacy will become a less serious issue because publicizing learnwares does not need to share the data. More importantly, it will enable common end users to attain tricky learning results that can only be attained by machine learning experts nowadays. Needless to say, a promising learnware industry will be open.

Acknowledgements The author wishes to thank Wang-Zhou Dai for converting a hand drawing sketch to comic, and thank LAMDA group members for helpful discussion.

References

1. Li N, Tsang I W, Zhou Z H. Efficient optimization of performance measures by classifier adaptation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013, 35(6): 1370–1382

2. Pan S J, Yang Q. A survey of transfer learning. IEEE Transactions on Knowledge and Data Engineering, 2010, 22(10): 1345–1359

3. Sugiyama M, Kawanabe M. Machine Learning in Non-Stationary Environments: Introduction to Covariate Shift Adaptation. Cambridge, MA: MIT Press, 2012

4. Da Q, Yu Y, Zhou Z H. Learning with augmented class by exploiting unlabeled data. In: Proceedings of the 28th AAAI Conference on Artificial Intelligence. 2014, 1760–1766

5. Mu X, Ting K M, Zhou Z H. Classification under streaming emerging new classes: a solution using completely random trees. CORR abs/1605.09131, 2016

6. Hou C, Zhou Z H. One-pass learning with incremental and decremental features. CORR abs/1605.09082, 2016

7. Dietterich T G. Towards robust artificial intelligence. In: Proceedings of the 30th AAAI Conference on Artificial Intelligence. 2016

8. Zhou Z H, Jiang Y, Chen S F. Extracting symbolic rules from trained neural network ensembles. AI Communications, 2003, 16(1): 3–15

9. Zhou Z H, Jiang Y. NeC4.5: Neural ensemble based C4.5. IEEE Transactions on Knowledge and Data Engineering, 2004, 16(6): 770–773

10. Zhou Z H. Ensemble Methods: Foundations and Algorithms. Boca Raton, FL: CRC Press, 2012

本文参与 腾讯云自媒体分享计划,分享自微信公众号。
原始发表:2016-06-02,如有侵权请联系 cloudcommunity@tencent.com 删除

本文分享自 新智元 微信公众号,前往查看

如有侵权,请联系 cloudcommunity@tencent.com 删除。

本文参与 腾讯云自媒体分享计划  ,欢迎热爱写作的你一起参与!

评论
登录后参与评论
0 条评论
热度
最新
推荐阅读
领券
问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档