学习
实践
活动
工具
TVP
写文章
专栏首页专知【论文推荐】最新六篇自动问答相关论文—排序函数、文本摘要评估、信息抽取框架、层次递归编码器、半监督问答

【论文推荐】最新六篇自动问答相关论文—排序函数、文本摘要评估、信息抽取框架、层次递归编码器、半监督问答

【导读】既前两天推出十三篇自动问答(Question Answering)相关文章,专知内容组今天又推出六篇自动问答相关文章,为大家进行介绍,欢迎查看!

14. Training a Ranking Function for Open-Domain Question Answering(训练排序函数对开放式问题进行回答)



作者:Phu Mon Htut,Samuel R. Bowman,Kyunghyun Cho

机构:New York University

摘要:In recent years, there have been amazing advances in deep learning methods for machine reading. In machine reading, the machine reader has to extract the answer from the given ground truth paragraph. Recently, the state-of-the-art machine reading models achieve human level performance in SQuAD which is a reading comprehension-style question answering (QA) task. The success of machine reading has inspired researchers to combine information retrieval with machine reading to tackle open-domain QA. However, these systems perform poorly compared to reading comprehension-style QA because it is difficult to retrieve the pieces of paragraphs that contain the answer to the question. In this study, we propose two neural network rankers that assign scores to different passages based on their likelihood of containing the answer to a given question. Additionally, we analyze the relative importance of semantic similarity and word level relevance matching in open-domain QA.

期刊:arXiv, 2018年4月12日

网址

http://www.zhuanzhi.ai/document/bd075132f2b950ddfe3129606b6695cd

15. A Semantic QA-Based Approach for Text Summarization Evaluation(一种基于QA的文本摘要评估方法)



作者:Ping Chen,Fei Wu,Tong Wang,Wei Ding

机构:University of Massachusetts Boston

摘要:Many Natural Language Processing and Computational Linguistics applications involves the generation of new texts based on some existing texts, such as summarization, text simplification and machine translation. However, there has been a serious problem haunting these applications for decades, that is, how to automatically and accurately assess quality of these applications. In this paper, we will present some preliminary results on one especially useful and challenging problem in NLP system evaluation: how to pinpoint content differences of two text passages (especially for large pas-sages such as articles and books). Our idea is intuitive and very different from existing approaches. We treat one text passage as a small knowledge base, and ask it a large number of questions to exhaustively identify all content points in it. By comparing the correctly answered questions from two text passages, we will be able to compare their content precisely. The experiment using 2007 DUC summarization corpus clearly shows promising results.

期刊:arXiv, 2018年4月11日

网址

http://www.zhuanzhi.ai/document/bd075132f2b950ddfe3129606b6695cd

16. QA4IE: A Question Answering based Framework for Information Extraction(QA4IE:一个基于问答的信息抽取框架)



作者:Lin Qiu,Hao Zhou,Yanru Qu,Weinan Zhang,Suoheng Li,Shu Rong,Dongyu Ru,Lihua Qian,Kewei Tu,Yong Yu

机构:Shanghai Jiao Tong University

摘要:Information Extraction (IE) refers to automatically extracting structured relation tuples from unstructured texts. Common IE solutions, including Relation Extraction (RE) and open IE systems, can hardly handle cross-sentence tuples, and are severely restricted by limited relation types as well as informal relation specifications (e.g., free-text based relation tuples). In order to overcome these weaknesses, we propose a novel IE framework named QA4IE, which leverages the flexible question answering (QA) approaches to produce high quality relation triples across sentences. Based on the framework, we develop a large IE benchmark with high quality human evaluation. This benchmark contains 293K documents, 2M golden relation triples, and 636 relation types. We compare our system with some IE baselines on our benchmark and the results show that our system achieves great improvements.

期刊:arXiv, 2018年4月10日

网址

http://www.zhuanzhi.ai/document/100545a8407b2ec4d748cbf7579edbeb

17. Learning to Rank Question-Answer Pairs using Hierarchical Recurrent Encoder with Latent Topic Clustering(学习用层次式递归编码器和潜在的主题聚类对问题回答对进行排序)



作者:Seunghyun Yoon,Joongbo Shin,Kyomin Jung

机构:Seoul National University

摘要:In this paper, we propose a novel end-to-end neural architecture for ranking candidate answers, that adapts a hierarchical recurrent neural network and a latent topic clustering module. With our proposed model, a text is encoded to a vector representation from an word-level to a chunk-level to effectively capture the entire meaning. In particular, by adapting the hierarchical structure, our model shows very small performance degradations in longer text comprehension while other state-of-the-art recurrent neural network models suffer from it. Additionally, the latent topic clustering module extracts semantic information from target samples. This clustering module is useful for any text related tasks by allowing each data sample to find its nearest topic cluster, thus helping the neural network model analyze the entire data. We evaluate our models on the Ubuntu Dialogue Corpus and consumer electronic domain question answering dataset, which is related to Samsung products. The proposed model shows state-of-the-art results for ranking question-answer pairs.

期刊:arXiv, 2018年4月9日

网址

http://www.zhuanzhi.ai/document/3ed10f562bf0cb5d81c85a7ee339969f

18. Simple and Effective Semi-Supervised Question Answering(简单有效的半监督问答)



作者:Bhuwan Dhingra,Danish Pruthi,Dheeraj Rajagopal

机构:Carnegie Mellon University

摘要:Recent success of deep learning models for the task of extractive Question Answering (QA) is hinged on the availability of large annotated corpora. However, large domain specific annotated corpora are limited and expensive to construct. In this work, we envision a system where the end user specifies a set of base documents and only a few labelled examples. Our system exploits the document structure to create cloze-style questions from these base documents; pre-trains a powerful neural network on the cloze style questions; and further fine-tunes the model on the labeled examples. We evaluate our proposed system across three diverse datasets from different domains, and find it to be highly effective with very little labeled data. We attain more than 50% F1 score on SQuAD and TriviaQA with less than a thousand labelled examples. We are also releasing a set of 3.2M cloze-style questions for practitioners to use while building QA systems.

期刊:arXiv, 2018年4月3日

网址

http://www.zhuanzhi.ai/document/6036d026edb9f761e3a61d4ac11b61bc

19. Pay More Attention - Neural Architectures for Question-Answering(更多的关注-基于神经结构的问答)



作者:Zia Hasan,Sebastian Fischer

摘要:Machine comprehension is a representative task of natural language understanding. Typically, we are given context paragraph and the objective is to answer a question that depends on the context. Such a problem requires to model the complex interactions between the context paragraph and the question. Lately, attention mechanisms have been found to be quite successful at these tasks and in particular, attention mechanisms with attention flow from both context-to-question and question-to-context have been proven to be quite useful. In this paper, we study two state-of-the-art attention mechanisms called Bi-Directional Attention Flow (BiDAF) and Dynamic Co-Attention Network (DCN) and propose a hybrid scheme combining these two architectures that gives better overall performance. Moreover, we also suggest a new simpler attention mechanism that we call Double Cross Attention (DCA) that provides better results compared to both BiDAF and Co-Attention mechanisms while providing similar performance as the hybrid scheme. The objective of our paper is to focus particularly on the attention layer and to suggest improvements on that. Our experimental evaluations show that both our proposed models achieve superior results on the Stanford Question Answering Dataset (SQuAD) compared to BiDAF and DCN attention mechanisms.

期刊:arXiv, 2018年3月25日

网址

http://www.zhuanzhi.ai/document/9923a1c986b1e65f0eb7060a305c864d

-END-

专 · 知

文章分享自微信公众号:
专知

本文参与 腾讯云自媒体分享计划 ,欢迎热爱写作的你一起参与!

作者:专知内容组
原始发表时间:2018-05-10
如有侵权,请联系 cloudcommunity@tencent.com 删除。
登录 后参与评论
0 条评论

相关文章

  • AAAI2019 | 腾讯AI Lab详解自然语言处理领域三大研究方向及入选论文

    美国人工智能年会(the Association for the Advance of Artificial Intelligence),简称AAAI,是人工智...

    腾讯技术工程官方号
  • 学界 | NLP年度盛宴EMNLP/CoNLL 2018最精彩论文精选(摘要+评论)

    每年,全球最顶尖的NLP研究者们会在这两个大会上展示最前沿的NLP科研成果。然而,动辄上千篇的论文却令很多关注NLP领域最新动态的小伙伴望而却步。

    大数据文摘
  • 学界 | NLP年度盛宴EMNLP/CoNLL 2018最精彩论文精选(摘要+评论)

    每年,全球最顶尖的NLP研究者们会在这两个大会上展示最前沿的NLP科研成果。然而,动辄上千篇的论文却令很多关注NLP领域最新动态的小伙伴望而却步。

    zenRRan
  • (含源码)「自然语言处理(NLP)」Question Answering(QA)论文整理(七)

    本次文章主要介绍了ERNIE-GEN(语言生成任务)、统一预训练语言模型(UniLM)、问答系统数据集(CoQA)、端到端神经生成问答(GENQA)、生成式问答...

    ShuYini
  • 3 天,我把 NLP 中的预训练模型、图神经网络、模型压缩、知识图谱彻底撸清楚了!

    大家都知道NLP近几年非常火,而且发展也特别快。那些耳熟的BERT、GPT-3、图神经网络、知识图谱等技术实际上也就是这几年发展起来的,特别像图神经网络在这两年...

    zenRRan
  • (含源码)问答对生成(QAG)| 你竟还在手工梳理问答对!?

    AI时代,不同的行业都有自己的智能客服,比如银行智能客服、导购智能客服、后期服务支持智能客服等。训练客服机器人需要利用高质量、大量业务场景问答对(QA)作为语...

    ShuYini
  • 最新综述文章推荐:自然语言生成、深度学习算法、多媒体大数据分析

    【导读】专知内容组整理了最近人工智能领域相关期刊的5篇最新综述文章,为大家进行介绍,欢迎查看! 1 ▌自然语言生成综述:任务,应用,评价 ---- ---- ...

    WZEARW
  • 【NLP论文速递&&源码】Bert模型优化、自回归预训练、键值记忆网络、大规模问答系统训练

    本次论文分享内容主要包括以下内容:鲁棒优化Bert模型(RoBERTa)、自回归预训练模型(XLNet)、无监督多任务学习语言模型、生成预训练语言理解、深层上下...

    ShuYini
  • 终于来了!这份NLP算法工程师学习路线yyds!

    大家都知道NLP近几年非常火,而且相关技术发展也特别快,像BERT、GPT-3、图神经网络、知识图谱等技术被大量应用于项目实践中,这也推动了NLP在产业中的持续...

    zenRRan
  • 深度 | 腾讯AI Lab独家解析ACL 2017,聚焦三大研究领域

    本文转载自「腾讯AI实验室」,搜索「tencent_ailab」即可关注 本文是腾讯AI Lab微信号第三篇文章,深度解析本届ACL热门研究。文章第一部分是三大...

    AI科技评论
  • 专栏 | 腾讯AI Lab独家解析:深度聚焦ACL 2017三大研究领域

    机器之心
  • 独家解析 ACL 2017,聚焦三大研究领域

    本文深度解析本届 ACL 热门研究。文章第一部分是三大前沿领域重点文章解析,包括信息抽取、问答系统和机器翻译等。第二部分是 ACL 简介及我们 NLP 团队首次...

    腾讯AI实验室
  • 7 Papers | KDD2019最佳论文;AutoML SOTA 综述

    摘要:在本文中,研究者深入探索了真实世界图谱谱密度的核心。他们借用了凝聚态物理学中开发的工具,并添加了新的适应性来处理常见图形的谱特征。他们计算了单个计算节点上...

    机器之心
  • 深度聚焦ACL 2017三大研究领域 | 腾讯AI Lab独家解析

    文章转载自微信号腾讯AI实验室(tencent_ailab) 腾讯AI Lab去年四月成立,今年是首次参展ACL,共计三篇文章被录取,位居国内企业前列。此次团队...

    腾讯高校合作
  • 掌握BERT、GPT-3、图神经网络、知识图谱等大厂必备技能!

    金三银四很快就到了,铁子们做好跳槽拿高薪的准备了吗? 回想去年的算法岗,可谓是从灰飞烟灭到人间炼狱。之后的趋势都变成了这样:转行的开始转行,换专业的开始换专业...

    机器学习AI算法工程
  • 入门 | 献给新手的深度学习综述

    论文:Recent Advances in Deep Learning: An Overview

    机器之心
  • 掌握BERT、GPT-3、图神经网络、知识图谱等大厂必备技能!

    金三银四很快就到了,铁子们做好跳槽拿高薪的准备了吗? 回想去年的算法岗,可谓是从灰飞烟灭到人间炼狱。之后的趋势都变成了这样:转行的开始转行,换专业的开始换专业...

    昱良

扫码关注腾讯云开发者

领取腾讯云代金券