<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.aclweb.org/aclwiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Raojinfeng</id>
	<title>ACL Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.aclweb.org/aclwiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Raojinfeng"/>
	<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/Special:Contributions/Raojinfeng"/>
	<updated>2026-04-11T10:23:48Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.6</generator>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Data_sets_for_NLG_blog&amp;diff=12954</id>
		<title>Data sets for NLG blog</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Data_sets_for_NLG_blog&amp;diff=12954"/>
		<updated>2020-08-06T20:09:33Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This blog is a supplement to [[Data sets for NLG]], which lists comments about these data sets from users, authors and other interested parties.   We are especially interested in comments about appropriate and inappropriate usage of a data set, &amp;quot;best practice&amp;quot; use of a data set, useful additional information about a data set (eg, scope, how it was constructed), and pointers to related data sets which may be more appropriate for some users.   Links to relevant papers and other resources are welcome.&lt;br /&gt;
&lt;br /&gt;
We&#039;d love to see more content here, please email Ehud Reiter (e.reiter@abdn.ac.uk) with contributions or other comments&lt;br /&gt;
&lt;br /&gt;
=== E2E ===&lt;br /&gt;
The E2E dataset was used in the [http://www.macs.hw.ac.uk/InteractionLab/E2E/ E2E challenge].&lt;br /&gt;
&lt;br /&gt;
=== SumTime ===&lt;br /&gt;
The SumTime corpus is structured as a database, and presented in text (CSV) and MDB (Microsoft Access) formats.&lt;br /&gt;
&lt;br /&gt;
A good example of the use of Sumtime is [https://doi.org/10.1017/S1351324907004664 Automatic generation of weather forecast texts using comprehensive probabilistic generation-space models].&lt;br /&gt;
&lt;br /&gt;
=== Tuna ===&lt;br /&gt;
[http://www.lrec-conf.org/proceedings/lrec2010/pdf/251_Paper.pdf Dutch] and [https://www.aclweb.org/anthology/W17-3532 Mandarin] versions of Tuna have been developed.&lt;br /&gt;
&lt;br /&gt;
=== WebNLG ===&lt;br /&gt;
Thiago Castro Ferreira and Diego Moussallem spent six months producing an enriched version of WebNLG with high-quality annotations.  This is available on [https://github.com/ThiagoCF05/webnlg GitHub]&lt;br /&gt;
&lt;br /&gt;
The WebNLG dataset was used in the [http://webnlg.loria.fr/pages/results.html WebNLG challenge].&lt;br /&gt;
&lt;br /&gt;
=== Weather ===&lt;br /&gt;
The weather dataset leverages tree-structured meaning representations for better discourse-level structuring, and collects ~30K human annotated utterances. &lt;br /&gt;
&lt;br /&gt;
This is available on [https://github.com/facebookresearch/TreeNLG Github].&lt;br /&gt;
&lt;br /&gt;
=== Weathergov ===&lt;br /&gt;
The Weathergov corpus contains the output of a template-based weather forecast generator, not human-written forecasts ([https://ehudreiter.com/2017/05/09/weathergov/ blog post]). Hence ML on Weathergov is an exercise in reverse engineering a template-based NLG system, not in training an NLG system from human data.  If you want to train on human-written weather forecasts, consider using the [https://github.com/facebookresearch/TreeNLG Weather corpus] and [https://ehudreiter.files.wordpress.com/2016/12/sumtime.zip SumTime corpus] instead.&lt;br /&gt;
&lt;br /&gt;
=== WikiBio ===&lt;br /&gt;
No manual verification or filtering  [https://ehudreiter.com/2019/09/26/generated-texts-must-be-accurate/#comment-15983]&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Data_sets_for_NLG&amp;diff=12953</id>
		<title>Data sets for NLG</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Data_sets_for_NLG&amp;diff=12953"/>
		<updated>2020-08-06T20:00:25Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;!-- MoinMoin name:  DataSets --&amp;gt;&lt;br /&gt;
&amp;lt;!-- Comment:         --&amp;gt;&lt;br /&gt;
&amp;lt;!-- WikiMedia name: DataSets --&amp;gt;&lt;br /&gt;
&amp;lt;!-- Page revision:  00000001 --&amp;gt;&lt;br /&gt;
&amp;lt;!-- Original date:  Fri Nov 11 09:00:35 2005 (1131699635000000) --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This page lists data sets and corpora used for research in natural language generation.  They are available for download over the web. If you know of a dataset which is not listed here, you can email siggen-board@aclweb.org, or just click on Edit in the upper left corner of this page and add the system yourself.&lt;br /&gt;
&lt;br /&gt;
We also have a [[Data sets for NLG blog|blog page]] about data sets, which includes comments about appropriate and inappropriate usage, additional information about data sets, and pointers to related resources.&lt;br /&gt;
&lt;br /&gt;
==Data-to-text/Concept-to-text Generation==&lt;br /&gt;
These datasets contain data and corresponding texts based on this data.&lt;br /&gt;
&lt;br /&gt;
=== boxscore-data (Rotowire) ===&lt;br /&gt;
https://github.com/harvardnlp/boxscore-data/&lt;br /&gt;
&lt;br /&gt;
This dataset consists of (human-written) NBA basketball game summaries aligned with their corresponding box- and line-scores.&lt;br /&gt;
&lt;br /&gt;
=== E2E === &lt;br /&gt;
http://www.macs.hw.ac.uk/InteractionLab/E2E/#data   ([[Data sets for NLG blog#E2E|blog comments]])&lt;br /&gt;
&lt;br /&gt;
Crowdsourced restaurant descriptions with corresponding restaurant data.  English.&lt;br /&gt;
&lt;br /&gt;
=== Methodius Corpus ===&lt;br /&gt;
https://www.inf.ed.ac.uk/research/isdd/admin/package?view=1&amp;amp;id=197&lt;br /&gt;
&lt;br /&gt;
This dataset consists of 5000 short texts describing ancient Greek artefacts, generated by the Methodius NLG system. Each text is linked to its corresponding content plan (including rhetorical relations) and OpenCCG logical form (which describes the syntactic structure).&lt;br /&gt;
&lt;br /&gt;
=== Personage Stylistic Variation for NLG ===&lt;br /&gt;
https://nlds.soe.ucsc.edu/stylistic-variation-nlg&lt;br /&gt;
&lt;br /&gt;
This dataset provides training data for natural language generation of restaurant descriptions in different Big-Five personality styles.&lt;br /&gt;
&lt;br /&gt;
=== Personage Sentence Planning for NLG === &lt;br /&gt;
https://nlds.soe.ucsc.edu/sentence-planning-NLG&lt;br /&gt;
&lt;br /&gt;
This dataset provides training data for natural language generation of restaurant descriptions using sentence planning operations of various kinds.&lt;br /&gt;
&lt;br /&gt;
=== SUMTIME === &lt;br /&gt;
https://ehudreiter.files.wordpress.com/2016/12/sumtime.zip  ([[Data sets for NLG blog#SumTime|blog comments]])&lt;br /&gt;
&lt;br /&gt;
Weather forecasts written by human forecasters, with corresponding forecast data, for UK North Sea marine forecasts.&lt;br /&gt;
&lt;br /&gt;
=== ToTTo ===&lt;br /&gt;
https://github.com/google-research-datasets/ToTTo/&lt;br /&gt;
&lt;br /&gt;
100,000 examples of descriptions of the content of highlighted cells in a Wikipedia table. &lt;br /&gt;
&lt;br /&gt;
=== Weather ===&lt;br /&gt;
https://github.com/facebookresearch/TreeNLG&lt;br /&gt;
&lt;br /&gt;
~30K human annotated utterances for tree-structured weather meaning representations.&lt;br /&gt;
&lt;br /&gt;
=== WeatherGov ===&lt;br /&gt;
https://cs.stanford.edu/~pliang/data/weather-data.zip  ([[Data sets for NLG blog#Weathergov|blog comments]])&lt;br /&gt;
&lt;br /&gt;
Computer-generated weather forecasts from weather.gov (US public forecast), along with corresponding weather data.&lt;br /&gt;
&lt;br /&gt;
=== WebNLG=== &lt;br /&gt;
http://webnlg.loria.fr/pages/data.html   ([[Data sets for NLG blog#WebNLG|blog comments]])&lt;br /&gt;
&lt;br /&gt;
Crowdsourced descriptions of semantic web entities, with corresponding RDF triples.&lt;br /&gt;
&lt;br /&gt;
=== WikiBio (Wikipedia biography dataset) ===&lt;br /&gt;
https://github.com/DavidGrangier/wikipedia-biography-dataset   ([[Data sets for NLG blog#WikiBio|blog comments]])&lt;br /&gt;
&lt;br /&gt;
This dataset gathers 728,321 biographies from Wikipedia. It consists of the first paragraph and the infobox (both tokenized).&lt;br /&gt;
&lt;br /&gt;
=== WikiBio German and French(Wikipedia biography dataset) ===&lt;br /&gt;
https://github.com/PrekshaNema25/StructuredData_To_Descriptions&lt;br /&gt;
&lt;br /&gt;
This dataset consists of the first paragraph and the infobox from German and French Wikipedia biography pages.&lt;br /&gt;
&lt;br /&gt;
=== Wikipedia Person and Animal Dataset ===&lt;br /&gt;
https://eaglew.github.io/dataset/narrating&lt;br /&gt;
&lt;br /&gt;
This dataset gathers 428,748 person and 12,236 animal infobox with descriptions based on Wikipedia dump (2018/04/01) and Wikidata (2018/04/12).&lt;br /&gt;
&lt;br /&gt;
=== The Wikipedia company corpus ===&lt;br /&gt;
https://gricad-gitlab.univ-grenoble-alpes.fr/getalp/wikipediacompanycorpus&lt;br /&gt;
&lt;br /&gt;
Company descriptions collected from Wikipedia. The dataset contains semantic representations, short, and long descriptions for 51K companies in English&lt;br /&gt;
&lt;br /&gt;
== Referring Expressions Generation==&lt;br /&gt;
Referring expression generation is a sub-task of NLG that focuses only on the generation of referring expressions (descriptions) that identify specific entities called targets.&lt;br /&gt;
&lt;br /&gt;
=== GRE3D3 and GRE3D7: Spatial Relations in Referring Expressions ===&lt;br /&gt;
http://jetteviethen.net/research/spatial.html&lt;br /&gt;
&lt;br /&gt;
Two web-based production experiments were conducted by Jette Viethen under the supervision of Robert Dale.&lt;br /&gt;
The resulting corpora GRE3D3 and GRE3D7 contain 720  and 4480 referring expressions, respectively. Each referring expression describes a simple object in a simple 3D scene. GRE3D3 scenes contain 3 objects and GRE3D7 scenes contain 7 objects.&lt;br /&gt;
&lt;br /&gt;
=== RefClef, RefCOCO, RefCOCO+ and RefCOCOg ===&lt;br /&gt;
https://github.com/lichengunc/refer&lt;br /&gt;
&lt;br /&gt;
Referring expressions for objects in images, and the corresponding images.&lt;br /&gt;
&lt;br /&gt;
=== The REAL dataset ===&lt;br /&gt;
https://datastorre.stir.ac.uk/handle/11667/82&lt;br /&gt;
&lt;br /&gt;
Referring expressions for real-wrold objects in images, and the corresponding images.&lt;br /&gt;
&lt;br /&gt;
=== GeoDescriptors ===&lt;br /&gt;
https://gitlab.citius.usc.es/alejandro.ramos/geodescriptors &lt;br /&gt;
&lt;br /&gt;
Geographical descriptions (eg, &amp;quot;Norte de Galicia&amp;quot;) and corresponding regions on a map&lt;br /&gt;
&lt;br /&gt;
=== TUNA Reference Corpus ===&lt;br /&gt;
https://www.abdn.ac.uk/ncs/departments/computing-science/corpus-496.php  ([[Data sets for NLG blog#Tuna|blog comments]])&lt;br /&gt;
&lt;br /&gt;
https://www.abdn.ac.uk/ncs/documents/corpus.zip    [direct download]&lt;br /&gt;
&lt;br /&gt;
The TUNA Reference Corpus is a semantically and pragmatically transparent corpus of identifying references to objects in visual domains. It was constructed via an online experiment and has since been used in a number of evaluation studies on Referring Expressions Generation, as well as in two Shared Tasks: the Attribute Selection for Referring Expressions Generation task (2007), and the Referring Expression Generation task (2008). Main authors: Kees van Deemter, Albert Gatt, Ielka van der Sluis. &lt;br /&gt;
&lt;br /&gt;
=== COCONUT Corpus ===&lt;br /&gt;
http://www.pitt.edu/~coconut/coconut-corpus.html&lt;br /&gt;
&lt;br /&gt;
http://www.pitt.edu/%7Ecoconut/corpora/corpus.tar.gz    [direct download]&lt;br /&gt;
&lt;br /&gt;
COCONUT was a project on “Cooperative, coordinated natural language utterances”. The COCONUT corpus is a collection of computer-mediated dialogues in which two subjects collaborate on a simple task, namely buying furniture. SGML annotations were added according to the [http://www.pitt.edu/%7Epjordan/papers/coconut-manual.pdf COCONUT-DRI coding scheme].&lt;br /&gt;
&lt;br /&gt;
=== Stars2 corpus of referring expressions === &lt;br /&gt;
A collection of 884 annotated definite descriptions produced by 56 subjects in collaborative communication involving speaker-hearer pairs in situations designed so as to challenge existing REG algorithms, with a particular focus on the issue of attribute choice in referential overspeci�fication. &lt;br /&gt;
Link: https://drive.google.com/file/d/0B-KyU7T8S8bLZ1lEQmJRdUc1V28/view?usp=sharing&lt;br /&gt;
Cite: https://link.springer.com/article/10.1007/s10579-016-9350-y &lt;br /&gt;
&lt;br /&gt;
=== b5 corpus of text and referring expressions labelled with personality information ===&lt;br /&gt;
A collection of crowd sourced scene descriptions and an annotated REG corpus, both of which labelled with Big Five personality scores of their authors. Suitable for studies in personality-dependent text generation and referring expression generation.&lt;br /&gt;
Link: https://drive.google.com/open?id=0B-KyU7T8S8bLTHpaMnh2U2NWZzQ&lt;br /&gt;
Cite: https://www.aclweb.org/anthology/L18-1183&lt;br /&gt;
&lt;br /&gt;
==Surface Realisation ==&lt;br /&gt;
&lt;br /&gt;
=== Surface Realization Shared Task 2018 (SR&#039;18) dataset ===&lt;br /&gt;
http://taln.upf.edu/pages/msr2018-ws/SRST.html#data&lt;br /&gt;
&lt;br /&gt;
Description: A multilingual dataset automatically converted from the Universal Dependencies v2.0, comprising unordered syntactic structures (10 languages) and predicate-argument structures (3 languages).&lt;br /&gt;
&lt;br /&gt;
=== Finnish morphology ===&lt;br /&gt;
&lt;br /&gt;
https://www.kaggle.com/mikahama/finnish-locative-cases-for-nouns&lt;br /&gt;
&lt;br /&gt;
Dataset for picking the correct locative case for Finnish nouns (e.g Venäjä&#039;&#039;&#039;llä&#039;&#039;&#039; vs Suome&#039;&#039;&#039;ssa&#039;&#039;&#039;)&lt;br /&gt;
&lt;br /&gt;
https://www.kaggle.com/mikahama/cases-of-complements-of-finnish-verbs&lt;br /&gt;
&lt;br /&gt;
Dataset for picking the right case for objects of verbs in Finnish (e.g. näen talo&#039;&#039;&#039;n&#039;&#039;&#039; vs uneksin talo&#039;&#039;&#039;sta&#039;&#039;&#039;)&lt;br /&gt;
&lt;br /&gt;
== Dialogue ==&lt;br /&gt;
&lt;br /&gt;
=== Alex Context NLG Dataset===&lt;br /&gt;
https://github.com/UFAL-DSG/alex_context_nlg_dataset&lt;br /&gt;
&lt;br /&gt;
A dataset for NLG in dialogue systems in the public transport information domain. It includes preceding context along with each data instance, which should allow NLG systems trained on this data to adapt to user&#039;s way of speaking, which should improve perceived naturalness. Papers: http://workshop.colips.org/re-wochat/documents/02_Paper_6.pdf, https://www.aclweb.org/anthology/W16-3622&lt;br /&gt;
&lt;br /&gt;
=== Cam4NLG === &lt;br /&gt;
https://github.com/shawnwun/RNNLG/tree/master/data&lt;br /&gt;
&lt;br /&gt;
Cam4NLG: Cam4NLG contains 4 NLG datasets for dialogue system development, each of them is in a unique domain. Each data point contains a (dialogue act, ground truth, handcrafted baseline) tuple. &lt;br /&gt;
&lt;br /&gt;
===CLASSiC WOZ corpus on InformationPresentation in Spoken Dialogue Systems===&lt;br /&gt;
http://www.classic-project.org/corpora&lt;br /&gt;
&lt;br /&gt;
CLASSiC is a project on [http://www.classic-project.org/ Computational Learning in Adaptive Systems for Spoken Conversation]. The Wizard-of-Oz corpus on Information Presentation in Spoken Dialogue Systems contains the wizards&#039; choices on Information Presentation strategy (summary, compare, recommend , or a combination of those) and attribute selection. The domain is restaurant search in Edinburgh. Objective measures (such as dialogue length, number of database hits, number of sentences generated etc.), as well as subjective measures (the user scores) were logged.&lt;br /&gt;
&lt;br /&gt;
=== CODA corpus Release 1.0 === &lt;br /&gt;
http://computing.open.ac.uk/coda/resources/code_form.html&lt;br /&gt;
&lt;br /&gt;
This release contains approximately 700 turns of human-authored expository dialogue (by Mark Twain and George Berkeley) which has been aligned with monologue that expresses the same information as the dialogue. The monologue side is annotated with Coherence Relations (RST), and the dialogue side with Dialogue Act tags.&lt;br /&gt;
&lt;br /&gt;
=== Hotel Dialogs for NLG === &lt;br /&gt;
https://nlds.soe.ucsc.edu/hotels&lt;br /&gt;
&lt;br /&gt;
This set of hotel corpora includes a set of paraphrases, room and property descriptions, and full hotel dialogues aimed at exploring different ways of eliciting dialogic, conversational descriptions about hotels.&lt;br /&gt;
&lt;br /&gt;
== Summarisation ==&lt;br /&gt;
&lt;br /&gt;
=== CASS (French) ===&lt;br /&gt;
https://github.com/euranova/CASS-dataset&lt;br /&gt;
&lt;br /&gt;
This dataset is composed of decisions made by the French Court of cassation and summaries of these decisions made by lawyer.&lt;br /&gt;
&lt;br /&gt;
=== TL;DR ===&lt;br /&gt;
https://toolbox.google.com/datasetsearch/search?query=Webis-TLDR-17%20Corpus&amp;amp;docid=kzcwbWD9z3B4Ah3wAAAAAA%3D%3D&lt;br /&gt;
&lt;br /&gt;
Dataset for abstractive summarization constructed using Reddit posts. It is the largest corpus (approximately 3 Million posts) for informal text such as Social Media text,  which can be used to train neural networks for summarization technology.&lt;br /&gt;
&lt;br /&gt;
== Image description ==&lt;br /&gt;
&lt;br /&gt;
===Chinese===&lt;br /&gt;
* Flickr8k-CN: http://lixirong.net/datasets/flickr8kcn&lt;br /&gt;
&lt;br /&gt;
===Dutch===&lt;br /&gt;
&lt;br /&gt;
* DIDEC: http://didec.uvt.nl&lt;br /&gt;
* Flickr30K https://github.com/cltl/DutchDescriptions&lt;br /&gt;
&lt;br /&gt;
===German===&lt;br /&gt;
* Multi30K: http://www.statmt.org/wmt16/multimodal-task.html&lt;br /&gt;
&lt;br /&gt;
== Question Generation ==&lt;br /&gt;
&lt;br /&gt;
=== QGSTEC 2010 Generating Questions from Sentences Corpus ===&lt;br /&gt;
http://computing.open.ac.uk/coda/resources/qg_form.html&lt;br /&gt;
&lt;br /&gt;
A corpus of over 1000 questions (both human and machine generated). The automatically generated questions have been rated by several raters according to five criteria (relevance, question type, syntactic correctness and fluency, ambiguity, and variety).&lt;br /&gt;
&lt;br /&gt;
=== QGSTEC+ ===&lt;br /&gt;
https://github.com/Keith-Godwin/QG-STEC-plus &lt;br /&gt;
&lt;br /&gt;
Improved annotations for the QGSTEC corpus (with higher inter-rater reliability) as described in [http://oro.open.ac.uk/47284/ Godwin and Piwek (2016)].&lt;br /&gt;
&lt;br /&gt;
== Paper Generation ==&lt;br /&gt;
&lt;br /&gt;
=== ACL Title and Abstract Dataset ===&lt;br /&gt;
https://github.com/EagleW/ACL_titles_abstracts_dataset&lt;br /&gt;
&lt;br /&gt;
This dataset gathers 10,874 title and abstract pairs from the ACL Anthology Network (until 2016).&lt;br /&gt;
&lt;br /&gt;
=== PubMed Term, Abstract, Conclusion, Title Dataset ===&lt;br /&gt;
https://eaglew.github.io/dataset/paperrobot_writing&lt;br /&gt;
&lt;br /&gt;
This dataset gathers three types of pairs: Title-to-Abstract (Training: 22,811/Development: 2095/Test: 2095), Abstract-to-Conclusion and Future work (Training: 22,811/Development: 2095/Test: 2095), Conclusion and Future work-to-Title (Training: 15,902/Development: 2095/Test: 2095) from PubMed. Each pair contains a pair of input and output as well as the corresponding terms(from original KB and link prediction results).&lt;br /&gt;
&lt;br /&gt;
==Challenge Data Repository ==&lt;br /&gt;
&lt;br /&gt;
https://sites.google.com/site/genchalrepository/ &lt;br /&gt;
&lt;br /&gt;
== Other ==&lt;br /&gt;
=== PIL: Patient Information Leaflet corpus ===&lt;br /&gt;
http://mcs.open.ac.uk/nlg/old_projects/pills/corpus/PIL/&lt;br /&gt;
&lt;br /&gt;
http://mcs.open.ac.uk/nlg/old_projects/pills/corpus/PIL-corpus-2.0.tar.gz     [direct download]&lt;br /&gt;
&lt;br /&gt;
The Patient Information Leaflet (PIL) corpus] is a [http://www.itri.brighton.ac.uk/projects/pills/corpus/PIL/searchtool/search.html searchable] and [http://mcs.open.ac.uk/nlg/old_projects/pills/corpus/PIL/ browsable] collection of patient information leaflets available in various document formats as well as structurally annotated SGML. The PIL corpus was initially developed as part of the ICONOCLAST project at ITRI, Brighton.&lt;br /&gt;
&lt;br /&gt;
=== Validity of BLEU Evaluation Metric ===&lt;br /&gt;
https://abdn.pure.elsevier.com/en/datasets/data-for-structured-review-of-the-validity-of-bleu&lt;br /&gt;
&lt;br /&gt;
https://abdn.pure.elsevier.com/files/125166547/bleu_survey_data.zip    [direct download]&lt;br /&gt;
&lt;br /&gt;
Correlations between BLEU and human evaluations (for MT as well as NLG), extracted from papers in the ACL Anthology&lt;br /&gt;
&lt;br /&gt;
[[Category:Knowledge Collections and Datasets]]&lt;br /&gt;
{{SIGGEN Wiki}}&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12672</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12672"/>
		<updated>2019-09-12T23:34:37Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - Holographic Dual LSTM Architecture&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.815&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.770&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2019) - Hybrid Co-Attention Network (HCAN)&lt;br /&gt;
| Rao et al. (2019)&lt;br /&gt;
| 0.774&lt;br /&gt;
| 0.843&lt;br /&gt;
|-&lt;br /&gt;
| Tayyar Madabushi (2018) - Question Classification + PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Tayyar Madabushi et al. (2018)&lt;br /&gt;
| 0.836&lt;br /&gt;
| 0.863&lt;br /&gt;
|-&lt;br /&gt;
| Kamath (2019) - Question Classification + RNN + Pre-Attention&lt;br /&gt;
| Kamath et al. (2019)&lt;br /&gt;
| 0.852&lt;br /&gt;
| 0.891&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.784&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2017) - BiMPM&lt;br /&gt;
| Wang et al.  (2017)&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Bian et al.  (2017) - Compare-Aggregate&lt;br /&gt;
| Bian et al.  (2017)&lt;br /&gt;
| 0.821&lt;br /&gt;
| 0.899&lt;br /&gt;
|-&lt;br /&gt;
| Shen et al.  (2017) - IWAN&lt;br /&gt;
| Shen et al.  (2017)&lt;br /&gt;
| 0.822&lt;br /&gt;
| 0.889&lt;br /&gt;
|-&lt;br /&gt;
| Tran et al. (2018) - IWAN + sCARNN&lt;br /&gt;
| Tran et al. (2018)&lt;br /&gt;
| 0.829&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2018) - Multi-Cast Attention Networks (MCAN)&lt;br /&gt;
| Tay et al. (2018)&lt;br /&gt;
| 0.838&lt;br /&gt;
| 0.904&lt;br /&gt;
|-&lt;br /&gt;
| Tayyar Madabushi (2018) - Question Classification + PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Tayyar Madabushi et al. (2018)&lt;br /&gt;
| 0.865&lt;br /&gt;
| 0.904&lt;br /&gt;
|-&lt;br /&gt;
| Yoon et al. (2019) - Compare-Aggregate + LanguageModel + LatentClustering&lt;br /&gt;
| Yoon et al. (2019)&lt;br /&gt;
| 0.868&lt;br /&gt;
| 0.928&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
* Yi Tay, Minh C. Phan, Luu Anh Tuan and Siu Cheung Hui. 2017 [https://arxiv.org/abs/1707.06372 Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture]. In SIGIR 2017.&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui. 2017 [https://arxiv.org/pdf/1707.07847 Enabling Efficient Question Answer Retrieval via Hyperbolic Neural Networks]. In eprint arXiv: 1707.07847.&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
* Zhiguo Wang, Wael Hamza and Radu Florian. 2017.  [https://arxiv.org/pdf/1702.03814.pdf Bilateral Multi-Perspective Matching for Natural Language Sentences]. In eprint arXiv:1702.03814.&lt;br /&gt;
* Weijie Bian, Si Li, Zhao Yang, Guang Chen, Zhiqing Lin. 2017. [https://dl.acm.org/citation.cfm?id=3133089&amp;amp;CFID=791659397&amp;amp;CFTOKEN=43388059 A Compare-Aggregate Model with Dynamic-Clip Attention for Answer Selection]. In CIKM 2017.&lt;br /&gt;
* Gehui Shen, Yunlun Yang, Zhi-Hong Deng. 2017. [https://aclanthology.info/pdf/D/D17/D17-1122.pdf Inter-Weighted Alignment Network for Sentence Pair Modeling.]. In EMNLP 2017.&lt;br /&gt;
* Quan Hung Tran, Tuan Manh Lai, Gholamreza Haffari, Ingrid Zukerman, Trung Bui, Hung Bui, [http://www.aclweb.org/anthology/N18-1115 The Context-dependent Additive Recurrent Neural Net], In NAACL 2018&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui, [https://arxiv.org/abs/1806.00778 Multi-Cast Attention Networks], In KDD 2018&lt;br /&gt;
* Harish Tayyar Madabushi, Mark Lee and John Barnden. [https://aclanthology.coli.uni-saarland.de/papers/C18-1278/c18-1278 Integrating Question Classification and Deep Learning for improved Answer Selection], In COLING 2018&lt;br /&gt;
* Seunghyun Yoon, Franck Dernoncourt, Doo Soon Kim, Trung Bui, Kyomin Jung. 2019. [https://arxiv.org/abs/1905.12897 A Compare-Aggregate Model with Latent Clustering for Answer Selection]. In eprint arXiv:1905.12897.&lt;br /&gt;
* Sanjay Kamath, Brigitte Grau and Yue Ma. 2019. [https://hal.archives-ouvertes.fr/hal-02104488/ Predicting and Integrating Expected Answer Types into a Simple Recurrent Neural Network Model for Answer Sentence Selection]. In CICLING 2019&lt;br /&gt;
* Jinfeng Rao, Linqing Liu, Yi Tay, Wei Yang, Peng Shi, Jimmy Lin, [https://jinfengr.github.io/publications/Rao_etal_EMNLP2019.pdf Bridging the Gap between Relevance Matching and Semantic Matching for Short Text Similarity Modeling], In EMNLP 2019&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12240</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12240"/>
		<updated>2018-06-01T18:27:38Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - Holographic Dual LSTM Architecture&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.815&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.770&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.784&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2017) - BiMPM&lt;br /&gt;
| Wang et al.  (2017)&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Bian et al.  (2017) - Compare-Aggregate&lt;br /&gt;
| Bian et al.  (2017)&lt;br /&gt;
| 0.821&lt;br /&gt;
| 0.899&lt;br /&gt;
|-&lt;br /&gt;
| Shen et al.  (2017) - IWAN&lt;br /&gt;
| Shen et al.  (2017)&lt;br /&gt;
| 0.822&lt;br /&gt;
| 0.889&lt;br /&gt;
|-&lt;br /&gt;
| Tran et al. (2018) - IWAN + sCARNN&lt;br /&gt;
| Tran et al. (2018)&lt;br /&gt;
| 0.829&lt;br /&gt;
| 0.875&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
* Yi Tay, Minh C. Phan, Luu Anh Tuan and Siu Cheung Hui. 2017 [https://arxiv.org/abs/1707.06372 Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture]. In SIGIR 2017.&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui. 2017 [https://arxiv.org/pdf/1707.07847 Enabling Efficient Question Answer Retrieval via Hyperbolic Neural Networks]. In eprint arXiv: 1707.07847.&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
* Zhiguo Wang, Wael Hamza and Radu Florian. 2017.  [https://arxiv.org/pdf/1702.03814.pdf Bilateral Multi-Perspective Matching for Natural Language Sentences]. In eprint arXiv:1702.03814.&lt;br /&gt;
* Weijie Bian, Si Li, Zhao Yang, Guang Chen, Zhiqing Lin. 2017. [https://dl.acm.org/citation.cfm?id=3133089&amp;amp;CFID=791659397&amp;amp;CFTOKEN=43388059 A Compare-Aggregate Model with Dynamic-Clip Attention for Answer Selection]. In CIKM 2017.&lt;br /&gt;
* Gehui Shen, Yunlun Yang, Zhi-Hong Deng. 2017. [https://aclanthology.info/pdf/D/D17/D17-1122.pdf Inter-Weighted Alignment Network for Sentence Pair Modeling.]. In EMNLP 2017.&lt;br /&gt;
* Quan Hung Tran, Tuan Manh Lai, Gholamreza Haffari, Ingrid Zukerman, Trung Bui, Hung Bui, [http://www.aclweb.org/anthology/N18-1115 The Context-dependent Additive Recurrent Neural Net], In NAACL 2018&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12074</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12074"/>
		<updated>2017-11-16T15:52:57Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - Holographic Dual LSTM Architecture&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.815&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.770&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.784&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2017) - BiMPM&lt;br /&gt;
| Wang et al.  (2017)&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Bian et al.  (2017) - Compare-Aggregate&lt;br /&gt;
| Bian et al.  (2017)&lt;br /&gt;
| 0.821&lt;br /&gt;
| 0.899&lt;br /&gt;
|-&lt;br /&gt;
| Shen et al.  (2017) - IWAN&lt;br /&gt;
| Shen et al.  (2017)&lt;br /&gt;
| 0.822&lt;br /&gt;
| 0.889&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
* Yi Tay, Minh C. Phan, Luu Anh Tuan and Siu Cheung Hui. 2017 [https://arxiv.org/abs/1707.06372 Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture]. In SIGIR 2017.&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui. 2017 [https://arxiv.org/pdf/1707.07847 Enabling Efficient Question Answer Retrieval via Hyperbolic Neural Networks]. In eprint arXiv: 1707.07847.&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
* Zhiguo Wang, Wael Hamza and Radu Florian. 2017.  [https://arxiv.org/pdf/1702.03814.pdf Bilateral Multi-Perspective Matching for Natural Language Sentences]. In eprint arXiv:1702.03814.&lt;br /&gt;
* Weijie Bian, Si Li, Zhao Yang, Guang Chen, Zhiqing Lin. 2017. [https://dl.acm.org/citation.cfm?id=3133089&amp;amp;CFID=791659397&amp;amp;CFTOKEN=43388059 A Compare-Aggregate Model with Dynamic-Clip Attention for Answer Selection]. In CIKM 2017.&lt;br /&gt;
* Gehui Shen, Yunlun Yang, Zhi-Hong Deng. 2017. [https://aclanthology.info/pdf/D/D17/D17-1123.pdf Inter-Weighted Alignment Network for Sentence Pair Modeling.]. In EMNLP 2017.&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12073</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12073"/>
		<updated>2017-11-15T16:20:27Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - Holographic Dual LSTM Architecture&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.815&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.770&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.784&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2017) - BiMPM&lt;br /&gt;
| Wang et al.  (2017)&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Bian et al.  (2017) - Compare-Aggregate&lt;br /&gt;
| Bian et al.  (2017)&lt;br /&gt;
| 0.821&lt;br /&gt;
| 0.899&lt;br /&gt;
|-&lt;br /&gt;
| Shen et al.  (2017) - IWAN&lt;br /&gt;
| Shen et al.  (2017)&lt;br /&gt;
| 0.822&lt;br /&gt;
| 0.889&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
* Yi Tay, Minh C. Phan, Luu Anh Tuan and Siu Cheung Hui. 2017 [https://arxiv.org/abs/1707.06372 Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture]. In SIGIR 2017.&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui. 2017 [https://arxiv.org/pdf/1707.07847 Enabling Efficient Question Answer Retrieval via Hyperbolic Neural Networks]. In eprint arXiv: 1707.07847.&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
* Zhiguo Wang, Wael Hamza and Radu Florian. 2017.  [https://arxiv.org/pdf/1702.03814.pdf Bilateral Multi-Perspective Matching for Natural Language Sentences]. In eprint arXiv:1702.03814.&lt;br /&gt;
* Weijie Bian, Si Li, Zhao Yang, Guang Chen, Zhiqing Lin. 2017. [https://aclanthology.info/pdf/D/D17/D17-1123.pdf A Compare-Aggregate Model with Dynamic-Clip Attention for Answer Selection]. In CIKM 2017.&lt;br /&gt;
* Gehui Shen, Yunlun Yang, Zhi-Hong Deng. 2017. [https://aclanthology.info/pdf/D/D17/D17-1123.pdf Inter-Weighted Alignment Network for Sentence Pair Modeling.]. In EMNLP 2017.&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12072</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=12072"/>
		<updated>2017-11-15T16:19:09Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: /* Answer Sentence Selection */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - Holographic Dual LSTM Architecture&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.815&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Tay (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.770&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Tay et al. (2017) - HyperQA (Hyperbolic Embeddings)&lt;br /&gt;
| Tay et al. (2017)&lt;br /&gt;
| 0.784&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2017) - BiMPM&lt;br /&gt;
| Wang et al.  (2017)&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.875&lt;br /&gt;
|-&lt;br /&gt;
| Bian et al.  (2017) - Compare-Aggregate&lt;br /&gt;
| Bian et al.  (2017)&lt;br /&gt;
| 0.821&lt;br /&gt;
| 0.899&lt;br /&gt;
|-&lt;br /&gt;
| Shen et al.  (2017) - IWAN&lt;br /&gt;
| Shen et al.  (2017)&lt;br /&gt;
| 0.822&lt;br /&gt;
| 0.889&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
* Yi Tay, Minh C. Phan, Luu Anh Tuan and Siu Cheung Hui. 2017 [https://arxiv.org/abs/1707.06372 Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture]. In SIGIR 2017.&lt;br /&gt;
* Yi Tay, Luu Anh Tuan, Siu Cheung Hui. 2017 [https://arxiv.org/pdf/1707.07847 Enabling Efficient Question Answer Retrieval via Hyperbolic Neural Networks]. In eprint arXiv: 1707.07847.&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
* Zhiguo Wang, Wael Hamza and Radu Florian. 2017.  [https://arxiv.org/pdf/1702.03814.pdf Bilateral Multi-Perspective Matching for Natural Language Sentences]. In eprint arXiv:1702.03814.&lt;br /&gt;
* Gehui Shen, Yunlun Yang, Zhi-Hong Deng. 2017. [https://aclanthology.info/pdf/D/D17/D17-1123.pdf Inter-Weighted Alignment Network for Sentence Pair Modeling.]. In EMNLP 2017.&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11713</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11713"/>
		<updated>2016-12-06T19:08:27Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. &lt;br /&gt;
* Over time, the original dataset diverged to two versions due to different pre-processing in recent publications: both have the same training set but their development and test sets differ. The Raw version has 82 questions in the development set and 100 questions in the test set; The Clean version (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) removed questions with no answers or with only positive/negative answers, thus has only 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
* Note: MAP/MRR scores on the two versions of TREC QA data (Clean vs Raw) are not comparable according to [https://dl.acm.org/authorize.cfm?key=N27026 Rao et al. (2016)]. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Yang (2016) - Attention-Based Neural Matching Model&lt;br /&gt;
| Yang et al. (2016)&lt;br /&gt;
| 0.750&lt;br /&gt;
| 0.811&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2016) - Pairwise Word Interaction Modelling&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.758&lt;br /&gt;
| 0.822&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.762&lt;br /&gt;
| 0.830&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version of TREC QA&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - L.D.C Model&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015) - Multi-Perspective CNN&lt;br /&gt;
| He and Lin (2015)&lt;br /&gt;
| 0.777&lt;br /&gt;
| 0.836&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - PairwiseRank + Multi-Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In Coling 2016.&lt;br /&gt;
* Hua He, Kevin Gimpel and Jimmy Lin. 2015. [http://aclweb.org/anthology/D/D15/D15-1181.pdf Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks]. In EMNLP 2015.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Liu Yang, Qingyao Ai, Jiafeng Guo, W. Bruce Croft. 2016. [http://maroo.cs.umass.edu/pub/web/getpdf.php?id=1240 aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model]. In CIKM 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [https://dl.acm.org/authorize.cfm?key=N27026 Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016.&lt;br /&gt;
[[Category:State of the art]]&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11669</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11669"/>
		<updated>2016-10-20T17:19:48Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. Over time, this dataset diverged to two versions: both have the same training set but their development and test sets differ due to different pre-processing. The raw version has 82 questions in the development set and 100 questions in the test set. Recently, the raw dataset was cleaned by some researchers (Wang and Ittycheriah et al. 2015, Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) by removing those questions with no answers or with only positive/negative answers, leaving 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Raw Version&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015)&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.755&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - Pairwise + Multiple Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm - Clean Version&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - Lexical Decomposition and Composition&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - Pairwise + Multiple Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In eprint arXiv:1602.07019.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [http://www.cs.umd.edu/~jinfeng/publications/PairwiseNeuralNetwork_CIKM2016.pdf Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016&lt;br /&gt;
[[Category:State of the art]]&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11668</id>
		<title>Question Answering (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Question_Answering_(State_of_the_art)&amp;diff=11668"/>
		<updated>2016-10-20T17:17:23Z</updated>

		<summary type="html">&lt;p&gt;Raojinfeng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Answer Sentence Selection ==&lt;br /&gt;
&lt;br /&gt;
The task of answer sentence selection is designed for the open-domain question answering setting. Given a question and a set of candidate sentences, the task is to choose the correct sentence that contains the exact answer and can sufficiently support the answer choice. &lt;br /&gt;
&lt;br /&gt;
* [http://cs.stanford.edu/people/mengqiu/data/qg-emnlp07-data.tgz QA Answer Sentence Selection Dataset]: labeled sentences using TREC QA track data, provided by [http://cs.stanford.edu/people/mengqiu/ Mengqiu Wang] and first used in [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf Wang et al. (2007)]. Over time, this dataset diverged to two versions: both have the same training set but their development and test sets differ due to different pre-processing. The raw version has 82 questions in the development set and 100 questions in the test set. Recently, the raw dataset was cleaned by some researchers (Tan et al. 2015, dos Santos et al. 2016, Wang et al. 2016) by removing those questions with no answers or with only positive/negative answers, leaving 65 questions in the development set and 68 questions in the test set. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm -- Raw Version&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Punyakanok (2004)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.419&lt;br /&gt;
| 0.494&lt;br /&gt;
|-&lt;br /&gt;
| Cui (2005)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.427&lt;br /&gt;
| 0.526&lt;br /&gt;
|-&lt;br /&gt;
| Wang (2007)&lt;br /&gt;
| Wang et al. (2007)&lt;br /&gt;
| 0.603&lt;br /&gt;
| 0.685&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S (2010)&lt;br /&gt;
| Heilman and Smith (2010)&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.692&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;M (2010)&lt;br /&gt;
| Wang and Manning (2010)&lt;br /&gt;
| 0.595&lt;br /&gt;
| 0.695&lt;br /&gt;
|-&lt;br /&gt;
| Yao (2013)&lt;br /&gt;
| Yao et al. (2013)&lt;br /&gt;
| 0.631&lt;br /&gt;
| 0.748&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2013)&lt;br /&gt;
| Severyn and Moschitti (2013)&lt;br /&gt;
| 0.678&lt;br /&gt;
| 0.736&lt;br /&gt;
|-&lt;br /&gt;
| Shnarch (2013) - Backward &lt;br /&gt;
| Shnarch (2013)&lt;br /&gt;
| 0.686&lt;br /&gt;
| 0.754&lt;br /&gt;
|-&lt;br /&gt;
| Yih (2013) - LCLR&lt;br /&gt;
| Yih et al. (2013)&lt;br /&gt;
| 0.709&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Yu (2014) - TRAIN-ALL bigram+count&lt;br /&gt;
| Yu et al. (2014)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.785&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;N (2015) - Three-Layer BLSTM+BM25&lt;br /&gt;
| Wang and Nyberg (2015)&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.791&lt;br /&gt;
|-&lt;br /&gt;
| Feng (2015) - Architecture-II&lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.711&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| S&amp;amp;M (2015)&lt;br /&gt;
| Severyn and Moschitti (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| W&amp;amp;I (2015)&lt;br /&gt;
| Wang and Ittycheriah (2015)&lt;br /&gt;
| 0.746&lt;br /&gt;
| 0.820&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;L (2015)&lt;br /&gt;
| He and Lin (2016)&lt;br /&gt;
| 0.755&lt;br /&gt;
| 0.825&lt;br /&gt;
|-&lt;br /&gt;
| Rao (2016) - Pairwise + Multiple Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.780&lt;br /&gt;
| 0.834&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm -- Clean Version&lt;br /&gt;
! Reference&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_average_precision MAP]&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Mean_reciprocal_rank MRR]&lt;br /&gt;
|-&lt;br /&gt;
| Tan (2015) - QA-LSTM/CNN+attention &lt;br /&gt;
| Tan et al. (2015)&lt;br /&gt;
| 0.728&lt;br /&gt;
| 0.832&lt;br /&gt;
|-&lt;br /&gt;
| dos Santos (2016) - Attentive Pooling CNN &lt;br /&gt;
| dos Santos et al. (2016)&lt;br /&gt;
| 0.753&lt;br /&gt;
| 0.851&lt;br /&gt;
|-&lt;br /&gt;
| Wang et al.  (2016) - Lexical Decomposition and Composition&lt;br /&gt;
| Wang et al. (2016)&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.845&lt;br /&gt;
|-&lt;br /&gt;
| Rao et al.  (2016) - Pairwise + Multiple Perspective CNN&lt;br /&gt;
| Rao et al. (2016)&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.877&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
* Vasin Punyakanok, Dan Roth, and Wen-Tau Yih. 2004. [http://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi04a.pdf Mapping dependencies trees: An application to question answering]. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Fort Lauderdale, FL, USA.&lt;br /&gt;
* Hang Cui, Renxu Sun, Keya Li, Min-Yen Kan, and Tat-Seng Chua. 2005. [http://ws.csie.ncku.edu.tw/login/upload/2005/paper/Question%20answering%20Question%20answering%20passage%20retrieval%20using%20dependency%20relations.pdf Question answering passage retrieval using dependency relations]. In Proceedings of the 28th ACM-SIGIR International Conference on Research and Development in Information Retrieval, Salvador, Brazil.&lt;br /&gt;
* Wang, Mengqiu and Smith, Noah A. and Mitamura, Teruko. 2007. [http://www.aclweb.org/anthology/D/D07/D07-1003.pdf What is the Jeopardy Model? A Quasi-Synchronous Grammar for QA]. In EMNLP-CoNLL 2007.&lt;br /&gt;
* Heilman, Michael and Smith, Noah A. 2010. [http://www.aclweb.org/anthology/N10-1145 Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions]. In NAACL-HLT 2010.&lt;br /&gt;
* Wang, Mengqiu and Manning, Christopher. 2010. [http://aclweb.org/anthology//C/C10/C10-1131.pdf Probabilistic Tree-Edit Models with Structured Latent Variables for Textual Entailment and Question Answering]. In COLING 2010.&lt;br /&gt;
* E. Shnarch. 2013. Probabilistic Models for Lexical Inference. Ph.D. thesis, Bar Ilan University.&lt;br /&gt;
* Yao, Xuchen and Van Durme, Benjamin and Callison-Burch, Chris and Clark, Peter. 2013. [http://www.aclweb.org/anthology/N13-1106.pdf Answer Extraction as Sequence Tagging with Tree Edit Distance]. In NAACL-HLT 2013.&lt;br /&gt;
* Yih, Wen-tau and Chang, Ming-Wei and Meek, Christopher and Pastusiak, Andrzej. 2013. [http://research.microsoft.com/pubs/192357/QA-SentSel-Updated-PostACL.pdf Question Answering Using Enhanced Lexical Semantic Models]. In ACL 2013.&lt;br /&gt;
* Severyn, Aliaksei and Moschitti, Alessandro. 2013. [http://www.aclweb.org/anthology/D13-1044.pdf Automatic Feature Engineering for Answer Selection and Extraction]. In EMNLP 2013.&lt;br /&gt;
* Lei Yu, Karl Moritz Hermann, Phil Blunsom, and Stephen Pulman. 2014. [http://arxiv.org/pdf/1412.1632v1.pdf Deep Learning for Answer Sentence Selection]. In NIPS deep learning workshop.&lt;br /&gt;
* Di Wang and Eric Nyberg. 2015. [http://www.aclweb.org/anthology/P15-2116 A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering]. In ACL 2015.&lt;br /&gt;
* Minwei Feng, Bing Xiang, Michael R. Glass, Lidan Wang, Bowen Zhou. 2015. [http://arxiv.org/abs/1508.01585 Applying deep learning to answer selection: A study and an open task]. In ASRU 2015.&lt;br /&gt;
* Aliaksei Severyn and Alessandro Moschitti. 2015. [http://disi.unitn.it/~severyn/papers/sigir-2015-long.pdf Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks]. In SIGIR 2015.&lt;br /&gt;
* Zhiguo Wang and Abraham Ittycheriah. 2015. [http://arxiv.org/abs/1507.02628 FAQ-based Question Answering via Word Alignment]. In eprint arXiv:1507.02628.&lt;br /&gt;
* Ming Tan, Cicero dos Santos, Bing Xiang &amp;amp; Bowen Zhou. 2015. [http://arxiv.org/abs/1511.04108 LSTM-Based Deep Learning Models for Nonfactoid Answer Selection]. In eprint arXiv:1511.04108.&lt;br /&gt;
* Cicero dos Santos, Ming Tan, Bing Xiang &amp;amp; Bowen Zhou. 2016. [http://arxiv.org/abs/1602.03609 Attentive Pooling Networks]. In eprint arXiv:1602.03609.&lt;br /&gt;
* Zhiguo Wang, Haitao Mi and Abraham Ittycheriah. 2016. [http://arxiv.org/pdf/1602.07019v1.pdf Sentence Similarity Learning by Lexical Decomposition and Composition]. In eprint arXiv:1602.07019.&lt;br /&gt;
* Hua He and Jimmy Lin. 2016. [https://cs.uwaterloo.ca/~jimmylin/publications/He_etal_NAACL-HTL2016.pdf Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement]. In NAACL 2016.&lt;br /&gt;
* Jinfeng Rao, Hua He and Jimmy Lin. 2016. [http://www.cs.umd.edu/~jinfeng/publications/PairwiseNeuralNetwork_CIKM2016.pdf Noise-Contrastive Estimation for Answer Selection with Deep Neural Networks]. In CIKM 2016&lt;br /&gt;
[[Category:State of the art]]&lt;/div&gt;</summary>
		<author><name>Raojinfeng</name></author>
	</entry>
</feed>