<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.aclweb.org/aclwiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Doboandris</id>
	<title>ACL Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.aclweb.org/aclwiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Doboandris"/>
	<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/Special:Contributions/Doboandris"/>
	<updated>2026-04-21T18:38:49Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.6</generator>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12976</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12976"/>
		<updated>2020-09-06T08:05:27Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC20-hybrid&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC20-corpus&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC20-hybrid&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC20-corpus&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Dobó and Csirik (2020)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &#039;&#039;IEEE 1ICSC 2018&#039;&#039;, pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2020). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &#039;&#039;Journal of Quantitative Linguistics&#039;&#039;, 27(3), pp. 244-271.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019). [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &#039;&#039;AIAI 2019: Artificial Intelligence Applications and Innovations&#039;&#039;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12975</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12975"/>
		<updated>2020-09-06T07:59:36Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó, PhD]&lt;br /&gt;
&lt;br /&gt;
Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics,&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2019	[http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf PhD in Computer Science] (summa cum laude) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2010	[https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
Artificial Intelligence I. tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Formal Languages tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Databases tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
Introduction to Informatics tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
English:	Level C2, Cambridge ESOL (2010)&lt;br /&gt;
&lt;br /&gt;
German:		Level B2, Goethe Institut (2005)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prizes and further information&lt;br /&gt;
&lt;br /&gt;
2013	Best Young Researcher Prize&lt;br /&gt;
	MSZNY 2013 - IX. Magyar Számítógépes Nyelvészeti Konferencia, Szeged&lt;br /&gt;
&lt;br /&gt;
2010	III. prize&lt;br /&gt;
	Regional Scientific Conference of Students (TDK), Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2000	I. place&lt;br /&gt;
	Makkosházi Mathematics Competition (city level)&lt;br /&gt;
&lt;br /&gt;
Erdős number:	3  (András Dobó – János Csirik – Vilmos Totik – Pál Erdős)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Dobó, A., Csirik, J.: [https://doi.org/10.1080/09296174.2019.1570897 A Comprehensive Study of the Parameters in the Creation and Comparison of Feature Vectors in Distributional Semantic Models]. Journal of Quantitative Linguistics. 27(3), 244-271. (2020)&lt;br /&gt;
&lt;br /&gt;
2.	Dobó, A.: [http://journal.sepln.org/index.php/pln/article/viewFile/6205/3656 A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. Procesamiento del Lenguaje Natural. 64, 127-130. (2020)&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A.: [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. Ph.D. thesis, University of Szeged (2019) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the Best Parameter Settings in the Creation and Comparison of Feature Vectors in Distributional Semantic Models Across Multiple Languages]. In: MacIntyre J., Maglogiannis I., Iliadis L., Pimenidis E. (eds) Artificial Intelligence Applications and Innovations. AIAI 2019. IFIP Advances in Information and Communication Technology, vol 559. 487-499. Springer, Cham. (2019)&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A.: [http://siba-ese.unisalento.it/index.php/ejasa/article/download/19296/17273 A measure of adjusted difference between values of a variable]. Electronic Journal of Applied Statistical Analysis. 12(1), 153-175. (2019)&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A.: [http://www.rcs.cic.ipn.mx/rcs/2018_147_6/Multi-D%20Kneser-Ney%20Smoothing%20Preserving%20the%20Original%20Marginal%20Distributions.pdf Multi-D Kneser-Ney Smoothing Preserving the Original Marginal Distributions]. Research in Computing Science. 147(6), 11-25. (2018)&lt;br /&gt;
&lt;br /&gt;
7.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V., Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. 333-341. Springer International Publishing, Switzerland (2014)&lt;br /&gt;
&lt;br /&gt;
8.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V., Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. 359-360. University of Szeged, Szeged (2014)&lt;br /&gt;
&lt;br /&gt;
9.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Computing%20semantic%20similarity%20using%20large%20static%20corpora.pdf Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. 491-502. Springer-Verlag, Berlin Heidelberg (2013)&lt;br /&gt;
&lt;br /&gt;
10.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A., Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 213-224. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
11.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A., Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 35-46. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
12.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/index.php/pln/article/viewFile/842/697 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural. 46, 59-66. (2011)&lt;br /&gt;
&lt;br /&gt;
13.	Dobó, A.: [http://www.inf.u-szeged.hu/sites/default/files/kutatas/konferenciak/tdk2010osz/Dobo_Andras.pdf Angol szavak szinonimáinak automatikus keresése]. TDK. University of Szeged (2010)&lt;br /&gt;
&lt;br /&gt;
14.	Dobó, A.: [https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Interpreting Noun Compounds]. University of Oxford (2010)&lt;br /&gt;
&lt;br /&gt;
15.	Dobó, A.: [http://diploma.bibl.u-szeged.hu/3370/ A Közelítő és szimbolikus számítások tárgy során MATLAB-ban írt zárthelyi dolgozatok automatizált javítása]. University of Szeged (2009)&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12974</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12974"/>
		<updated>2020-09-06T07:51:37Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &#039;&#039;IEEE 1ICSC 2018&#039;&#039;, pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &#039;&#039;Journal of Quantitative Linguistics&#039;&#039;, 27(3), pp. 244-271.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &#039;&#039;AIAI 2019: Artificial Intelligence Applications and Innovations&#039;&#039;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12973</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12973"/>
		<updated>2020-09-06T07:50:14Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &amp;quot;IEEE 1ICSC 2018. pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &amp;quot;Journal of Quantitative Linguistics&amp;quot;, 27(3), pp. 244-271.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &amp;quot;AIAI 2019: Artificial Intelligence Applications and Innovations&amp;quot;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12876</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12876"/>
		<updated>2020-05-16T23:10:35Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó, PhD]&lt;br /&gt;
&lt;br /&gt;
Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics,&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2019	[http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf PhD in Computer Science] (summa cum laude) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2010	[https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
Artificial Intelligence I. tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Formal Languages tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Databases tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
Introduction to Informatics tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
English:	Level C2, Cambridge ESOL (2010)&lt;br /&gt;
&lt;br /&gt;
German:		Level B2, Goethe Institut (2005)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prizes and further information&lt;br /&gt;
&lt;br /&gt;
2013	Best Young Researcher Prize&lt;br /&gt;
	MSZNY 2013 - IX. Magyar Számítógépes Nyelvészeti Konferencia, Szeged&lt;br /&gt;
&lt;br /&gt;
2010	III. prize&lt;br /&gt;
	Regional Scientific Conference of Students (TDK), Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2000	I. place&lt;br /&gt;
	Makkosházi Mathematics Competition (city level)&lt;br /&gt;
&lt;br /&gt;
Erdős number:	3  (András Dobó – János Csirik – Vilmos Totik – Pál Erdős)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Dobó, A.: [http://journal.sepln.org/index.php/pln/article/viewFile/6205/3656 A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. Procesamiento del Lenguaje Natural. 64, 127-130. (2020)&lt;br /&gt;
&lt;br /&gt;
2.	Dobó, A.: [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. Ph.D. thesis, University of Szeged (2019) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the Best Parameter Settings in the Creation and Comparison of Feature Vectors in Distributional Semantic Models Across Multiple Languages]. In: MacIntyre J., Maglogiannis I., Iliadis L., Pimenidis E. (eds) Artificial Intelligence Applications and Innovations. AIAI 2019. IFIP Advances in Information and Communication Technology, vol 559. 487-499. Springer, Cham. (2019)&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: [https://doi.org/10.1080/09296174.2019.1570897 A Comprehensive Study of the Parameters in the Creation and Comparison of Feature Vectors in Distributional Semantic Models]. Journal of Quantitative Linguistics (2019)&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A.: [http://siba-ese.unisalento.it/index.php/ejasa/article/download/19296/17273 A measure of adjusted difference between values of a variable]. Electronic Journal of Applied Statistical Analysis. 12(1), 153-175. (2019)&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A.: [http://www.rcs.cic.ipn.mx/rcs/2018_147_6/Multi-D%20Kneser-Ney%20Smoothing%20Preserving%20the%20Original%20Marginal%20Distributions.pdf Multi-D Kneser-Ney Smoothing Preserving the Original Marginal Distributions]. Research in Computing Science. 147(6), 11-25 (2018)&lt;br /&gt;
&lt;br /&gt;
7.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V., Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. 333-341. Springer International Publishing, Switzerland (2014)&lt;br /&gt;
&lt;br /&gt;
8.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V., Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. 359-360. University of Szeged, Szeged (2014)&lt;br /&gt;
&lt;br /&gt;
9.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Computing%20semantic%20similarity%20using%20large%20static%20corpora.pdf Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. 491-502. Springer-Verlag, Berlin Heidelberg (2013)&lt;br /&gt;
&lt;br /&gt;
10.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A., Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 213-224. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
11.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A., Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 35-46. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
12.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/index.php/pln/article/viewFile/842/697 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural. 46, 59-66. (2011)&lt;br /&gt;
&lt;br /&gt;
13.	Dobó, A.: [http://www.inf.u-szeged.hu/sites/default/files/kutatas/konferenciak/tdk2010osz/Dobo_Andras.pdf Angol szavak szinonimáinak automatikus keresése]. TDK. University of Szeged (2010)&lt;br /&gt;
&lt;br /&gt;
14.	Dobó, A.: [https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Interpreting Noun Compounds]. University of Oxford (2010)&lt;br /&gt;
&lt;br /&gt;
15.	Dobó, A.: [http://diploma.bibl.u-szeged.hu/3370/ A Közelítő és szimbolikus számítások tárgy során MATLAB-ban írt zárthelyi dolgozatok automatizált javítása]. University of Szeged (2009)&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12783</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12783"/>
		<updated>2020-02-04T00:59:18Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: Updated information&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó, PhD]&lt;br /&gt;
&lt;br /&gt;
Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics,&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2019	[http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf PhD in Computer Science] (summa cum laude) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2010	[https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
Artificial Intelligence I. tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Formal Languages tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Databases tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
Introduction to Informatics tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
English:	Level C2, Cambridge ESOL (2010)&lt;br /&gt;
&lt;br /&gt;
German:		Level B2, Goethe Institut (2005)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prizes and further information&lt;br /&gt;
&lt;br /&gt;
2013	Best Young Researcher Prize&lt;br /&gt;
	MSZNY 2013 - IX. Magyar Számítógépes Nyelvészeti Konferencia, Szeged&lt;br /&gt;
&lt;br /&gt;
2010	III. prize&lt;br /&gt;
	Regional Scientific Conference of Students (TDK), Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2000	I. place&lt;br /&gt;
	Makkosházi Mathematics Competition (city level)&lt;br /&gt;
&lt;br /&gt;
Erdős number:	3  (András Dobó – János Csirik – Vilmos Totik – Pál Erdős)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Dobó, A.: A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages. Procesamiento del Lenguaje Natural. 64. (2020) (in press)&lt;br /&gt;
&lt;br /&gt;
2.	Dobó, A.: [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. Ph.D. thesis, University of Szeged (2019) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
3.	Dobó A., Csirik J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the Best Parameter Settings in the Creation and Comparison of Feature Vectors in Distributional Semantic Models Across Multiple Languages]. In: MacIntyre J., Maglogiannis I., Iliadis L., Pimenidis E. (eds) Artificial Intelligence Applications and Innovations. AIAI 2019. IFIP Advances in Information and Communication Technology, vol 559. 487-499. Springer, Cham. (2019)&lt;br /&gt;
&lt;br /&gt;
4.	Dobó A., Csirik J.: [https://doi.org/10.1080/09296174.2019.1570897 A Comprehensive Study of the Parameters in the Creation and Comparison of Feature Vectors in Distributional Semantic Models]. Journal of Quantitative Linguistics (2019)&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A.: [http://siba-ese.unisalento.it/index.php/ejasa/article/download/19296/17273 A measure of adjusted difference between values of a variable]. Electronic Journal of Applied Statistical Analysis. 12(1), 153-175. (2019)&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A.: [http://www.rcs.cic.ipn.mx/rcs/2018_147_6/Multi-D%20Kneser-Ney%20Smoothing%20Preserving%20the%20Original%20Marginal%20Distributions.pdf Multi-D Kneser-Ney Smoothing Preserving the Original Marginal Distributions]. Research in Computing Science. 147(6), 11-25 (2018)&lt;br /&gt;
&lt;br /&gt;
7.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. 333-341. Springer International Publishing, Switzerland (2014)&lt;br /&gt;
&lt;br /&gt;
8.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. 359-360. University of Szeged, Szeged (2014)&lt;br /&gt;
&lt;br /&gt;
9.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Computing%20semantic%20similarity%20using%20large%20static%20corpora.pdf Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. 491-502. Springer-Verlag, Berlin Heidelberg (2013)&lt;br /&gt;
&lt;br /&gt;
10.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 213-224. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
11.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 35-46. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
12.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/index.php/pln/article/viewFile/842/697 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural. 46, 59-66. (2011)&lt;br /&gt;
&lt;br /&gt;
13.	Dobó, A.: [http://www.inf.u-szeged.hu/sites/default/files/kutatas/konferenciak/tdk2010osz/Dobo_Andras.pdf Angol szavak szinonimáinak automatikus keresése]. TDK. University of Szeged (2010)&lt;br /&gt;
&lt;br /&gt;
14.	Dobó, A.: [https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Interpreting Noun Compounds]. University of Oxford (2010)&lt;br /&gt;
&lt;br /&gt;
15.	Dobó, A.: [http://diploma.bibl.u-szeged.hu/3370/ A Közelítő és szimbolikus számítások tárgy során MATLAB-ban írt zárthelyi dolgozatok automatizált javítása]. University of Szeged (2009)&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12679</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12679"/>
		<updated>2019-09-16T00:36:16Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &amp;quot;IEEE 1ICSC 2018. pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &amp;quot;Journal of Quantitative Linguistics&amp;quot;, pp. 1-28.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &amp;quot;AIAI 2019: Artificial Intelligence Applications and Innovations&amp;quot;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=12678</id>
		<title>WordSimilarity-353 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=12678"/>
		<updated>2019-09-16T00:35:50Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
* [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ WordSimilarity-353 Test Collection]&lt;br /&gt;
* contains two sets of English word pairs along with human-assigned similarity judgements&lt;br /&gt;
* first set (set1) contains 153 word pairs along with their similarity scores assigned by 13 subjects&lt;br /&gt;
* second set (set2) contains 200 word pairs with similarity assessed by 16 subjects&lt;br /&gt;
* WordSimilarity-353 dataset is available [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ here]&lt;br /&gt;
* performance is measured by [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rank correlation coefficient]&lt;br /&gt;
* introduced by [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Finkelstein et al. (2002)]&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* [https://www.wikidata.org/wiki/Q31845205 Wikidata] and [https://tools.wmflabs.org/scholia/use/Q31845205 Scholia]&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of increasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! Spearman&#039;s rho&lt;br /&gt;
! Pearson&#039;s r&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.305&lt;br /&gt;
| 0.271&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath 1997&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.318&lt;br /&gt;
| 0.354&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.341&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.357&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.353&lt;br /&gt;
| 0.365&lt;br /&gt;
|-&lt;br /&gt;
| ROGET&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.415&lt;br /&gt;
| 0.536&lt;br /&gt;
|-&lt;br /&gt;
| C&amp;amp;W&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.5&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.48&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.579&lt;br /&gt;
| 0.577&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.492&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.563&lt;br /&gt;
|-&lt;br /&gt;
| simVB+simWN&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.55&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.622&lt;br /&gt;
| 0.629&lt;br /&gt;
|-&lt;br /&gt;
| HSMN+csmRNN&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.65&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.706&lt;br /&gt;
| 0.705&lt;br /&gt;
|-&lt;br /&gt;
| Multi-prototype&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.71&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Multi-lingual SSA&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.674&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.733&lt;br /&gt;
| 0.704&lt;br /&gt;
|-&lt;br /&gt;
| ESA&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.748&lt;br /&gt;
| 0.503&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.795&lt;br /&gt;
| 0.276&lt;br /&gt;
|-&lt;br /&gt;
| TSA&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.80&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| CLEAR&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| ConceptNet Numberbatch&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.828&lt;br /&gt;
| N/A&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in alphabetical order.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Finkelstein, Lev, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. (2002) [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Placing Search in Context: The Concept Revisited]. ACM Transactions on Information Systems, 20(1):116-131.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Halawi, Guy, Gideon Dror, Evgeniy Gabrilovich, and Yehuda Koren. (2012). [http://gabrilovich.com/publications/papers/Halawi2012LSL.pdf Large-scale learning of word relatedness with constraints]. In Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 1406-1414. ACM.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.aaai.org/ocs/index.php/AAAI/AAAI11/paper/download/3616/3972/ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Huang, Eric H., Richard Socher, Christopher D. Manning, and Andrew Y. Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers - Volume 1 (ACL &#039;12), Vol. 1. Association for Computational Linguistics, Stroudsburg, PA, USA, 873-882.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Luong, Minh-Thang, Richard Socher, and Christopher D. Manning. (2013). [http://nlp.stanford.edu/~lmthang/data/papers/conll13_morpho.pdf Better word representations with recursive neural networks for morphology]. CoNLL-2013: 104.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., D. Jurgens and R. Navigli. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Radinsky, Kira, Eugene Agichtein, Evgeniy Gabrilovich, and Shaul Markovitch. (2011). [http://gabrilovich.com/publications/papers/Radinsky2011WTS.pdf A word at a time: computing word relatedness using temporal semantic analysis]. In Proceedings of the 20th international conference on World wide web, pp. 337-346. ACM.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, Rob, Joshua Chin and Catherine Havasi. (2017). [http://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14972 ConceptNet 5.5: An Open Multilingual Graph of General Knowledge]. Proceedings of The 31st AAAI Conference on Artificial Intelligence, San Francisco, CA.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael and Simone Paolo Ponzetto. (2006). [http://www.aaai.org/Papers/AAAI/2006/AAAI06-223.pdf WikiRelate! Computing Semantic Relatedness Using Wikipedia]. Proceedings of The 21st National Conference on Artificial Intelligence (AAAI), Boston, MA.&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=SimLex-999_(State_of_the_art)&amp;diff=12677</id>
		<title>SimLex-999 (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=SimLex-999_(State_of_the_art)&amp;diff=12677"/>
		<updated>2019-09-16T00:34:47Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.cl.cam.ac.uk/~fh295/simlex.html SimLex-999] aims at a cleaner benchmark of similarity (but not relatedness). Pairs of words were chosen to represent different ranges of similarity and with either high or low association. Subjects were instructed to differentiate between similarity and relatedness and rate regarding the former only.&lt;br /&gt;
&lt;br /&gt;
See also: [[Similarity (State of the art)]], [[Similar-Associated-Both Test Collection (State of the art)]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm !! Reference for algorithm !! Reference for reported results  !! Type !! Spearman&#039;s rho !! Pearson&#039;s r !! Notes&lt;br /&gt;
|-&lt;br /&gt;
| Re16&lt;br /&gt;
| Recski et al. (2016)&amp;lt;ref name=recski16&amp;gt;Recski, G., Iklódi, E., Pajkossy, K., &amp;amp; Kornai, A. (2016). [https://www.aclweb.org/anthology/W16-1622 Measuring semantic similarity of words using concept networks]. In: Proceedings of the 1st Workshop on Representation Learning for NLP, pp. 193-200.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Recski et al. (2016)&amp;lt;ref name=recski16/&amp;gt;&lt;br /&gt;
| Hybrid || 0.76 || -&lt;br /&gt;
|-&lt;br /&gt;
| SVR4&lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| Combined || 0.642 || 0.658&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19&amp;gt;Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Hybrid || 0.621 || 0.481&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&amp;lt;ref name=speer17&amp;gt;Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. AAAI-17, pp. 4444-4451.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Hybrid || 0.616 || 0.634&lt;br /&gt;
|-&lt;br /&gt;
| joint(SP+,skip-gram)&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus&amp;gt;Schwartz, R., Reichart, Roi, Rappoport, A. (2015). Symmetric Pattern Based Word Embeddings for Improved Word Similarity Prediction, CoNLL 2015.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus/&amp;gt;&lt;br /&gt;
| Distributional || 0.56 || - || Trained on word2vec corpus, best results for pure distributional model.&lt;br /&gt;
|-&lt;br /&gt;
| UMBC&lt;br /&gt;
| Han et al. (2013)&amp;lt;ref&amp;gt;Han, L., Kashyap, A., Finin, T., Mayfield, J., Weese, J.: UMBC EBIQUITY-CORE: Semantic textual similarity systems. In: Proceedings of the Second Joint Conference on Lexical and Computational Semantics, vol. 1, pp. 44–52 (2013)&amp;lt;/ref&amp;gt; &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| || 0.558 || 0.557 || without using POS information&lt;br /&gt;
|-&lt;br /&gt;
| SP+&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus&amp;gt;Schwartz, R., Reichart, Roi, Rappoport, A. (2015). Symmetric Pattern Based Word Embeddings for Improved Word Similarity Prediction, CoNLL 2015.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus/&amp;gt;&lt;br /&gt;
| Distributional || 0.52 || -&lt;br /&gt;
|-&lt;br /&gt;
| RNNenc&lt;br /&gt;
| Hill et al. (2014b)&amp;lt;ref name=rnnenc&amp;gt;Hill, F., Cho, K., Jean, S., Devin, C., &amp;amp; Bengio, Y. (2014b). Not All Neural Embeddings are Born Equal, 1–5.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014b)&amp;lt;ref name=rnnenc/&amp;gt;&lt;br /&gt;
| Distributional, multilingual || 0.52 || -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&amp;lt;ref name=salle18&amp;gt;Salle A., Idiart M., and Villavicencio A. (2018). [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&amp;lt;/ref&amp;gt; &lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.417 || 0.426&lt;br /&gt;
|-&lt;br /&gt;
| Word2vec&lt;br /&gt;
| Mikolov et al. (2013)&amp;lt;ref&amp;gt;Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex&amp;gt;Hill, F., Reichart, R., &amp;amp; Korhonen, A. (2014a). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Distributional || 0.414 || - || Trained on Wikipedia&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&amp;lt;ref name=pennington14&amp;gt;Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. EMNLP 2014, pp. 1532-1543.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.406 || 0.433&lt;br /&gt;
|-&lt;br /&gt;
| Lesk&lt;br /&gt;
| &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea&amp;gt;Banjade, R., Maharjan, N., Niraula, N., Rus, V., &amp;amp; Gautam, D. (2015). Lemon and Tea Are Not Similar: Measuring Word-to-Word Similarity by Combining Different Methods. Computational Linguistics and Intelligent Text Processing, 9041, 335–346. doi:10.1007/978-3-319-18111-0_25&amp;lt;/ref&amp;gt;&lt;br /&gt;
| || 0.404 || 0.347&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.393 || 0.401&lt;br /&gt;
|-&lt;br /&gt;
| ESA&lt;br /&gt;
| &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| || 0.271 || 0.145&lt;br /&gt;
|-&lt;br /&gt;
| Neural language model&lt;br /&gt;
| Collobert &amp;amp; Weston (2008)&amp;lt;ref&amp;gt;R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex/&amp;gt; &lt;br /&gt;
| Distributional || 0.268 || - || Trained on Wikipedia&lt;br /&gt;
|-&lt;br /&gt;
| Neural language model with global context&lt;br /&gt;
| Huang et al. (2012)&amp;lt;ref&amp;gt;Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex/&amp;gt; &lt;br /&gt;
| Distributional || 0.098 || - || Trained on Wikipedia&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12676</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12676"/>
		<updated>2019-09-16T00:34:01Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset [Resnik, 1995]&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset [Miller and Charles, 1991], which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.893 [0.780, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.892 [0.778, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.853 [0.704, 0.930]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.832 [0.666, 0.919]&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.822 [0.648, 0.914]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12675</id>
		<title>RG-65 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12675"/>
		<updated>2019-09-16T00:33:40Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Rubenstein &amp;amp; Goodenough (RG-65) dataset&lt;br /&gt;
* 65 word pairs; &lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 51 subjects [Rubenstein and Goodenough, 1965].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Knowledge-based (Wiktionary)&lt;br /&gt;
| 0.920&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.901&lt;br /&gt;
| 0.896&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.899&lt;br /&gt;
| 0.914&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.890&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| NASARI&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.880&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Knowledge-based (WordNet)&lt;br /&gt;
| 0.868&lt;br /&gt;
| 0.810&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.838&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.833&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.830&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.732&lt;br /&gt;
|-&lt;br /&gt;
| Roget&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.818&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.731&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.787&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.797&lt;br /&gt;
| 0.852&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.788&lt;br /&gt;
| 0.834&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.769&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.763&lt;br /&gt;
| 0.792&lt;br /&gt;
|-&lt;br /&gt;
| ESA*&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based &lt;br /&gt;
| 0.749&lt;br /&gt;
| 0.716&lt;br /&gt;
|-&lt;br /&gt;
| SOCPMI*&lt;br /&gt;
| Islam and Inkpen (2006)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.741&lt;br /&gt;
| 0.729&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.732&lt;br /&gt;
| 0.737&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.731&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.640&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| LSA*&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.644&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| -&lt;br /&gt;
| 0.530&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Note: values reported by (Hassan and Mihalcea, 2011) are &amp;quot;based on the collected raw data from the respective authors&amp;quot;, and those highlighted by (*) are re-implementations.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, Eneko, Enrique Alfonseca, Keith Hall, Jana Kravalova, Marius Pasca, Aitor Soroa: [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. HLT-NAACL 2009: 19-27&lt;br /&gt;
&lt;br /&gt;
Camacho-Collados, José, Pilehvar, Mohammad Taher, and Navigli, Roberto: [http://aclweb.org/anthology/N/N15/N15-1059.pdf NASARI: a Novel Approach to a Semantically­-Aware Representation of Items]. NAACL 2015, pp. 567-577, Denver, USA.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.cse.unt.edu/~rada/papers/hassan.aaai11.pdf‎ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Hughes, Thad, Daniel Ramage, Lexical Semantic Relatedness with Random Graph Walks. EMNLP-CoNLL 2007: 581-589.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Milne, David, and Ian H. Witten, An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links, In Proceedings of AAAI 2008.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens, D. and Navigli, R. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T. and Navigli, R. [http://www.sciencedirect.com/science/article/pii/S000437021500106X From Senses to Texts: An All-in-one Graph-based Approach for Measuring Semantic Similarity]. Artificial Intelligence, Elsevier.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Rubenstein, Herbert,  and John B. Goodenough. Contextual correlates of synonymy. Communications of the ACM, 8(10):627–633, 1965.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael, Simone Paolo Ponzetto: WikiRelate! Computing Semantic Relatedness Using Wikipedia. AAAI 2006: 1419-1424&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12674</id>
		<title>TOEFL Synonym Questions (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12674"/>
		<updated>2019-09-16T00:32:59Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* &#039;&#039;&#039;the TOEFL questions are available on request by contacting [http://lsa.colorado.edu/mail_sub.html LSA Support at CU Boulder]&#039;&#039;&#039;, the people who manage the [http://lsa.colorado.edu/ LSA web site at Colorado]&lt;br /&gt;
* TOEFL = Test of English as a Foreign Language&lt;br /&gt;
* 80 multiple-choice synonym questions; 4 choices per question&lt;br /&gt;
* introduced in Landauer and Dumais (1997) as a way of evaluating algorithms for measuring degree of similarity between words&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sample question ==&lt;br /&gt;
&lt;br /&gt;
::{| border=&amp;quot;0&amp;quot; cellpadding=&amp;quot;1&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Stem:&lt;br /&gt;
|&lt;br /&gt;
| levied&lt;br /&gt;
|-&lt;br /&gt;
! Choices:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (b)&lt;br /&gt;
| believed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (c)&lt;br /&gt;
| requested&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (d)&lt;br /&gt;
| correlated&lt;br /&gt;
|-&lt;br /&gt;
! Solution:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for experiment&lt;br /&gt;
! Type&lt;br /&gt;
! Correct&lt;br /&gt;
! 95% confidence&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 20.31%&lt;br /&gt;
| 12.89–31.83%&lt;br /&gt;
|-&lt;br /&gt;
| LC&lt;br /&gt;
| Leacock and Chodrow (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 21.88%&lt;br /&gt;
| 13.91–33.21%&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 24.06%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| Random&lt;br /&gt;
| Random guessing&lt;br /&gt;
| 1 / 4 = 25.00%&lt;br /&gt;
| Random&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| JC&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 64.38%&lt;br /&gt;
| 52.90–74.80%&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Average non-English US college applicant&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Human&lt;br /&gt;
| 64.50%&lt;br /&gt;
| 53.01–74.88%&lt;br /&gt;
|-&lt;br /&gt;
| RI&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 72.50%&lt;br /&gt;
| 61.38-81.90%&lt;br /&gt;
|-&lt;br /&gt;
| DS&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.00%&lt;br /&gt;
| 62.72-82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.75%&lt;br /&gt;
| 62.72–82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PairClass&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 76.25%&lt;br /&gt;
| 65.42-85.06%&lt;br /&gt;
|-&lt;br /&gt;
| HSO&lt;br /&gt;
| Hirst and St.-Onge (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 77.91%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 78.75%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 80.00%&lt;br /&gt;
| 69.56–88.11%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| LC-IR&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| CWO&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 82.55%&lt;br /&gt;
| 72.38–90.09%&lt;br /&gt;
|-&lt;br /&gt;
| PPMIC&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 85.00%&lt;br /&gt;
| 75.26-92.00%&lt;br /&gt;
|-&lt;br /&gt;
| GLSA&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 86.25%&lt;br /&gt;
| 76.73-92.93%&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 87.50%&lt;br /&gt;
| 78.21-93.84%&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 88.75%&lt;br /&gt;
| 79.72-94.72%&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 90.00%&lt;br /&gt;
| 81.24-95.58%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 92.50%&lt;br /&gt;
| 84.39-97.20%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 95.0%&lt;br /&gt;
| 87.69-98.62%&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| WordNet graph-based (unsupervised)&lt;br /&gt;
| 96.25%&lt;br /&gt;
| 89.43-99.22%&lt;br /&gt;
|-&lt;br /&gt;
| PR&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 97.50%&lt;br /&gt;
| 91.26–99.70%&lt;br /&gt;
|-&lt;br /&gt;
| Sp19&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| PCCP&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 100.00%&lt;br /&gt;
| 96.32-100.00%&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Explanation of table ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Algorithm&#039;&#039;&#039; = name of algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for algorithm&#039;&#039;&#039; = where to find out more about given algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for experiment&#039;&#039;&#039; = where to find out more about evaluation of given algorithm with TOEFL questions&lt;br /&gt;
* &#039;&#039;&#039;Type&#039;&#039;&#039; = general type of algorithm: corpus-based, lexicon-based, hybrid&lt;br /&gt;
* &#039;&#039;&#039;Correct&#039;&#039;&#039; = percent of 80 questions that given algorithm answered correctly&lt;br /&gt;
* &#039;&#039;&#039;95% confidence&#039;&#039;&#039; = confidence interval calculated using the [[Statistical calculators|Binomial Exact Test]]&lt;br /&gt;
* table rows sorted in order of increasing percent correct&lt;br /&gt;
* several WordNet-based similarity measures are implemented in [http://www.d.umn.edu/~tpederse/ Ted Pedersen]&#039;s [http://www.d.umn.edu/~tpederse/similarity.html WordNet::Similarity] package&lt;br /&gt;
* LSA = Latent Semantic Analysis&lt;br /&gt;
* PCCP = Principal Component vectors with Caron P&lt;br /&gt;
* PMI-IR = Pointwise Mutual Information - Information Retrieval&lt;br /&gt;
* PR = Product Rule&lt;br /&gt;
* PPMIC = Positive Pointwise Mutual Information with Cosine&lt;br /&gt;
* GLSA = Generalized Latent Semantic Analysis&lt;br /&gt;
* CWO = Context Window Overlapping&lt;br /&gt;
* DS = Dependency Space&lt;br /&gt;
* RI = Random Indexing&lt;br /&gt;
&lt;br /&gt;
== Notes ==&lt;br /&gt;
&lt;br /&gt;
* the performance of a corpus-based algorithm depends on the corpus, so the difference in performance between two corpus-based systems may be due to the different corpora, rather than the different algorithms&lt;br /&gt;
* the TOEFL questions include nouns, verbs, and adjectives, but some of the WordNet-based algorithms were only designed to work with nouns; this explains some of the lower scores&lt;br /&gt;
* some of the algorithms may have been tuned on the TOEFL questions; read the references for details&lt;br /&gt;
* Landauer and Dumais (1997) report scores that were corrected for guessing by subtracting a penalty of 1/3 for each incorrect answer; they report a score of 52.5% when this penalty is applied; when the penalty is removed, their performance is 64.4% correct&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2007). [http://www.cs.bham.ac.uk/~jxb/PUBS/BRM.pdf Extracting semantic representations from word co-occurrence statistics: A computational study]. &#039;&#039;Behavior Research Methods&#039;&#039;, 39(3), 510-526.&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2012). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.228.9582&amp;amp;rep=rep1&amp;amp;type=pdf Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and SVD]. &#039;&#039;Behavior Research Methods&#039;&#039;,  44(3):890-907.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged. [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502&lt;br /&gt;
&lt;br /&gt;
Lushan Han. (2014). [http://ebiquity.umbc.edu/paper/html/id/658/Schema-Free-Querying-of-Semantic-Data Schema Free Querying of Semantic Data], Ph.D. dissertation, University of Maryland, Baltimore County, Baltimore, MD USA.&lt;br /&gt;
&lt;br /&gt;
Higgins, D. (2005). [http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.329.1517 Which Statistics Reflect Semantics? Rethinking Synonymy and Word Similarity.] In: Kepser, S., Reis, M. (eds.) &#039;&#039;Linguistic Evidence: Empirical, Theoretical and Computational Perspectives&#039;&#039;. Mouton de Gruyter, Berlin, pp. 265–284.&lt;br /&gt;
&lt;br /&gt;
Hirst, G., and St-Onge, D. (1998). [http://mirror.eacoss.org/documentation/ITLibrary/IRIS/Data/1997/Hirst/Lexical/1997-Hirst-Lexical.pdf Lexical chains as representation of context for the detection and correction of malapropisms]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, 305-332.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Jiang, J.J., and Conrath, D.W. (1997). [http://wortschatz.uni-leipzig.de/~sbordag/aalw05/Referate/03_Assoziationen_BudanitskyResnik/Jiang_Conrath_97.pdf Semantic similarity based on corpus statistics and lexical taxonomy]. &#039;&#039;Proceedings of the International Conference on Research in Computational Linguistics&#039;&#039;, Taiwan.&lt;br /&gt;
&lt;br /&gt;
Karlgren, J. and Sahlgren, M. (2001). [http://www.sics.se/~jussi/Artiklar/2001_RWIbook/KarlgrenSahlgren2001.pdf From Words to Understanding]. In Uesaka, Y., Kanerva, P., &amp;amp; Asoh, H. (Eds.), &#039;&#039;Foundations of Real-World Intelligence&#039;&#039;, Stanford: CSLI Publications, pp. 294–308. &lt;br /&gt;
&lt;br /&gt;
Landauer, T.K., and Dumais, S.T. (1997). [http://lsa.colorado.edu/papers/plato/plato.annote.html A solution to Plato&#039;s problem: The latent semantic analysis theory of the acquisition, induction, and representation of knowledge]. &#039;&#039;Psychological Review&#039;&#039;, 104(2):211–240.&lt;br /&gt;
&lt;br /&gt;
Leacock, C., and Chodorow, M. (1998). [http://books.google.ca/books?id=Rehu8OOzMIMC&amp;amp;lpg=PA265&amp;amp;ots=IpnaLkZUec&amp;amp;lr&amp;amp;pg=PA265#v=onepage&amp;amp;q&amp;amp;f=false Combining local context and WordNet similarity for word sense identification]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, pp. 265-283.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://www.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. &#039;&#039;Proceedings of the 15th International Conference on Machine Learning (ICML-98)&#039;&#039;, Madison, WI, pp. 296-304.&lt;br /&gt;
&lt;br /&gt;
Matveeva, I., Levow, G., Farahat, A., and Royer, C. (2005). [http://people.cs.uchicago.edu/~matveeva/SynGLSA_ranlp_final.pdf Generalized latent semantic analysis for term representation]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-05)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Pado, S., and Lapata, M. (2007). [http://www.nlpado.de/~sebastian/pub/papers/cl07_pado.pdf Dependency-based construction of semantic space models]. &#039;&#039;Computational Linguistics&#039;&#039;, 33(2), 161-199.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens D., and Navigli R. (2013). [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, disambiguate and walk: A unified approach for measuring semantic similarity]. &#039;&#039;Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013),&#039;&#039; Sofia, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Rapp, R. (2003). [http://www.amtaweb.org/summit/MTSummit/FinalPapers/19-Rapp-final.pdf Word sense discovery based on sense descriptor dissimilarity]. &#039;&#039;Proceedings of the Ninth Machine Translation Summit&#039;&#039;, pp. 315-322.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://citeseer.ist.psu.edu/resnik95using.html Using information content to evaluate semantic similarity]. &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI-95)&#039;&#039;, Montreal, pp. 448-453.&lt;br /&gt;
&lt;br /&gt;
Ruiz-Casado, M., Alfonseca, E. and Castells, P. (2005) [http://alfonseca.org/pubs/2005-ranlp1.pdf Using context-window overlapping in Synonym Discovery and Ontology Extension]. &#039;&#039;Proceedings of the International Conference Recent Advances in Natural Language Processing (RANLP-2005)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Terra, E., and Clarke, C.L.A. (2003). [http://acl.ldc.upenn.edu/N/N03/N03-1032.pdf Frequency estimates for statistical word similarity measures]. &#039;&#039;Proceedings of the Human Language Technology and North American Chapter of Association of Computational Linguistics Conference 2003 (HLT/NAACL 2003)&#039;&#039;, pp. 244–251.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2001). [http://arxiv.org/abs/cs.LG/0212033 Mining the Web for synonyms: PMI-IR versus LSA on TOEFL]. &#039;&#039;Proceedings of the Twelfth European Conference on Machine Learning (ECML-2001)&#039;&#039;, Freiburg, Germany, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D., Littman, M.L., Bigham, J., and Shnayder, V. (2003). [http://arxiv.org/abs/cs.CL/0309035 Combining independent modules to solve multiple-choice synonym and analogy problems]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, pp. 482-489.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2008). [http://arxiv.org/abs/0809.0124 A uniform approach to analogies, synonyms, antonyms, and associations]. &#039;&#039;Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)&#039;&#039;, Manchester, UK, pp. 905-912.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12673</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12673"/>
		<updated>2019-09-16T00:31:02Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó]&lt;br /&gt;
&lt;br /&gt;
PhD Student&lt;br /&gt;
&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
Artificial Intelligence I. tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Formal Languages tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Databases tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
Introduction to Informatics tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	[http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf PhD in Computer Science] - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2009-2010	[https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
English:	Level C2, Cambridge ESOL (2010)&lt;br /&gt;
&lt;br /&gt;
German:		Level B2, Goethe Institut (2005)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prizes and further information&lt;br /&gt;
&lt;br /&gt;
2013	Best Young Researcher Prize&lt;br /&gt;
	MSZNY 2013 - IX. Magyar Számítógépes Nyelvészeti Konferencia, Szeged&lt;br /&gt;
&lt;br /&gt;
2010	III. prize&lt;br /&gt;
	Regional Scientific Conference of Students (TDK), Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2000	I. place&lt;br /&gt;
	Makkosházi Mathematics Competition (city level)&lt;br /&gt;
&lt;br /&gt;
Erdős number:	3  (András Dobó – János Csirik – Vilmos Totik – Pál Erdős)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Dobó, A.: [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged (2019) - [https://github.com/doboandras/dsm-parameter-analysis GitHub repository]&lt;br /&gt;
&lt;br /&gt;
2.	Dobó A., Csirik J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the Best Parameter Settings in the Creation and Comparison of Feature Vectors in Distributional Semantic Models Across Multiple Languages]. In: MacIntyre J., Maglogiannis I., Iliadis L., Pimenidis E. (eds) Artificial Intelligence Applications and Innovations. AIAI 2019. IFIP Advances in Information and Communication Technology, vol 559. 487-499. Springer, Cham. (2019)&lt;br /&gt;
&lt;br /&gt;
3.	Dobó A., Csirik J.: [https://doi.org/10.1080/09296174.2019.1570897 A Comprehensive Study of the Parameters in the Creation and Comparison of Feature Vectors in Distributional Semantic Models]. Journal of Quantitative Linguistics (2019)&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A.: [http://siba-ese.unisalento.it/index.php/ejasa/article/download/19296/17273 A measure of adjusted difference between values of a variable]. Electronic Journal of Applied Statistical Analysis. 12(1), 153-175. (2019)&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A.: [http://www.rcs.cic.ipn.mx/rcs/2018_147_6/Multi-D%20Kneser-Ney%20Smoothing%20Preserving%20the%20Original%20Marginal%20Distributions.pdf Multi-D Kneser-Ney Smoothing Preserving the Original Marginal Distributions]. Research in Computing Science. 147(6), 11-25 (2018)&lt;br /&gt;
&lt;br /&gt;
6.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. 333-341. Springer International Publishing, Switzerland (2014)&lt;br /&gt;
&lt;br /&gt;
7.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. 359-360. University of Szeged, Szeged (2014)&lt;br /&gt;
&lt;br /&gt;
8.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Computing%20semantic%20similarity%20using%20large%20static%20corpora.pdf Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. 491-502. Springer-Verlag, Berlin Heidelberg (2013)&lt;br /&gt;
&lt;br /&gt;
9.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 213-224. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
10.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 35-46. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
11.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/index.php/pln/article/viewFile/842/697 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural. 46, 59-66. (2011)&lt;br /&gt;
&lt;br /&gt;
12.	Dobó, A.: [http://www.inf.u-szeged.hu/sites/default/files/kutatas/konferenciak/tdk2010osz/Dobo_Andras.pdf Angol szavak szinonimáinak automatikus keresése]. TDK. University of Szeged (2010)&lt;br /&gt;
&lt;br /&gt;
13.	Dobó, A.: [https://ora.ox.ac.uk/objects/uuid:1b771160-3f13-4362-a69a-c73401bec321/download_file?file_format=pdf&amp;amp;safe_filename=Dissertation.pdf&amp;amp;type_of_work=Thesis Interpreting Noun Compounds]. University of Oxford (2010)&lt;br /&gt;
&lt;br /&gt;
14.	Dobó, A.: [http://diploma.bibl.u-szeged.hu/3370/ A Közelítő és szimbolikus számítások tárgy során MATLAB-ban írt zárthelyi dolgozatok automatizált javítása]. University of Szeged (2009)&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12610</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=12610"/>
		<updated>2019-08-17T00:47:20Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó]&lt;br /&gt;
&lt;br /&gt;
PhD Student&lt;br /&gt;
&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
Artificial Intelligence I. tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Formal Languages tutorials (3 semesters)&lt;br /&gt;
&lt;br /&gt;
Databases tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
Introduction to Informatics tutorials (1 semester)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	[http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf PhD in Computer Science]&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2009-2010	[http://solo.bodleian.ox.ac.uk/primo_library/libweb/action/dlDisplay.do?vid=OXVU1&amp;amp;docId=oxfaleph017405947&amp;amp;fn=permalink Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
English:	Level C2, Cambridge ESOL (2010)&lt;br /&gt;
&lt;br /&gt;
German:		Level B2, Goethe Institut (2005)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prizes and further information&lt;br /&gt;
&lt;br /&gt;
2013	Best Young Researcher Prize&lt;br /&gt;
	MSZNY 2013 - IX. Magyar Számítógépes Nyelvészeti Konferencia, Szeged&lt;br /&gt;
&lt;br /&gt;
2010	III. prize&lt;br /&gt;
	Regional Scientific Conference of Students (TDK), Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
2000	I. place&lt;br /&gt;
	Makkosházi Mathematics Competition (city level)&lt;br /&gt;
&lt;br /&gt;
Erdős number:	3  (András Dobó – János Csirik – Vilmos Totik – Pál Erdős)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Dobó, A.: [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged (2019)&lt;br /&gt;
&lt;br /&gt;
2.	Dobó A., Csirik J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the Best Parameter Settings in the Creation and Comparison of Feature Vectors in Distributional Semantic Models Across Multiple Languages]. In: MacIntyre J., Maglogiannis I., Iliadis L., Pimenidis E. (eds) Artificial Intelligence Applications and Innovations. AIAI 2019. IFIP Advances in Information and Communication Technology, vol 559. 487-499. Springer, Cham. (2019)&lt;br /&gt;
&lt;br /&gt;
3.	Dobó A., Csirik J.: [https://doi.org/10.1080/09296174.2019.1570897 A Comprehensive Study of the Parameters in the Creation and Comparison of Feature Vectors in Distributional Semantic Models]. Journal of Quantitative Linguistics (2019)&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A.: [http://siba-ese.unisalento.it/index.php/ejasa/article/download/19296/17273 A measure of adjusted difference between values of a variable]. Electronic Journal of Applied Statistical Analysis. 12(1), 153-175. (2019)&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A.: [http://www.rcs.cic.ipn.mx/rcs/2018_147_6/Multi-D%20Kneser-Ney%20Smoothing%20Preserving%20the%20Original%20Marginal%20Distributions.pdf Multi-D Kneser-Ney Smoothing Preserving the Original Marginal Distributions]. Research in Computing Science. 147(6), 11-25 (2018)&lt;br /&gt;
&lt;br /&gt;
6.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. 333-341. Springer International Publishing, Switzerland (2014)&lt;br /&gt;
&lt;br /&gt;
7.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. 359-360. University of Szeged, Szeged (2014)&lt;br /&gt;
&lt;br /&gt;
8.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/~dobo/Publications/Computing%20semantic%20similarity%20using%20large%20static%20corpora.pdf Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. 491-502. Springer-Verlag, Berlin Heidelberg (2013)&lt;br /&gt;
&lt;br /&gt;
9.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 213-224. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
10.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. 35-46. University of Szeged, Szeged (2012)&lt;br /&gt;
&lt;br /&gt;
11.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/download/842/697 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural. 46, 59-66. (2011)&lt;br /&gt;
&lt;br /&gt;
12.	Dobó, A.: [http://www.inf.u-szeged.hu/sites/default/files/kutatas/konferenciak/tdk2010osz/Dobo_Andras.pdf Angol szavak szinonimáinak automatikus keresése]. TDK. University of Szeged (2010)&lt;br /&gt;
&lt;br /&gt;
13.	Dobó, A.: [http://solo.bodleian.ox.ac.uk/primo_library/libweb/action/dlDisplay.do?vid=OXVU1&amp;amp;docId=oxfaleph017405947&amp;amp;fn=permalink Interpreting Noun Compounds]. University of Oxford (2010)&lt;br /&gt;
&lt;br /&gt;
14.	Dobó, A.: [http://diploma.bibl.u-szeged.hu/3370/ A Közelítő és szimbolikus számítások tárgy során MATLAB-ban írt zárthelyi dolgozatok automatizált javítása]. University of Szeged (2009)&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12609</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12609"/>
		<updated>2019-08-17T00:37:31Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &amp;quot;IEEE 1ICSC 2018. pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &amp;quot;Journal of Quantitative Linguistics&amp;quot;, pp. 1-28.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [http://www.inf.u-szeged.hu/~dobo/Publications/Comparison%20of%20the%20best%20parameter%20settings%20of%20DSMs%20across%20languages.pdf Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &amp;quot;AIAI 2019: Artificial Intelligence Applications and Innovations&amp;quot;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12606</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12606"/>
		<updated>2019-08-13T00:51:13Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &amp;quot;IEEE 1ICSC 2018. pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &amp;quot;Journal of Quantitative Linguistics&amp;quot;, pp. 1-28.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [https://doi.org/10.1007/978-3-030-19823-7_41 Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &amp;quot;AIAI 2019: Artificial Intelligence Applications and Innovations&amp;quot;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12605</id>
		<title>MEN Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MEN_Test_Collection_(State_of_the_art)&amp;diff=12605"/>
		<updated>2019-08-13T00:49:52Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: Created page with &amp;quot;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013) * 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 p...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* State of the art on the [https://staff.fnwi.uva.nl/e.bruni/MEN MEN dataset] (Bruni et al., 2013)&lt;br /&gt;
* 3000 word pairs: 2000 pairs in the development part of the dataset, 1000 pairs in the test part of the dataset&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 50 subjects&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the test part of the dataset (1000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.867&lt;br /&gt;
| 0.866&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.869&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.866&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| Ch18&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Christopoulou et al. (2018)&lt;br /&gt;
| Corpus-based, predictive&lt;br /&gt;
| 0.84&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.808&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.798&lt;br /&gt;
| 0.798&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.781&lt;br /&gt;
| 0.749&lt;br /&gt;
|-&lt;br /&gt;
| Br13&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Bruni et al. (2013)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.78&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Dobó (2019), Dobó and Csirik (2019b)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.705&lt;br /&gt;
| 0.709&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results for the full dataset (3000 word pairs) ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.846&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-hybrid&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.862&lt;br /&gt;
| 0.865&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.861&lt;br /&gt;
| 0.859&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.809&lt;br /&gt;
| 0.803&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019c)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.802&lt;br /&gt;
| 0.801&lt;br /&gt;
|-&lt;br /&gt;
| DC19a-corpus&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Dobó and Csirik (2019a)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.771&lt;br /&gt;
| 0.746&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based, distributional&lt;br /&gt;
| 0.702&lt;br /&gt;
| 0.707&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Bruni, E., Tran, N. K., and Baroni, M. (2014). [https://www.jair.org/index.php/jair/article/download/10857/25905/ Multimodal distributional semantics]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039;, 49, pp. 1-47.&lt;br /&gt;
&lt;br /&gt;
Christopoulou, F., Briakou, E., Iosif, E., and Potamianos, A. (2018). [https://ieeexplore.ieee.org/abstract/document/8334459/ Mixture of topic-based distributional semantic and affective models]. &amp;quot;IEEE 1ICSC 2018. pp. 203-210. IEEE.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019a). [https://doi.org/10.1080/09296174.2019.1570897 A comprehensive study of the parameters in the creation and comparison of feature vectors in distributional semantic models]. &amp;quot;Journal of Quantitative Linguistics&amp;quot;, pp. 1-28.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2019b). [https://doi.org/10.1007/978-3-030-19823-7_41 Comparison of the best parameter settings in the creation and comparison of feature vectors in distributional semantic models across multiple languages]. &amp;quot;AIAI 2019: Artificial Intelligence Applications and Innovations&amp;quot;, pp. 487-499.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=State_of_the_art&amp;diff=12604</id>
		<title>State of the art</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=State_of_the_art&amp;diff=12604"/>
		<updated>2019-08-13T00:23:10Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The purpose of this section of the ACL wiki is to be a repository of &#039;&#039;k&#039;&#039;-best state-of-the-art results (i.e., methods and software) for various core natural language processing tasks. &lt;br /&gt;
&lt;br /&gt;
As a side effect, this should hopefully evolve into a knowledge base of standard evaluation methods and datasets for various tasks, as well as encourage more effort into reproducibility of results. This will help newcomers to a field appreciate what has been done so far and what the main tasks are, and will help keep active researchers informed on fields other than their specific research. The next time you need a system for PP attachment, or wonder what is the current state of word sense disambiguation, this will be the place to visit. &lt;br /&gt;
&lt;br /&gt;
Please contribute! (This is also a good place for you to display your results!)&lt;br /&gt;
&lt;br /&gt;
As a historical point of reference, you may want to refer to the [http://web.archive.org/web/20100325144600/http://cslu.cse.ogi.edu/HLTsurvey/ Survey of the State of the Art in Human Language Technology] ([http://www.lt-world.org/hlt_survey/master.pdf also available as PDF]), edited by R. Cole, J. Mariani, H. Uszkoreit, G. B. Varile, A. Zaenen, A. Zampolli, V. Zue, 1996.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Please keep this list in alphabetical order --&amp;gt;&lt;br /&gt;
* [[Analogy (State of the art)|Analogy]] -- [[SAT Analogy Questions (State of the art)|SAT]],  [[SemEval-2012 Task 2 (State of the art)|SemEval-2012 Task 2]], [[Syntactic Analogies (State of the art)|Syntactic Analogies]], [[Google analogy test set (State of the art)|Google analogy test set]], [[Bigger analogy test set (State of the art)|Bigger analogy test set]]&lt;br /&gt;
* [[Anaphora Resolution (State of the art)|Anaphora Resolution]] (stub)&lt;br /&gt;
* [[Automatic Text Summarization (State of the art)|Automatic Text Summarization]] (stub)&lt;br /&gt;
* [[Chunking (State of the art)|Chunking]] (stub)&lt;br /&gt;
* [[Dependency Parsing (State of the art)|Dependency Parsing]] (stub)&lt;br /&gt;
* [[Document Classification (State of the art)|Document Classification]] (stub)&lt;br /&gt;
* [[Language Identification (State of the art)|Language Identification]] (stub)&lt;br /&gt;
* [[Named Entity Recognition (State of the art)|Named Entity Recognition]]&lt;br /&gt;
* [[Noun-Modifier Semantic Relations (State of the art)|Noun-Modifier Semantic Relations]]&lt;br /&gt;
* [[NP Chunking (State of the art)|NP Chunking]] &lt;br /&gt;
* [[Paraphrase Identification (State of the art)|Paraphrase Identification]]&lt;br /&gt;
* [[Parsing (State of the art)|Parsing]] &lt;br /&gt;
* [[POS Induction (State of the art) |POS Induction]]&lt;br /&gt;
* [[POS Tagging (State of the art) |POS Tagging]]&lt;br /&gt;
* [[PP Attachment (State of the art)|PP Attachment]] (stub)&lt;br /&gt;
* [[Question Answering (State of the art)|Question Answering]]&lt;br /&gt;
* [[Semantic Role Labeling (State of the art)|Semantic Role Labeling]] (stub)&lt;br /&gt;
* [[Sentiment Analysis (State of the art)|Sentiment Analysis]] (stub)&lt;br /&gt;
* [[Similarity (State of the art)|Similarity]] -- [[ESL Synonym Questions (State of the art)|ESL]], [[SAT Analogy Questions (State of the art)|SAT]], [[TOEFL Synonym Questions (State of the art)|TOEFL]], [[RG-65 Test Collection (State of the art)|RG-65 Test Collection]], [[MC-28 Test Collection (State of the art)|MC-28 Test Collection]], [[SimLex-999 (State of the art)|SimLex-999 Similarity Test Collection]], [[WordSimilarity-353 Test Collection (State of the art)|WordSimilarity-353]], [[SemEval-2012 Task 2 (State of the art)|SemEval-2012 Task 2]], [[MEN Test Collection (State of the art)|MEN Test Collection]]&lt;br /&gt;
* [[Speech Recognition (State of the art)|Speech Recognition]] (article request)&lt;br /&gt;
* [[Temporal Information Extraction (State of the art)|Temporal Information Extraction]]&lt;br /&gt;
* [[Cleaneval (State of the art)| Web Corpus Cleaning]] (stub)&lt;br /&gt;
* [[Word Segmentation (State of the art)|Word Segmentation]] (stub)&lt;br /&gt;
* [[Word Sense Disambiguation (State of the art)|Word Sense Disambiguation]] (stub)&lt;br /&gt;
&amp;lt;!-- Please keep this list in alphabetical order --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Similarity_(State_of_the_art)&amp;diff=12603</id>
		<title>Similarity (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Similarity_(State_of_the_art)&amp;diff=12603"/>
		<updated>2019-08-13T00:21:39Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* see also: [[State of the art]]&lt;br /&gt;
&lt;br /&gt;
== Attributional similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;attributional similarity:&#039;&#039;&#039; the degree to which two words are synonymous&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[TOEFL Synonym Questions (State of the art)|TOEFL Synonym Questions]]&lt;br /&gt;
** [[ESL Synonym Questions (State of the art)|ESL Synonym Questions]]&lt;br /&gt;
** [[RG-65 Test Collection (State of the art)|RG-65 Test Collection]]&lt;br /&gt;
** [[MC-28 Test Collection (State of the art)|MC-28 Test Collection]]&lt;br /&gt;
** [[SimLex-999 (State of the art)|SimLex-999 Similarity Test Collection]]&lt;br /&gt;
** [[WordSimilarity-353 Test Collection (State of the art)|WordSimilarity-353 Test Collection]]&lt;br /&gt;
** [[MEN Test Collection (State of the art)|MEN Test Collection]]&lt;br /&gt;
&lt;br /&gt;
== Similarity versus Association ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;similarity versus association&#039;&#039;&#039;: the contrast between taxonomical similarity (co-hyponymy) and association (co-occurrence)&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Similar-Associated-Both Test Collection (State of the art)|Similar-Associated-Both Test Collection]]&lt;br /&gt;
** [[SimLex-999 (State of the art)|SimLex-999 Similarity Test Collection]]&lt;br /&gt;
&lt;br /&gt;
== Relational similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;relational similarity:&#039;&#039;&#039; the degree to which two relations are analogous&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[SAT Analogy Questions (State of the art)|SAT Analogy Questions]]&lt;br /&gt;
** [[SemEval-2012 Task 2 (State of the art)|SemEval-2012 Task 2: Measuring Degrees of Relational Similarity]]&lt;br /&gt;
** [[Syntactic Analogies|Microsoft Research Syntactic Analogies Dataset]]&lt;br /&gt;
&lt;br /&gt;
== Phrase similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;phrase similarity:&#039;&#039;&#039; the degree to which two phrases are similar&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Noun-Modifier Questions (State of the art)|Noun-Modifier Questions]]&lt;br /&gt;
&lt;br /&gt;
== Sentence similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;sentence similarity:&#039;&#039;&#039; sentence paraphrase, paraphrase identification, paraphrase recognition&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Paraphrase Identification (State of the art)|Microsoft Research Paraphrase Corpus]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
&lt;br /&gt;
* SemEval-2012 Task 2: [https://sites.google.com/site/semeval2012task2/ Measuring Degrees of Relational Similarity]&lt;br /&gt;
* SemEval-2012 Task 6: [http://www.cs.york.ac.uk/semeval-2012/task6/ Semantic Textual Similarity]&lt;br /&gt;
* SEM 2013 Shared Task: [http://ixa2.si.ehu.es/sts/ Semantic Textual Similarity]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=12602</id>
		<title>WordSimilarity-353 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=12602"/>
		<updated>2019-08-12T23:12:17Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
* [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ WordSimilarity-353 Test Collection]&lt;br /&gt;
* contains two sets of English word pairs along with human-assigned similarity judgements&lt;br /&gt;
* first set (set1) contains 153 word pairs along with their similarity scores assigned by 13 subjects&lt;br /&gt;
* second set (set2) contains 200 word pairs with similarity assessed by 16 subjects&lt;br /&gt;
* WordSimilarity-353 dataset is available [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ here]&lt;br /&gt;
* performance is measured by [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rank correlation coefficient]&lt;br /&gt;
* introduced by [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Finkelstein et al. (2002)]&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* [https://www.wikidata.org/wiki/Q31845205 Wikidata] and [https://tools.wmflabs.org/scholia/use/Q31845205 Scholia]&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of increasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! Spearman&#039;s rho&lt;br /&gt;
! Pearson&#039;s r&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.305&lt;br /&gt;
| 0.271&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath 1997&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.318&lt;br /&gt;
| 0.354&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.341&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.357&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.353&lt;br /&gt;
| 0.365&lt;br /&gt;
|-&lt;br /&gt;
| ROGET&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.415&lt;br /&gt;
| 0.536&lt;br /&gt;
|-&lt;br /&gt;
| C&amp;amp;W&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.5&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.48&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.579&lt;br /&gt;
| 0.577&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.492&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.563&lt;br /&gt;
|-&lt;br /&gt;
| simVB+simWN&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.55&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.622&lt;br /&gt;
| 0.629&lt;br /&gt;
|-&lt;br /&gt;
| HSMN+csmRNN&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.65&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.706&lt;br /&gt;
| 0.705&lt;br /&gt;
|-&lt;br /&gt;
| Multi-prototype&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.71&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Multi-lingual SSA&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.674&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.733&lt;br /&gt;
| 0.704&lt;br /&gt;
|-&lt;br /&gt;
| ESA&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.748&lt;br /&gt;
| 0.503&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.795&lt;br /&gt;
| 0.276&lt;br /&gt;
|-&lt;br /&gt;
| TSA&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.80&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| CLEAR&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| ConceptNet Numberbatch&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.828&lt;br /&gt;
| N/A&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in alphabetical order.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Finkelstein, Lev, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. (2002) [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Placing Search in Context: The Concept Revisited]. ACM Transactions on Information Systems, 20(1):116-131.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Halawi, Guy, Gideon Dror, Evgeniy Gabrilovich, and Yehuda Koren. (2012). [http://gabrilovich.com/publications/papers/Halawi2012LSL.pdf Large-scale learning of word relatedness with constraints]. In Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 1406-1414. ACM.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.aaai.org/ocs/index.php/AAAI/AAAI11/paper/download/3616/3972/ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Huang, Eric H., Richard Socher, Christopher D. Manning, and Andrew Y. Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers - Volume 1 (ACL &#039;12), Vol. 1. Association for Computational Linguistics, Stroudsburg, PA, USA, 873-882.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Luong, Minh-Thang, Richard Socher, and Christopher D. Manning. (2013). [http://nlp.stanford.edu/~lmthang/data/papers/conll13_morpho.pdf Better word representations with recursive neural networks for morphology]. CoNLL-2013: 104.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., D. Jurgens and R. Navigli. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Radinsky, Kira, Eugene Agichtein, Evgeniy Gabrilovich, and Shaul Markovitch. (2011). [http://gabrilovich.com/publications/papers/Radinsky2011WTS.pdf A word at a time: computing word relatedness using temporal semantic analysis]. In Proceedings of the 20th international conference on World wide web, pp. 337-346. ACM.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, Rob, Joshua Chin and Catherine Havasi. (2017). [http://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14972 ConceptNet 5.5: An Open Multilingual Graph of General Knowledge]. Proceedings of The 31st AAAI Conference on Artificial Intelligence, San Francisco, CA.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael and Simone Paolo Ponzetto. (2006). [http://www.aaai.org/Papers/AAAI/2006/AAAI06-223.pdf WikiRelate! Computing Semantic Relatedness Using Wikipedia]. Proceedings of The 21st National Conference on Artificial Intelligence (AAAI), Boston, MA.&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=SimLex-999_(State_of_the_art)&amp;diff=12601</id>
		<title>SimLex-999 (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=SimLex-999_(State_of_the_art)&amp;diff=12601"/>
		<updated>2019-08-12T23:00:55Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.cl.cam.ac.uk/~fh295/simlex.html SimLex-999] aims at a cleaner benchmark of similarity (but not relatedness). Pairs of words were chosen to represent different ranges of similarity and with either high or low association. Subjects were instructed to differentiate between similarity and relatedness and rate regarding the former only.&lt;br /&gt;
&lt;br /&gt;
See also: [[Similarity (State of the art)]], [[Similar-Associated-Both Test Collection (State of the art)]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{|&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm !! Reference for algorithm !! Reference for reported results  !! Type !! Spearman&#039;s rho !! Pearson&#039;s r !! Notes&lt;br /&gt;
|-&lt;br /&gt;
| Re16&lt;br /&gt;
| Recski et al. (2016)&amp;lt;ref name=recski16&amp;gt;Recski, G., Iklódi, E., Pajkossy, K., &amp;amp; Kornai, A. (2016). [https://www.aclweb.org/anthology/W16-1622 Measuring semantic similarity of words using concept networks]. In: Proceedings of the 1st Workshop on Representation Learning for NLP, pp. 193-200.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Recski et al. (2016)&amp;lt;ref name=recski16/&amp;gt;&lt;br /&gt;
| Hybrid || 0.76 || -&lt;br /&gt;
|-&lt;br /&gt;
| SVR4&lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| Combined || 0.642 || 0.658&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19&amp;gt;Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Hybrid || 0.621 || 0.481&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&amp;lt;ref name=speer17&amp;gt;Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. AAAI-17, pp. 4444-4451.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Hybrid || 0.616 || 0.634&lt;br /&gt;
|-&lt;br /&gt;
| joint(SP+,skip-gram)&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus&amp;gt;Schwartz, R., Reichart, Roi, Rappoport, A. (2015). Symmetric Pattern Based Word Embeddings for Improved Word Similarity Prediction, CoNLL 2015.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus/&amp;gt;&lt;br /&gt;
| Distributional || 0.56 || - || Trained on word2vec corpus, best results for pure distributional model.&lt;br /&gt;
|-&lt;br /&gt;
| UMBC&lt;br /&gt;
| Han et al. (2013)&amp;lt;ref&amp;gt;Han, L., Kashyap, A., Finin, T., Mayfield, J., Weese, J.: UMBC EBIQUITY-CORE: Semantic textual similarity systems. In: Proceedings of the Second Joint Conference on Lexical and Computational Semantics, vol. 1, pp. 44–52 (2013)&amp;lt;/ref&amp;gt; &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| || 0.558 || 0.557 || without using POS information&lt;br /&gt;
|-&lt;br /&gt;
| SP+&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus&amp;gt;Schwartz, R., Reichart, Roi, Rappoport, A. (2015). Symmetric Pattern Based Word Embeddings for Improved Word Similarity Prediction, CoNLL 2015.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Schwartz et al. (2015)&amp;lt;ref name=spplus/&amp;gt;&lt;br /&gt;
| Distributional || 0.52 || -&lt;br /&gt;
|-&lt;br /&gt;
| RNNenc&lt;br /&gt;
| Hill et al. (2014b)&amp;lt;ref name=rnnenc&amp;gt;Hill, F., Cho, K., Jean, S., Devin, C., &amp;amp; Bengio, Y. (2014b). Not All Neural Embeddings are Born Equal, 1–5.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014b)&amp;lt;ref name=rnnenc/&amp;gt;&lt;br /&gt;
| Distributional, multilingual || 0.52 || -&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&amp;lt;ref name=salle18&amp;gt;Salle A., Idiart M., and Villavicencio A. (2018). [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&amp;lt;/ref&amp;gt; &lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.417 || 0.426&lt;br /&gt;
|-&lt;br /&gt;
| Word2vec&lt;br /&gt;
| Mikolov et al. (2013)&amp;lt;ref&amp;gt;Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex&amp;gt;Hill, F., Reichart, R., &amp;amp; Korhonen, A. (2014a). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Distributional || 0.414 || - || Trained on Wikipedia&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&amp;lt;ref name=pennington14&amp;gt;Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. EMNLP 2014, pp. 1532-1543.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.406 || 0.433&lt;br /&gt;
|-&lt;br /&gt;
| Lesk&lt;br /&gt;
| &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea&amp;gt;Banjade, R., Maharjan, N., Niraula, N., Rus, V., &amp;amp; Gautam, D. (2015). Lemon and Tea Are Not Similar: Measuring Word-to-Word Similarity by Combining Different Methods. Computational Linguistics and Intelligent Text Processing, 9041, 335–346. doi:10.1007/978-3-319-18111-0_25&amp;lt;/ref&amp;gt;&lt;br /&gt;
| || 0.404 || 0.347&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Dobó (2019)&amp;lt;ref name=dobo19/&amp;gt;&lt;br /&gt;
| Distributional || 0.393 || 0.401&lt;br /&gt;
|-&lt;br /&gt;
| ESA&lt;br /&gt;
| &lt;br /&gt;
| Banjade et al. (2015)&amp;lt;ref name=lemontea/&amp;gt;&lt;br /&gt;
| || 0.271 || 0.145&lt;br /&gt;
|-&lt;br /&gt;
| Neural language model&lt;br /&gt;
| Collobert &amp;amp; Weston (2008)&amp;lt;ref&amp;gt;R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex/&amp;gt; &lt;br /&gt;
| Distributional || 0.268 || - || Trained on Wikipedia&lt;br /&gt;
|-&lt;br /&gt;
| Neural language model with global context&lt;br /&gt;
| Huang et al. (2012)&amp;lt;ref&amp;gt;Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.&amp;lt;/ref&amp;gt;&lt;br /&gt;
| Hill et al. (2014a)&amp;lt;ref name=simlex/&amp;gt; &lt;br /&gt;
| Distributional || 0.098 || - || Trained on Wikipedia&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12600</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12600"/>
		<updated>2019-08-12T22:54:15Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset [Resnik, 1995]&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset [Miller and Charles, 1991], which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.893 [0.780, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.892 [0.778, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.853 [0.704, 0.930]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.832 [0.666, 0.919]&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.822 [0.648, 0.914]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12599</id>
		<title>RG-65 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12599"/>
		<updated>2019-08-12T22:53:54Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Rubenstein &amp;amp; Goodenough (RG-65) dataset&lt;br /&gt;
* 65 word pairs; &lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 51 subjects [Rubenstein and Goodenough, 1965].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Knowledge-based (Wiktionary)&lt;br /&gt;
| 0.920&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.901&lt;br /&gt;
| 0.896&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.899&lt;br /&gt;
| 0.914&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.890&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| NASARI&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.880&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Knowledge-based (WordNet)&lt;br /&gt;
| 0.868&lt;br /&gt;
| 0.810&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.838&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.833&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.830&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.732&lt;br /&gt;
|-&lt;br /&gt;
| Roget&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.818&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.731&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.787&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.797&lt;br /&gt;
| 0.852&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.788&lt;br /&gt;
| 0.834&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.769&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.763&lt;br /&gt;
| 0.792&lt;br /&gt;
|-&lt;br /&gt;
| ESA*&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based &lt;br /&gt;
| 0.749&lt;br /&gt;
| 0.716&lt;br /&gt;
|-&lt;br /&gt;
| SOCPMI*&lt;br /&gt;
| Islam and Inkpen (2006)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.741&lt;br /&gt;
| 0.729&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.732&lt;br /&gt;
| 0.737&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.731&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.640&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| LSA*&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.644&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| -&lt;br /&gt;
| 0.530&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Note: values reported by (Hassan and Mihalcea, 2011) are &amp;quot;based on the collected raw data from the respective authors&amp;quot;, and those highlighted by (*) are re-implementations.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, Eneko, Enrique Alfonseca, Keith Hall, Jana Kravalova, Marius Pasca, Aitor Soroa: [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. HLT-NAACL 2009: 19-27&lt;br /&gt;
&lt;br /&gt;
Camacho-Collados, José, Pilehvar, Mohammad Taher, and Navigli, Roberto: [http://aclweb.org/anthology/N/N15/N15-1059.pdf NASARI: a Novel Approach to a Semantically­-Aware Representation of Items]. NAACL 2015, pp. 567-577, Denver, USA.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.cse.unt.edu/~rada/papers/hassan.aaai11.pdf‎ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Hughes, Thad, Daniel Ramage, Lexical Semantic Relatedness with Random Graph Walks. EMNLP-CoNLL 2007: 581-589.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Milne, David, and Ian H. Witten, An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links, In Proceedings of AAAI 2008.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens, D. and Navigli, R. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T. and Navigli, R. [http://www.sciencedirect.com/science/article/pii/S000437021500106X From Senses to Texts: An All-in-one Graph-based Approach for Measuring Semantic Similarity]. Artificial Intelligence, Elsevier.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Rubenstein, Herbert,  and John B. Goodenough. Contextual correlates of synonymy. Communications of the ACM, 8(10):627–633, 1965.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael, Simone Paolo Ponzetto: WikiRelate! Computing Semantic Relatedness Using Wikipedia. AAAI 2006: 1419-1424&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12598</id>
		<title>TOEFL Synonym Questions (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12598"/>
		<updated>2019-08-12T22:53:19Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* &#039;&#039;&#039;the TOEFL questions are available on request by contacting [http://lsa.colorado.edu/mail_sub.html LSA Support at CU Boulder]&#039;&#039;&#039;, the people who manage the [http://lsa.colorado.edu/ LSA web site at Colorado]&lt;br /&gt;
* TOEFL = Test of English as a Foreign Language&lt;br /&gt;
* 80 multiple-choice synonym questions; 4 choices per question&lt;br /&gt;
* introduced in Landauer and Dumais (1997) as a way of evaluating algorithms for measuring degree of similarity between words&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sample question ==&lt;br /&gt;
&lt;br /&gt;
::{| border=&amp;quot;0&amp;quot; cellpadding=&amp;quot;1&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Stem:&lt;br /&gt;
|&lt;br /&gt;
| levied&lt;br /&gt;
|-&lt;br /&gt;
! Choices:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (b)&lt;br /&gt;
| believed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (c)&lt;br /&gt;
| requested&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (d)&lt;br /&gt;
| correlated&lt;br /&gt;
|-&lt;br /&gt;
! Solution:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for experiment&lt;br /&gt;
! Type&lt;br /&gt;
! Correct&lt;br /&gt;
! 95% confidence&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 20.31%&lt;br /&gt;
| 12.89–31.83%&lt;br /&gt;
|-&lt;br /&gt;
| LC&lt;br /&gt;
| Leacock and Chodrow (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 21.88%&lt;br /&gt;
| 13.91–33.21%&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 24.06%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| Random&lt;br /&gt;
| Random guessing&lt;br /&gt;
| 1 / 4 = 25.00%&lt;br /&gt;
| Random&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| JC&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 64.38%&lt;br /&gt;
| 52.90–74.80%&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Average non-English US college applicant&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Human&lt;br /&gt;
| 64.50%&lt;br /&gt;
| 53.01–74.88%&lt;br /&gt;
|-&lt;br /&gt;
| RI&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 72.50%&lt;br /&gt;
| 61.38-81.90%&lt;br /&gt;
|-&lt;br /&gt;
| DS&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.00%&lt;br /&gt;
| 62.72-82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.75%&lt;br /&gt;
| 62.72–82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PairClass&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 76.25%&lt;br /&gt;
| 65.42-85.06%&lt;br /&gt;
|-&lt;br /&gt;
| HSO&lt;br /&gt;
| Hirst and St.-Onge (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 77.91%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 78.75%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 80.00%&lt;br /&gt;
| 69.56–88.11%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| LC-IR&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| CWO&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 82.55%&lt;br /&gt;
| 72.38–90.09%&lt;br /&gt;
|-&lt;br /&gt;
| PPMIC&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 85.00%&lt;br /&gt;
| 75.26-92.00%&lt;br /&gt;
|-&lt;br /&gt;
| GLSA&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 86.25%&lt;br /&gt;
| 76.73-92.93%&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 87.50%&lt;br /&gt;
| 78.21-93.84%&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 88.75%&lt;br /&gt;
| 79.72-94.72%&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 90.00%&lt;br /&gt;
| 81.24-95.58%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 92.50%&lt;br /&gt;
| 84.39-97.20%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 95.0%&lt;br /&gt;
| 87.69-98.62%&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| WordNet graph-based (unsupervised)&lt;br /&gt;
| 96.25%&lt;br /&gt;
| 89.43-99.22%&lt;br /&gt;
|-&lt;br /&gt;
| PR&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 97.50%&lt;br /&gt;
| 91.26–99.70%&lt;br /&gt;
|-&lt;br /&gt;
| Sp19&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| PCCP&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 100.00%&lt;br /&gt;
| 96.32-100.00%&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Explanation of table ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Algorithm&#039;&#039;&#039; = name of algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for algorithm&#039;&#039;&#039; = where to find out more about given algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for experiment&#039;&#039;&#039; = where to find out more about evaluation of given algorithm with TOEFL questions&lt;br /&gt;
* &#039;&#039;&#039;Type&#039;&#039;&#039; = general type of algorithm: corpus-based, lexicon-based, hybrid&lt;br /&gt;
* &#039;&#039;&#039;Correct&#039;&#039;&#039; = percent of 80 questions that given algorithm answered correctly&lt;br /&gt;
* &#039;&#039;&#039;95% confidence&#039;&#039;&#039; = confidence interval calculated using the [[Statistical calculators|Binomial Exact Test]]&lt;br /&gt;
* table rows sorted in order of increasing percent correct&lt;br /&gt;
* several WordNet-based similarity measures are implemented in [http://www.d.umn.edu/~tpederse/ Ted Pedersen]&#039;s [http://www.d.umn.edu/~tpederse/similarity.html WordNet::Similarity] package&lt;br /&gt;
* LSA = Latent Semantic Analysis&lt;br /&gt;
* PCCP = Principal Component vectors with Caron P&lt;br /&gt;
* PMI-IR = Pointwise Mutual Information - Information Retrieval&lt;br /&gt;
* PR = Product Rule&lt;br /&gt;
* PPMIC = Positive Pointwise Mutual Information with Cosine&lt;br /&gt;
* GLSA = Generalized Latent Semantic Analysis&lt;br /&gt;
* CWO = Context Window Overlapping&lt;br /&gt;
* DS = Dependency Space&lt;br /&gt;
* RI = Random Indexing&lt;br /&gt;
&lt;br /&gt;
== Notes ==&lt;br /&gt;
&lt;br /&gt;
* the performance of a corpus-based algorithm depends on the corpus, so the difference in performance between two corpus-based systems may be due to the different corpora, rather than the different algorithms&lt;br /&gt;
* the TOEFL questions include nouns, verbs, and adjectives, but some of the WordNet-based algorithms were only designed to work with nouns; this explains some of the lower scores&lt;br /&gt;
* some of the algorithms may have been tuned on the TOEFL questions; read the references for details&lt;br /&gt;
* Landauer and Dumais (1997) report scores that were corrected for guessing by subtracting a penalty of 1/3 for each incorrect answer; they report a score of 52.5% when this penalty is applied; when the penalty is removed, their performance is 64.4% correct&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2007). [http://www.cs.bham.ac.uk/~jxb/PUBS/BRM.pdf Extracting semantic representations from word co-occurrence statistics: A computational study]. &#039;&#039;Behavior Research Methods&#039;&#039;, 39(3), 510-526.&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2012). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.228.9582&amp;amp;rep=rep1&amp;amp;type=pdf Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and SVD]. &#039;&#039;Behavior Research Methods&#039;&#039;,  44(3):890-907.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502&lt;br /&gt;
&lt;br /&gt;
Lushan Han. (2014). [http://ebiquity.umbc.edu/paper/html/id/658/Schema-Free-Querying-of-Semantic-Data Schema Free Querying of Semantic Data], Ph.D. dissertation, University of Maryland, Baltimore County, Baltimore, MD USA.&lt;br /&gt;
&lt;br /&gt;
Higgins, D. (2005). [http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.329.1517 Which Statistics Reflect Semantics? Rethinking Synonymy and Word Similarity.] In: Kepser, S., Reis, M. (eds.) &#039;&#039;Linguistic Evidence: Empirical, Theoretical and Computational Perspectives&#039;&#039;. Mouton de Gruyter, Berlin, pp. 265–284.&lt;br /&gt;
&lt;br /&gt;
Hirst, G., and St-Onge, D. (1998). [http://mirror.eacoss.org/documentation/ITLibrary/IRIS/Data/1997/Hirst/Lexical/1997-Hirst-Lexical.pdf Lexical chains as representation of context for the detection and correction of malapropisms]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, 305-332.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Jiang, J.J., and Conrath, D.W. (1997). [http://wortschatz.uni-leipzig.de/~sbordag/aalw05/Referate/03_Assoziationen_BudanitskyResnik/Jiang_Conrath_97.pdf Semantic similarity based on corpus statistics and lexical taxonomy]. &#039;&#039;Proceedings of the International Conference on Research in Computational Linguistics&#039;&#039;, Taiwan.&lt;br /&gt;
&lt;br /&gt;
Karlgren, J. and Sahlgren, M. (2001). [http://www.sics.se/~jussi/Artiklar/2001_RWIbook/KarlgrenSahlgren2001.pdf From Words to Understanding]. In Uesaka, Y., Kanerva, P., &amp;amp; Asoh, H. (Eds.), &#039;&#039;Foundations of Real-World Intelligence&#039;&#039;, Stanford: CSLI Publications, pp. 294–308. &lt;br /&gt;
&lt;br /&gt;
Landauer, T.K., and Dumais, S.T. (1997). [http://lsa.colorado.edu/papers/plato/plato.annote.html A solution to Plato&#039;s problem: The latent semantic analysis theory of the acquisition, induction, and representation of knowledge]. &#039;&#039;Psychological Review&#039;&#039;, 104(2):211–240.&lt;br /&gt;
&lt;br /&gt;
Leacock, C., and Chodorow, M. (1998). [http://books.google.ca/books?id=Rehu8OOzMIMC&amp;amp;lpg=PA265&amp;amp;ots=IpnaLkZUec&amp;amp;lr&amp;amp;pg=PA265#v=onepage&amp;amp;q&amp;amp;f=false Combining local context and WordNet similarity for word sense identification]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, pp. 265-283.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://www.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. &#039;&#039;Proceedings of the 15th International Conference on Machine Learning (ICML-98)&#039;&#039;, Madison, WI, pp. 296-304.&lt;br /&gt;
&lt;br /&gt;
Matveeva, I., Levow, G., Farahat, A., and Royer, C. (2005). [http://people.cs.uchicago.edu/~matveeva/SynGLSA_ranlp_final.pdf Generalized latent semantic analysis for term representation]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-05)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Pado, S., and Lapata, M. (2007). [http://www.nlpado.de/~sebastian/pub/papers/cl07_pado.pdf Dependency-based construction of semantic space models]. &#039;&#039;Computational Linguistics&#039;&#039;, 33(2), 161-199.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation]. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens D., and Navigli R. (2013). [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, disambiguate and walk: A unified approach for measuring semantic similarity]. &#039;&#039;Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013),&#039;&#039; Sofia, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Rapp, R. (2003). [http://www.amtaweb.org/summit/MTSummit/FinalPapers/19-Rapp-final.pdf Word sense discovery based on sense descriptor dissimilarity]. &#039;&#039;Proceedings of the Ninth Machine Translation Summit&#039;&#039;, pp. 315-322.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://citeseer.ist.psu.edu/resnik95using.html Using information content to evaluate semantic similarity]. &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI-95)&#039;&#039;, Montreal, pp. 448-453.&lt;br /&gt;
&lt;br /&gt;
Ruiz-Casado, M., Alfonseca, E. and Castells, P. (2005) [http://alfonseca.org/pubs/2005-ranlp1.pdf Using context-window overlapping in Synonym Discovery and Ontology Extension]. &#039;&#039;Proceedings of the International Conference Recent Advances in Natural Language Processing (RANLP-2005)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Terra, E., and Clarke, C.L.A. (2003). [http://acl.ldc.upenn.edu/N/N03/N03-1032.pdf Frequency estimates for statistical word similarity measures]. &#039;&#039;Proceedings of the Human Language Technology and North American Chapter of Association of Computational Linguistics Conference 2003 (HLT/NAACL 2003)&#039;&#039;, pp. 244–251.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2001). [http://arxiv.org/abs/cs.LG/0212033 Mining the Web for synonyms: PMI-IR versus LSA on TOEFL]. &#039;&#039;Proceedings of the Twelfth European Conference on Machine Learning (ECML-2001)&#039;&#039;, Freiburg, Germany, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D., Littman, M.L., Bigham, J., and Shnayder, V. (2003). [http://arxiv.org/abs/cs.CL/0309035 Combining independent modules to solve multiple-choice synonym and analogy problems]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, pp. 482-489.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2008). [http://arxiv.org/abs/0809.0124 A uniform approach to analogies, synonyms, antonyms, and associations]. &#039;&#039;Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)&#039;&#039;, Manchester, UK, pp. 905-912.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12597</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=12597"/>
		<updated>2019-08-12T22:20:54Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset [Resnik, 1995]&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset [Miller and Charles, 1991], which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.893 [0.780, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.892 [0.778, 0.949]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.853 [0.704, 0.930]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.832 [0.666, 0.919]&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.822 [0.648, 0.914]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12596</id>
		<title>RG-65 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=RG-65_Test_Collection_(State_of_the_art)&amp;diff=12596"/>
		<updated>2019-08-12T22:09:03Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Rubenstein &amp;amp; Goodenough (RG-65) dataset&lt;br /&gt;
* 65 word pairs; &lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 51 subjects [Rubenstein and Goodenough, 1965].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] (ρ)&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient Pearson correlation] (r)&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Pilehvar and Navigli (2015)&lt;br /&gt;
| Knowledge-based (Wiktionary)&lt;br /&gt;
| 0.920&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| Sp17&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.901&lt;br /&gt;
| 0.896&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.899&lt;br /&gt;
| 0.914&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.890&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| NASARI&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Camacho-Collados et al. (2015)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.880&lt;br /&gt;
| 0.910&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Knowledge-based (WordNet)&lt;br /&gt;
| 0.868&lt;br /&gt;
| 0.810&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Hughes and Ramage (2007)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.838&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.833&lt;br /&gt;
| 0.861&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.830&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.813&lt;br /&gt;
| 0.732&lt;br /&gt;
|-&lt;br /&gt;
| Roget&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.818&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.804&lt;br /&gt;
| 0.731&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.801&lt;br /&gt;
| 0.787&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.797&lt;br /&gt;
| 0.852&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.788&lt;br /&gt;
| 0.834&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.769&lt;br /&gt;
| 0.770&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.763&lt;br /&gt;
| 0.792&lt;br /&gt;
|-&lt;br /&gt;
| ESA*&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based &lt;br /&gt;
| 0.749&lt;br /&gt;
| 0.716&lt;br /&gt;
|-&lt;br /&gt;
| SOCPMI*&lt;br /&gt;
| Islam and Inkpen (2006)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.741&lt;br /&gt;
| 0.729&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.732&lt;br /&gt;
| 0.737&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.731&lt;br /&gt;
| 0.800&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.640&lt;br /&gt;
| -&lt;br /&gt;
|-&lt;br /&gt;
| LSA*&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.609&lt;br /&gt;
| 0.644&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| -&lt;br /&gt;
| 0.530&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Note: values reported by (Hassan and Mihalcea, 2011) are &amp;quot;based on the collected raw data from the respective authors&amp;quot;, and those highlighted by (*) are re-implementations.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, Eneko, Enrique Alfonseca, Keith Hall, Jana Kravalova, Marius Pasca, Aitor Soroa: [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. HLT-NAACL 2009: 19-27&lt;br /&gt;
&lt;br /&gt;
Camacho-Collados, José, Pilehvar, Mohammad Taher, and Navigli, Roberto: [http://aclweb.org/anthology/N/N15/N15-1059.pdf NASARI: a Novel Approach to a Semantically­-Aware Representation of Items]. NAACL 2015, pp. 567-577, Denver, USA.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.cse.unt.edu/~rada/papers/hassan.aaai11.pdf‎ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Hughes, Thad, Daniel Ramage, Lexical Semantic Relatedness with Random Graph Walks. EMNLP-CoNLL 2007: 581-589.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Milne, David, and Ian H. Witten, An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links, In Proceedings of AAAI 2008.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens, D. and Navigli, R. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T. and Navigli, R. [http://www.sciencedirect.com/science/article/pii/S000437021500106X From Senses to Texts: An All-in-one Graph-based Approach for Measuring Semantic Similarity]. Artificial Intelligence, Elsevier.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Rubenstein, Herbert,  and John B. Goodenough. Contextual correlates of synonymy. Communications of the ACM, 8(10):627–633, 1965.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael, Simone Paolo Ponzetto: WikiRelate! Computing Semantic Relatedness Using Wikipedia. AAAI 2006: 1419-1424&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12595</id>
		<title>TOEFL Synonym Questions (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=12595"/>
		<updated>2019-08-12T21:39:57Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* &#039;&#039;&#039;the TOEFL questions are available on request by contacting [http://lsa.colorado.edu/mail_sub.html LSA Support at CU Boulder]&#039;&#039;&#039;, the people who manage the [http://lsa.colorado.edu/ LSA web site at Colorado]&lt;br /&gt;
* TOEFL = Test of English as a Foreign Language&lt;br /&gt;
* 80 multiple-choice synonym questions; 4 choices per question&lt;br /&gt;
* introduced in Landauer and Dumais (1997) as a way of evaluating algorithms for measuring degree of similarity between words&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sample question ==&lt;br /&gt;
&lt;br /&gt;
::{| border=&amp;quot;0&amp;quot; cellpadding=&amp;quot;1&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Stem:&lt;br /&gt;
|&lt;br /&gt;
| levied&lt;br /&gt;
|-&lt;br /&gt;
! Choices:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (b)&lt;br /&gt;
| believed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (c)&lt;br /&gt;
| requested&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (d)&lt;br /&gt;
| correlated&lt;br /&gt;
|-&lt;br /&gt;
! Solution:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for experiment&lt;br /&gt;
! Type&lt;br /&gt;
! Correct&lt;br /&gt;
! 95% confidence&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 20.31%&lt;br /&gt;
| 12.89–31.83%&lt;br /&gt;
|-&lt;br /&gt;
| LC&lt;br /&gt;
| Leacock and Chodrow (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 21.88%&lt;br /&gt;
| 13.91–33.21%&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 24.06%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| Random&lt;br /&gt;
| Random guessing&lt;br /&gt;
| 1 / 4 = 25.00%&lt;br /&gt;
| Random&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| JC&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 64.38%&lt;br /&gt;
| 52.90–74.80%&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Average non-English US college applicant&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Human&lt;br /&gt;
| 64.50%&lt;br /&gt;
| 53.01–74.88%&lt;br /&gt;
|-&lt;br /&gt;
| RI&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 72.50%&lt;br /&gt;
| 61.38-81.90%&lt;br /&gt;
|-&lt;br /&gt;
| DS&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.00%&lt;br /&gt;
| 62.72-82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.75%&lt;br /&gt;
| 62.72–82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PairClass&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 76.25%&lt;br /&gt;
| 65.42-85.06%&lt;br /&gt;
|-&lt;br /&gt;
| HSO&lt;br /&gt;
| Hirst and St.-Onge (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 77.91%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 78.75%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| Sa18&lt;br /&gt;
| Salle et al. (2018)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 80.00%&lt;br /&gt;
| 69.56–88.11%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| LC-IR&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-corpus&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| CWO&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 82.55%&lt;br /&gt;
| 72.38–90.09%&lt;br /&gt;
|-&lt;br /&gt;
| PPMIC&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 85.00%&lt;br /&gt;
| 75.26-92.00%&lt;br /&gt;
|-&lt;br /&gt;
| GLSA&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 86.25%&lt;br /&gt;
| 76.73-92.93%&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 87.50%&lt;br /&gt;
| 78.21-93.84%&lt;br /&gt;
|-&lt;br /&gt;
| DC13&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 88.75%&lt;br /&gt;
| 79.72-94.72%&lt;br /&gt;
|-&lt;br /&gt;
| Pe14&lt;br /&gt;
| Pennington et al. (2014)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 90.00%&lt;br /&gt;
| 81.24-95.58%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 92.50%&lt;br /&gt;
| 84.39-97.20%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 95.0%&lt;br /&gt;
| 87.69-98.62%&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| WordNet graph-based (unsupervised)&lt;br /&gt;
| 96.25%&lt;br /&gt;
| 89.43-99.22%&lt;br /&gt;
|-&lt;br /&gt;
| PR&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 97.50%&lt;br /&gt;
| 91.26–99.70%&lt;br /&gt;
|-&lt;br /&gt;
| Sp19&lt;br /&gt;
| Speer et al. (2017)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| Do19-hybrid&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Dobó (2019)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 98.75%&lt;br /&gt;
| 93.23–99.97%&lt;br /&gt;
|-&lt;br /&gt;
| PCCP&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 100.00%&lt;br /&gt;
| 96.32-100.00%&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Explanation of table ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Algorithm&#039;&#039;&#039; = name of algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for algorithm&#039;&#039;&#039; = where to find out more about given algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for experiment&#039;&#039;&#039; = where to find out more about evaluation of given algorithm with TOEFL questions&lt;br /&gt;
* &#039;&#039;&#039;Type&#039;&#039;&#039; = general type of algorithm: corpus-based, lexicon-based, hybrid&lt;br /&gt;
* &#039;&#039;&#039;Correct&#039;&#039;&#039; = percent of 80 questions that given algorithm answered correctly&lt;br /&gt;
* &#039;&#039;&#039;95% confidence&#039;&#039;&#039; = confidence interval calculated using the [[Statistical calculators|Binomial Exact Test]]&lt;br /&gt;
* table rows sorted in order of increasing percent correct&lt;br /&gt;
* several WordNet-based similarity measures are implemented in [http://www.d.umn.edu/~tpederse/ Ted Pedersen]&#039;s [http://www.d.umn.edu/~tpederse/similarity.html WordNet::Similarity] package&lt;br /&gt;
* LSA = Latent Semantic Analysis&lt;br /&gt;
* PCCP = Principal Component vectors with Caron P&lt;br /&gt;
* PMI-IR = Pointwise Mutual Information - Information Retrieval&lt;br /&gt;
* PR = Product Rule&lt;br /&gt;
* PPMIC = Positive Pointwise Mutual Information with Cosine&lt;br /&gt;
* GLSA = Generalized Latent Semantic Analysis&lt;br /&gt;
* CWO = Context Window Overlapping&lt;br /&gt;
* DS = Dependency Space&lt;br /&gt;
* RI = Random Indexing&lt;br /&gt;
&lt;br /&gt;
== Notes ==&lt;br /&gt;
&lt;br /&gt;
* the performance of a corpus-based algorithm depends on the corpus, so the difference in performance between two corpus-based systems may be due to the different corpora, rather than the different algorithms&lt;br /&gt;
* the TOEFL questions include nouns, verbs, and adjectives, but some of the WordNet-based algorithms were only designed to work with nouns; this explains some of the lower scores&lt;br /&gt;
* some of the algorithms may have been tuned on the TOEFL questions; read the references for details&lt;br /&gt;
* Landauer and Dumais (1997) report scores that were corrected for guessing by subtracting a penalty of 1/3 for each incorrect answer; they report a score of 52.5% when this penalty is applied; when the penalty is removed, their performance is 64.4% correct&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2007). [http://www.cs.bham.ac.uk/~jxb/PUBS/BRM.pdf Extracting semantic representations from word co-occurrence statistics: A computational study]. &#039;&#039;Behavior Research Methods&#039;&#039;, 39(3), 510-526.&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2012). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.228.9582&amp;amp;rep=rep1&amp;amp;type=pdf Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and SVD]. &#039;&#039;Behavior Research Methods&#039;&#039;,  44(3):890-907.&lt;br /&gt;
&lt;br /&gt;
Dobó, A. (2019). [http://doktori.bibl.u-szeged.hu/10120/1/AndrasDoboThesis2019.pdf A comprehensive analysis of the parameters in the creation and comparison of feature vectors in distributional semantic models for multiple languages]. University of Szeged.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502&lt;br /&gt;
&lt;br /&gt;
Lushan Han. (2014). [http://ebiquity.umbc.edu/paper/html/id/658/Schema-Free-Querying-of-Semantic-Data Schema Free Querying of Semantic Data], Ph.D. dissertation, University of Maryland, Baltimore County, Baltimore, MD USA.&lt;br /&gt;
&lt;br /&gt;
Higgins, D. (2005). [http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.329.1517 Which Statistics Reflect Semantics? Rethinking Synonymy and Word Similarity.] In: Kepser, S., Reis, M. (eds.) &#039;&#039;Linguistic Evidence: Empirical, Theoretical and Computational Perspectives&#039;&#039;. Mouton de Gruyter, Berlin, pp. 265–284.&lt;br /&gt;
&lt;br /&gt;
Hirst, G., and St-Onge, D. (1998). [http://mirror.eacoss.org/documentation/ITLibrary/IRIS/Data/1997/Hirst/Lexical/1997-Hirst-Lexical.pdf Lexical chains as representation of context for the detection and correction of malapropisms]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, 305-332.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Jiang, J.J., and Conrath, D.W. (1997). [http://wortschatz.uni-leipzig.de/~sbordag/aalw05/Referate/03_Assoziationen_BudanitskyResnik/Jiang_Conrath_97.pdf Semantic similarity based on corpus statistics and lexical taxonomy]. &#039;&#039;Proceedings of the International Conference on Research in Computational Linguistics&#039;&#039;, Taiwan.&lt;br /&gt;
&lt;br /&gt;
Karlgren, J. and Sahlgren, M. (2001). [http://www.sics.se/~jussi/Artiklar/2001_RWIbook/KarlgrenSahlgren2001.pdf From Words to Understanding]. In Uesaka, Y., Kanerva, P., &amp;amp; Asoh, H. (Eds.), &#039;&#039;Foundations of Real-World Intelligence&#039;&#039;, Stanford: CSLI Publications, pp. 294–308. &lt;br /&gt;
&lt;br /&gt;
Landauer, T.K., and Dumais, S.T. (1997). [http://lsa.colorado.edu/papers/plato/plato.annote.html A solution to Plato&#039;s problem: The latent semantic analysis theory of the acquisition, induction, and representation of knowledge]. &#039;&#039;Psychological Review&#039;&#039;, 104(2):211–240.&lt;br /&gt;
&lt;br /&gt;
Leacock, C., and Chodorow, M. (1998). [http://books.google.ca/books?id=Rehu8OOzMIMC&amp;amp;lpg=PA265&amp;amp;ots=IpnaLkZUec&amp;amp;lr&amp;amp;pg=PA265#v=onepage&amp;amp;q&amp;amp;f=false Combining local context and WordNet similarity for word sense identification]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, pp. 265-283.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://www.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. &#039;&#039;Proceedings of the 15th International Conference on Machine Learning (ICML-98)&#039;&#039;, Madison, WI, pp. 296-304.&lt;br /&gt;
&lt;br /&gt;
Matveeva, I., Levow, G., Farahat, A., and Royer, C. (2005). [http://people.cs.uchicago.edu/~matveeva/SynGLSA_ranlp_final.pdf Generalized latent semantic analysis for term representation]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-05)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Pado, S., and Lapata, M. (2007). [http://www.nlpado.de/~sebastian/pub/papers/cl07_pado.pdf Dependency-based construction of semantic space models]. &#039;&#039;Computational Linguistics&#039;&#039;, 33(2), 161-199.&lt;br /&gt;
&lt;br /&gt;
Pennington, J., Socher, R., and Manning, C. (2014). [https://www.aclweb.org/anthology/D14-1162 Glove: Global vectors for word representation. &#039;&#039;EMNLP 2014&#039;&#039;, pp. 1532-1543.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens D., and Navigli R. (2013). [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, disambiguate and walk: A unified approach for measuring semantic similarity]. &#039;&#039;Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013),&#039;&#039; Sofia, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Rapp, R. (2003). [http://www.amtaweb.org/summit/MTSummit/FinalPapers/19-Rapp-final.pdf Word sense discovery based on sense descriptor dissimilarity]. &#039;&#039;Proceedings of the Ninth Machine Translation Summit&#039;&#039;, pp. 315-322.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://citeseer.ist.psu.edu/resnik95using.html Using information content to evaluate semantic similarity]. &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI-95)&#039;&#039;, Montreal, pp. 448-453.&lt;br /&gt;
&lt;br /&gt;
Ruiz-Casado, M., Alfonseca, E. and Castells, P. (2005) [http://alfonseca.org/pubs/2005-ranlp1.pdf Using context-window overlapping in Synonym Discovery and Ontology Extension]. &#039;&#039;Proceedings of the International Conference Recent Advances in Natural Language Processing (RANLP-2005)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Salle A., Idiart M., and Villavicencio A. (2018) [https://github.com/alexandres/lexvec/blob/master/README.md LexVec]&lt;br /&gt;
&lt;br /&gt;
Speer, R., Chin, J., and Havasi, C. (2017). [https://www.aaai.org/ocs/index.php/AAAI/AAAI17/paper/download/14972/14051 Conceptnet 5.5: An open multilingual graph of general knowledge]. &#039;&#039;AAAI-17&#039;&#039;, pp. 4444-4451.&lt;br /&gt;
&lt;br /&gt;
Terra, E., and Clarke, C.L.A. (2003). [http://acl.ldc.upenn.edu/N/N03/N03-1032.pdf Frequency estimates for statistical word similarity measures]. &#039;&#039;Proceedings of the Human Language Technology and North American Chapter of Association of Computational Linguistics Conference 2003 (HLT/NAACL 2003)&#039;&#039;, pp. 244–251.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2001). [http://arxiv.org/abs/cs.LG/0212033 Mining the Web for synonyms: PMI-IR versus LSA on TOEFL]. &#039;&#039;Proceedings of the Twelfth European Conference on Machine Learning (ECML-2001)&#039;&#039;, Freiburg, Germany, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D., Littman, M.L., Bigham, J., and Shnayder, V. (2003). [http://arxiv.org/abs/cs.CL/0309035 Combining independent modules to solve multiple-choice synonym and analogy problems]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, pp. 482-489.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2008). [http://arxiv.org/abs/0809.0124 A uniform approach to analogies, synonyms, antonyms, and associations]. &#039;&#039;Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)&#039;&#039;, Manchester, UK, pp. 905-912.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11292</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11292"/>
		<updated>2015-11-02T16:29:04Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset [Resnik, 1995]&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset [Miller and Charles, 1991], which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| DC-best&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=11290</id>
		<title>WordSimilarity-353 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=WordSimilarity-353_Test_Collection_(State_of_the_art)&amp;diff=11290"/>
		<updated>2015-11-01T13:45:51Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
* [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ WordSimilarity-353 Test Collection]&lt;br /&gt;
* contains two sets of English word pairs along with human-assigned similarity judgements&lt;br /&gt;
* first set (set1) contains 153 word pairs along with their similarity scores assigned by 13 subjects&lt;br /&gt;
* second set (set2) contains 200 word pairs with similarity assessed by 16 subjects&lt;br /&gt;
* WordSimilarity-353 dataset is available [http://www.cs.technion.ac.il/~gabr/resources/data/wordsim353/ here]&lt;br /&gt;
* performance is measured by [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rank correlation coefficient]&lt;br /&gt;
* introduced by [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Finkelstein et al. (2002)]&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of increasing [http://en.wikipedia.org/wiki/Spearman_rank_correlation Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! Spearman&#039;s rho&lt;br /&gt;
! Pearson&#039;s r&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| WNE&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.305&lt;br /&gt;
| 0.271&lt;br /&gt;
|-&lt;br /&gt;
| J&amp;amp;C&lt;br /&gt;
| Jiang and Conrath 1997&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.318&lt;br /&gt;
| 0.354&lt;br /&gt;
|-&lt;br /&gt;
| L&amp;amp;C&lt;br /&gt;
| Leacock and Chodorow (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.341&lt;br /&gt;
|-&lt;br /&gt;
| H&amp;amp;S&lt;br /&gt;
| Hirst and St-Onge (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.302&lt;br /&gt;
| 0.356&lt;br /&gt;
|-&lt;br /&gt;
| Lin&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.348&lt;br /&gt;
| 0.357&lt;br /&gt;
|-&lt;br /&gt;
| Resnik&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.353&lt;br /&gt;
| 0.365&lt;br /&gt;
|-&lt;br /&gt;
| ROGET&lt;br /&gt;
| Jarmasz (2003)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.415&lt;br /&gt;
| 0.536&lt;br /&gt;
|-&lt;br /&gt;
| C&amp;amp;W&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Collobert and Weston (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.5&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| WikiRelate&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Strube and Ponzetto (2006)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.48&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.492&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer et al. (1997)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.581&lt;br /&gt;
| 0.563&lt;br /&gt;
|-&lt;br /&gt;
| simVB+simWN&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Finkelstein et al. (2002)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| N/A&lt;br /&gt;
| 0.55&lt;br /&gt;
|-&lt;br /&gt;
| SSA&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Hassan and Mihalcea (2011)&lt;br /&gt;
| Knowledge-based&lt;br /&gt;
| 0.622&lt;br /&gt;
| 0.629&lt;br /&gt;
|-&lt;br /&gt;
| HSMN+csmRNN&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Luong et al. (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.65&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Multi-prototype&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Huang et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.71&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Multi-lingual SSA&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Hassan et al. (2011)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.713&lt;br /&gt;
| 0.674&lt;br /&gt;
|-&lt;br /&gt;
| ESA&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.748&lt;br /&gt;
| 0.503&lt;br /&gt;
|-&lt;br /&gt;
| TSA&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Radinsky et al. (2011)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.80&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| CLEAR&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Halawi et al. (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
| Y&amp;amp;Q&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Yih and Qazvinian (2012)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81&lt;br /&gt;
| N/A&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in alphabetical order.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Finkelstein, Lev, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. (2002) [http://www.cs.technion.ac.il/~gabr/papers/tois_context.pdf Placing Search in Context: The Concept Revisited]. ACM Transactions on Information Systems, 20(1):116-131.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, Evgeniy, and Shaul Markovitch, [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI), Hyderabad, India, 2007.&lt;br /&gt;
&lt;br /&gt;
Halawi, Guy, Gideon Dror, Evgeniy Gabrilovich, and Yehuda Koren. (2012). [http://gabrilovich.com/publications/papers/Halawi2012LSL.pdf Large-scale learning of word relatedness with constraints]. In Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 1406-1414. ACM.&lt;br /&gt;
&lt;br /&gt;
Hassan, Samer, and Rada Mihalcea: [http://www.aaai.org/ocs/index.php/AAAI/AAAI11/paper/download/3616/3972/ Semantic Relatedness Using Salient Semantic Analysis]. AAAI 2011&lt;br /&gt;
&lt;br /&gt;
Hirst, Graeme and David St-Onge. Lexical chains as representations of context for the detection and correction of malapropisms. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 305–332, 1998.&lt;br /&gt;
&lt;br /&gt;
Huang, Eric H., Richard Socher, Christopher D. Manning, and Andrew Y. Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers - Volume 1 (ACL &#039;12), Vol. 1. Association for Computational Linguistics, Stroudsburg, PA, USA, 873-882.&lt;br /&gt;
&lt;br /&gt;
Islam, A., and Inkpen, D. 2006. [http://www.site.uottawa.ca/~mdislam/publications/LREC_06_242.pdf Second order co-occurrence pmi for determining the semantic similarity of words]. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2006) 1033–1038.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M. 2003. [http://www.arxiv.org/pdf/1204.0140 Roget’s thesaurus as a Lexical Resource for Natural Language Processing]. Ph.D. Dissertation, Ottawa Carleton Institute for Computer Science, School of Information Technology and Engineering, University of Ottawa.&lt;br /&gt;
&lt;br /&gt;
Jiang, Jay J. and David W. Conrath. Semantic similarity based on corpus statistics and lexical taxonomy. In Proceedings of International Conference on Research in Computational Linguistics (ROCLING X), Taiwan, pages 19–33, 1997.&lt;br /&gt;
&lt;br /&gt;
Landauer, T. K.; L, T. K.; Laham, D.; Rehder, B.; and Schreiner, M. E. 1997. How well can passage meaning be derived without using word order? a comparison of latent semantic analysis and humans.&lt;br /&gt;
&lt;br /&gt;
Leacock, Claudia and Martin Chodorow. Combining local context and WordNet similarity for word sense identification. In Christiane Fellbaum, editor, WordNet: An Electronic Lexical Database. The MIT Press, Cambridge, MA, pages 265–283, 1998.&lt;br /&gt;
&lt;br /&gt;
Lin, Dekang. An information-theoretic definition of similarity. In Proceedings of the 15th International Conference on Machine Learning, Madison,WI, pages 296–304, 1998.&lt;br /&gt;
&lt;br /&gt;
Luong, Minh-Thang, Richard Socher, and Christopher D. Manning. (2013). [http://nlp.stanford.edu/~lmthang/data/papers/conll13_morpho.pdf Better word representations with recursive neural networks for morphology]. CoNLL-2013: 104.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., D. Jurgens and R. Navigli. [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity]. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013), Sofia, Bulgaria, August 4-9, 2013, pp. 1341-1351.&lt;br /&gt;
&lt;br /&gt;
Radinsky, Kira, Eugene Agichtein, Evgeniy Gabrilovich, and Shaul Markovitch. (2011). [http://gabrilovich.com/publications/papers/Radinsky2011WTS.pdf A word at a time: computing word relatedness using temporal semantic analysis]. In Proceedings of the 20th international conference on World wide web, pp. 337-346. ACM.&lt;br /&gt;
&lt;br /&gt;
Resnik, Philip. Using information content to evaluate semantic similarity. In Proceedings of the 14th International Joint Conference on Artificial Intelligence, pages 448–453, Montreal, Canada, 1995.&lt;br /&gt;
&lt;br /&gt;
Strube, Michael and Simone Paolo Ponzetto. (2006). [http://www.aaai.org/Papers/AAAI/2006/AAAI06-223.pdf WikiRelate! Computing Semantic Relatedness Using Wikipedia]. Proceedings of The 21st National Conference on Artificial Intelligence (AAAI), Boston, MA.&lt;br /&gt;
&lt;br /&gt;
Yih, W. and Qazvinian, V. (2012). [http://aclweb.org/anthology/N/N12/N12-1077.pdf Measuring Word Relatedness Using Heterogeneous Vector Space Models]. Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT 2012).&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11289</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11289"/>
		<updated>2015-11-01T13:25:59Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset [Resnik, 1995]&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset [Miller and Charles, 1991], which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| bnc-bagofwords-num-cos-qw+enwiki-parsed-num-cos-freq&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=State_of_the_art&amp;diff=11288</id>
		<title>State of the art</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=State_of_the_art&amp;diff=11288"/>
		<updated>2015-11-01T13:23:17Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The purpose of this section of the ACL wiki is to be a repository of &#039;&#039;k&#039;&#039;-best state-of-the-art results (i.e., methods and software) for various core natural language processing tasks. &lt;br /&gt;
&lt;br /&gt;
As a side effect, this should hopefully evolve into a knowledge base of standard evaluation methods and datasets for various tasks, as well as encourage more effort into reproducibility of results. This will help newcomers to a field appreciate what has been done so far and what the main tasks are, and will help keep active researchers informed on fields other than their specific research. The next time you need a system for PP attachment, or wonder what is the current state of word sense disambiguation, this will be the place to visit. &lt;br /&gt;
&lt;br /&gt;
Please contribute! (This is also a good place for you to display your results!)&lt;br /&gt;
&lt;br /&gt;
As a historical point of reference, you may want to refer to the [http://web.archive.org/web/20100325144600/http://cslu.cse.ogi.edu/HLTsurvey/ Survey of the State of the Art in Human Language Technology] ([http://www.lt-world.org/hlt_survey/master.pdf also available as PDF]), edited by R. Cole, J. Mariani, H. Uszkoreit, G. B. Varile, A. Zaenen, A. Zampolli, V. Zue, 1996.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- Please keep this list in alphabetical order --&amp;gt;&lt;br /&gt;
* [[Anaphora Resolution (State of the art)|Anaphora Resolution]] (stub)&lt;br /&gt;
* [[Automatic Text Summarization (State of the art)|Automatic Text Summarization]] (stub)&lt;br /&gt;
* [[Chunking (State of the art)|Chunking]] (stub)&lt;br /&gt;
* [[Dependency Parsing (State of the art)|Dependency Parsing]] (stub)&lt;br /&gt;
* [[Document Classification (State of the art)|Document Classification]] (stub)&lt;br /&gt;
* [[Language Identification (State of the art)|Language Identification]] (stub)&lt;br /&gt;
* [[Named Entity Recognition (State of the art)|Named Entity Recognition]]&lt;br /&gt;
* [[Noun-Modifier Semantic Relations (State of the art)|Noun-Modifier Semantic Relations]]&lt;br /&gt;
* [[NP Chunking (State of the art)|NP Chunking]] &lt;br /&gt;
* [[Paraphrase Identification (State of the art)|Paraphrase Identification]]&lt;br /&gt;
* [[Parsing (State of the art)|Parsing]] &lt;br /&gt;
* [[POS Induction (State of the art) |POS Induction]]&lt;br /&gt;
* [[POS Tagging (State of the art) |POS Tagging]]&lt;br /&gt;
* [[PP Attachment (State of the art)|PP Attachment]] (stub)&lt;br /&gt;
* [[Question Answering (State of the art)|Question Answering]]&lt;br /&gt;
* [[Semantic Role Labeling (State of the art)|Semantic Role Labeling]] (stub)&lt;br /&gt;
* [[Sentiment Analysis (State of the art)|Sentiment Analysis]] (stub)&lt;br /&gt;
* [[Similarity (State of the art)|Similarity]] -- [[ESL Synonym Questions (State of the art)|ESL]], [[SAT Analogy Questions (State of the art)|SAT]], [[TOEFL Synonym Questions (State of the art)|TOEFL]], [[RG-65 Test Collection (State of the art)|RG-65 Test Collection]], [[MC-28 Test Collection (State of the art)|MC-28 Test Collection]], [[WordSimilarity-353 Test Collection (State of the art)|WordSimilarity-353]], [[SemEval-2012 Task 2 (State of the art)|SemEval-2012 Task 2]]&lt;br /&gt;
* [[Speech Recognition (State of the art)|Speech Recognition]] (article request)&lt;br /&gt;
* [[Temporal Information Extraction (State of the art)|Temporal Information Extraction]]&lt;br /&gt;
* [[Cleaneval (State of the art)| Web Corpus Cleaning]] (stub)&lt;br /&gt;
* [[Word Segmentation (State of the art)|Word Segmentation]] (stub)&lt;br /&gt;
* [[Word Sense Disambiguation (State of the art)|Word Sense Disambiguation]] (stub)&lt;br /&gt;
&amp;lt;!-- Please keep this list in alphabetical order --&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11287</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11287"/>
		<updated>2015-10-31T21:44:28Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó]&lt;br /&gt;
&lt;br /&gt;
PhD Student&lt;br /&gt;
&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
2013/2014 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2013/2014 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2011/2012 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Databases tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Introduction to informatics tutorials&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	PhD in Computer Science&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
	Research topic: Automatic interpretation of English and Hungarian noun&lt;br /&gt;
	compounds&lt;br /&gt;
&lt;br /&gt;
2009-2010	[http://solo.bodleian.ox.ac.uk/primo_library/libweb/action/dlDisplay.do?vid=OXVU1&amp;amp;docId=oxfaleph017405947&amp;amp;fn=permalink Master of Science in Computer Science]&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	[http://diploma.bibl.u-szeged.hu/3370/ Bachelor of Science in Computer Program Designer]&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
2010	English, level C2, Cambridge ESOL&lt;br /&gt;
&lt;br /&gt;
2005	German, level B2, Goethe Institut&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://link.springer.com/chapter/10.1007/978-3-319-13817-6_32 Information Extraction from Hungarian, English and German CVs for a Career Portal]. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. Springer International Publishing, Switzerland (2014) 333-341&lt;br /&gt;
&lt;br /&gt;
2.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: [http://rgai.inf.u-szeged.hu/mszny2014/MSZNY2014_press_b5.pdf#page=369 Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz]. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2014) 359-360&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. Springer-Verlag, Berlin Heidelberg (2013) 491-502&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=221 Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 213-224&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A., Pulman, S.G.: [http://www.inf.u-szeged.hu/projectdirs/mszny2013/images/stories/kepek/MSZNY2013_press.pdf#page=43 Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével]. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 35-46&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A., Pulman, S.G.: [http://journal.sepln.org/index.php/pln/article/view/842 Interpreting noun compounds using paraphrases]. Procesamiento del Lenguaje Natural, Vol. 46 (2011) 59-66&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11286</id>
		<title>MC-28 Test Collection (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=MC-28_Test_Collection_(State_of_the_art)&amp;diff=11286"/>
		<updated>2015-10-31T20:02:41Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: Created page with &amp;quot;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset * 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset, which is a subset of the RG-65 Test Collection...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* state of the art in Miller &amp;amp; Charles 28 (MC-28) dataset&lt;br /&gt;
* 28 word pairs of the original Miller &amp;amp; Charles 30 (MC-30) dataset, which is a subset of the [[RG-65 Test Collection (State of the art)|Rubenstein &amp;amp; Goodenough (RG-65) dataset]]; two word pairs have generally been omitted for semantic similarity evaluation, as words in these word pairs have not been included in previous versions of WordNet&lt;br /&gt;
* Similarity of each pair is scored according to a scale from 0 to 4 (the higher the &amp;quot;similarity of meaning,&amp;quot; the higher the number);&lt;br /&gt;
* The similarity values in the dataset are the means of judgments made by 38 subjects [Miller and Charles, 1991].&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed in order of decreasing [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman&#039;s rho].&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for reported results&lt;br /&gt;
! Type&lt;br /&gt;
! [http://en.wikipedia.org/wiki/Spearman%27s_rank_correlation_coefficient Spearman correlation] [with 95% confidence intervals]&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Human upper bound&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Human&lt;br /&gt;
| 0.934 [0.861, 0.969]&lt;br /&gt;
|-&lt;br /&gt;
| PPR&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.92 [0.833, 0.962]&lt;br /&gt;
|-&lt;br /&gt;
| Gloss Vector&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Patwardhan and Pedersen (2006)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.91 [0.813, 0.957]&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.87 [0.736, 0.938]&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 0.856 [0.710, 0.931]&lt;br /&gt;
|-&lt;br /&gt;
| KC&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Kulkarni and Caragea (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.835 [0.671, 0.921]&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.82 [0.644, 0.913]&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 0.81 [0.627, 0.908]&lt;br /&gt;
|-&lt;br /&gt;
| bnc-bagofwords-num-cos-qw+enwiki-parsed-num-cos-freq&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.773 [0.562, 0.889]&lt;br /&gt;
|-&lt;br /&gt;
| GM&lt;br /&gt;
| Gabrilovich and Markovitch (2007)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 0.72 [0.475, 0.861]&lt;br /&gt;
|-&lt;br /&gt;
| WLM&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Milne and Witten (2008)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.70 [0.443, 0.850]&lt;br /&gt;
|-&lt;br /&gt;
| SH&lt;br /&gt;
| Sahami and Heilman (2006)&lt;br /&gt;
| Agirre et al. (2009)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 0.618 [0.319, 0.805]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Listed alphabetically.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Agirre, E., Alfonseca, E., Hall, K., Kravalova, J., Paşca, M., Soroa, A. (2009). [http://www.aclweb.org/anthology/N09-1003 A Study on Similarity and Relatedness Using Distributional and WordNet-based Approaches]. In: &#039;&#039;10th Annual Conference of the North American Chapter of the Association for Computational Linguistics - Human Language Technologies&#039;&#039;. Association for Computa-tional Linguistics, Stroudsburg. pp. 19–27.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Gabrilovich, E., and Markovitch, S. (2007). [http://www.cs.technion.ac.il/~gabr/papers/ijcai-2007-sim.pdf Computing Semantic Relatedness using Wikipedia-based Explicit Semantic Analysis], &#039;&#039;Proceedings of The 20th International Joint Conference on Artificial Intelligence (IJCAI)&#039;&#039;, Hyderabad, India.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Kulkarni, S., Caragea, D. (2009). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.157.4693&amp;amp;rep=rep1&amp;amp;type=pdf Computation of the Semantic Relatedness between Words using Concept Clouds]. In: &#039;&#039;International Conference on Knowledge Discovery and Information Re-trieval&#039;&#039;. INSTICC Press, Setubal. pp. 183–188.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://webdocs.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. In &#039;&#039;Proceedings of the 15th International Conference on Machine Learning&#039;&#039;, Madison,WI, pp. 296–304.&lt;br /&gt;
&lt;br /&gt;
Miller, G., and Charles, W. (1991). [http://www.tandfonline.com/doi/abs/10.1080/01690969108406936#.VjUdmjZRGUk Contextual correlates of semantic similarity]. &#039;&#039;Language and Cognitive Processes&#039;&#039;, 6(1), 1–28.&lt;br /&gt;
&lt;br /&gt;
Milne, D., Witten, I.H. (2008). [http://www.aaai.org/Papers/Workshops/2008/WS-08-15/WS08-15-005.pdf An Effective, Low-Cost Measure of Semantic Relatedness Obtained from Wikipedia Links], In &#039;&#039;Proceeding of AAAI Workshop on Wikipedia and Artificial Intelligence: an Evolving Synergy&#039;&#039;, AAAI Press, Chicago, USA pp. 25-30.&lt;br /&gt;
&lt;br /&gt;
Patwardhan, S., and Pedersen, T. (2006). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6642&amp;amp;rep=rep1&amp;amp;type=pdf#page=7 Using WordNet-based Context Vectors to Estimate the Se-mantic Relatedness of Concepts]. In: &#039;&#039;11th Conference of the European Chapter of the Association for Computational Linguistics&#039;&#039;. Association for Computational Linguistics, Stroudsbur. pp. 1–8.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://arxiv.org/pdf/cmp-lg/9511007 Using information content to evaluate semantic similarity]. In &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence&#039;&#039;, Montreal, Canada, pages 448–453.&lt;br /&gt;
&lt;br /&gt;
Sahami, M., Heilman, T.D. (2006). [http://robotics.stanford.edu/users/sahami/papers-dir/www2006.pdf A web-based kernel function for measuring the similarity of short text snippets]. In: &#039;&#039;15th international conference on World Wide Web&#039;&#039;. ACM Press, New York. pp. 377–386.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=Similarity_(State_of_the_art)&amp;diff=11285</id>
		<title>Similarity (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=Similarity_(State_of_the_art)&amp;diff=11285"/>
		<updated>2015-10-31T19:55:42Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: /* Attributional similarity */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* see also: [[State of the art]]&lt;br /&gt;
&lt;br /&gt;
== Attributional similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;attributional similarity:&#039;&#039;&#039; the degree to which two words are synonymous&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[TOEFL Synonym Questions (State of the art)|TOEFL Synonym Questions]]&lt;br /&gt;
** [[ESL Synonym Questions (State of the art)|ESL Synonym Questions]]&lt;br /&gt;
** [[RG-65 Test Collection (State of the art)|RG-65 Test Collection]]&lt;br /&gt;
** [[MC-28 Test Collection (State of the art)|MC-28 Test Collection]]&lt;br /&gt;
** [[SimLex-999 (State of the art)|SimLex-999 Similarity Test Collection]]&lt;br /&gt;
** [[WordSimilarity-353 Test Collection (State of the art)|WordSimilarity-353 Test Collection]]&lt;br /&gt;
&lt;br /&gt;
== Similarity versus Association ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;similarity versus association&#039;&#039;&#039;: the contrast between taxonomical similarity (co-hyponymy) and association (co-occurrence)&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Similar-Associated-Both Test Collection (State of the art)|Similar-Associated-Both Test Collection]]&lt;br /&gt;
** [[SimLex-999 (State of the art)|SimLex-999 Similarity Test Collection]]&lt;br /&gt;
&lt;br /&gt;
== Relational similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;relational similarity:&#039;&#039;&#039; the degree to which two relations are analogous&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[SAT Analogy Questions (State of the art)|SAT Analogy Questions]]&lt;br /&gt;
** [[SemEval-2012 Task 2 (State of the art)|SemEval-2012 Task 2: Measuring Degrees of Relational Similarity]]&lt;br /&gt;
** [[Syntactic Analogies|Microsoft Research Syntactic Analogies Dataset]]&lt;br /&gt;
&lt;br /&gt;
== Phrase similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;phrase similarity:&#039;&#039;&#039; the degree to which two phrases are similar&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Noun-Modifier Questions (State of the art)|Noun-Modifier Questions]]&lt;br /&gt;
&lt;br /&gt;
== Sentence similarity ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;sentence similarity:&#039;&#039;&#039; sentence paraphrase, paraphrase identification, paraphrase recognition&lt;br /&gt;
* state-of-the-art results for:&lt;br /&gt;
** [[Paraphrase Identification (State of the art)|Microsoft Research Paraphrase Corpus]]&lt;br /&gt;
&lt;br /&gt;
== External links ==&lt;br /&gt;
&lt;br /&gt;
* SemEval-2012 Task 2: [https://sites.google.com/site/semeval2012task2/ Measuring Degrees of Relational Similarity]&lt;br /&gt;
* SemEval-2012 Task 6: [http://www.cs.york.ac.uk/semeval-2012/task6/ Semantic Textual Similarity]&lt;br /&gt;
* SEM 2013 Shared Task: [http://ixa2.si.ehu.es/sts/ Semantic Textual Similarity]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11277</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11277"/>
		<updated>2015-10-31T09:25:48Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[http://www.inf.u-szeged.hu/~dobo/ András Dobó]&lt;br /&gt;
&lt;br /&gt;
PhD Student&lt;br /&gt;
&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 [http://www.inf.u-szeged.hu/~dobo/ www.inf.u-szeged.hu/~dobo]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
2013/2014 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2013/2014 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2011/2012 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Databases tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Introduction to informatics tutorials&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	PhD in Computer Science&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
	Research topic: Automatic interpretation of English and Hungarian noun&lt;br /&gt;
	compounds&lt;br /&gt;
&lt;br /&gt;
2009-2010	Master of Science in Computer Science&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	Bachelor of Science in Computer Program Designer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
2010	English, level C2, Cambridge ESOL&lt;br /&gt;
&lt;br /&gt;
2005	German, level B2, Goethe Institut&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: Information Extraction from Hungarian, English and German CVs for a Career Portal. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. Springer International Publishing, Switzerland (2014) 333-341&lt;br /&gt;
&lt;br /&gt;
2.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2014) 359-360&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: Computing semantic similarity using large static corpora. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. Springer-Verlag, Berlin Heidelberg (2013) 491-502&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 213-224&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A., Pulman, S.G.: Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 35-46&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A., Pulman, S.G.: Interpreting noun compounds using paraphrases. Procesamiento del Lenguaje Natural, Vol. 46 (2011) 59-66&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11276</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11276"/>
		<updated>2015-10-30T22:22:43Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;András Dobó&lt;br /&gt;
&lt;br /&gt;
PhD Student&lt;br /&gt;
&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
Web:	 www.inf.u-szeged.hu/~dobo&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
2013/2014 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2013/2014 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2011/2012 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Databases tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Introduction to informatics tutorials&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	PhD in Computer Science&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
	Research topic: Automatic interpretation of English and Hungarian noun&lt;br /&gt;
	compounds&lt;br /&gt;
&lt;br /&gt;
2009-2010	Master of Science in Computer Science&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	Bachelor of Science in Computer Program Designer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
2010	English, level C2, Cambridge ESOL&lt;br /&gt;
&lt;br /&gt;
2005	German, level B2, Goethe Institut&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: Information Extraction from Hungarian, English and German CVs for a Career Portal. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. Springer International Publishing, Switzerland (2014) 333-341&lt;br /&gt;
&lt;br /&gt;
2.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2014) 359-360&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: Computing semantic similarity using large static corpora. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. Springer-Verlag, Berlin Heidelberg (2013) 491-502&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 213-224&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A., Pulman, S.G.: Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 35-46&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A., Pulman, S.G.: Interpreting noun compounds using paraphrases. Procesamiento del Lenguaje Natural, Vol. 46 (2011) 59-66&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=11272</id>
		<title>TOEFL Synonym Questions (State of the art)</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=TOEFL_Synonym_Questions_(State_of_the_art)&amp;diff=11272"/>
		<updated>2015-10-29T20:48:39Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;* TOEFL = Test of English as a Foreign Language&lt;br /&gt;
* 80 multiple-choice synonym questions; 4 choices per question&lt;br /&gt;
* the TOEFL questions are available on request by contacting [http://lsa.colorado.edu/mail_sub.html LSA Support at CU Boulder], the people who manage the [http://lsa.colorado.edu/ LSA web site at Colorado]&lt;br /&gt;
* introduced in Landauer and Dumais (1997) as a way of evaluating algorithms for measuring degree of similarity between words&lt;br /&gt;
* subsequently used by many other researchers&lt;br /&gt;
* see also: [[Similarity (State of the art)]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sample question ==&lt;br /&gt;
&lt;br /&gt;
::{| border=&amp;quot;0&amp;quot; cellpadding=&amp;quot;1&amp;quot; cellspacing=&amp;quot;1&amp;quot; &lt;br /&gt;
|-&lt;br /&gt;
! Stem:&lt;br /&gt;
|&lt;br /&gt;
| levied&lt;br /&gt;
|-&lt;br /&gt;
! Choices:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (b)&lt;br /&gt;
| believed&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (c)&lt;br /&gt;
| requested&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
| (d)&lt;br /&gt;
| correlated&lt;br /&gt;
|-&lt;br /&gt;
! Solution:&lt;br /&gt;
| (a)&lt;br /&gt;
| imposed&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Table of results ==&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellpadding=&amp;quot;5&amp;quot; cellspacing=&amp;quot;1&amp;quot; width=&amp;quot;100%&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! Algorithm&lt;br /&gt;
! Reference for algorithm&lt;br /&gt;
! Reference for experiment&lt;br /&gt;
! Type&lt;br /&gt;
! Correct&lt;br /&gt;
! 95% confidence&lt;br /&gt;
|-&lt;br /&gt;
| RES&lt;br /&gt;
| Resnik (1995)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 20.31%&lt;br /&gt;
| 12.89–31.83%&lt;br /&gt;
|-&lt;br /&gt;
| LC&lt;br /&gt;
| Leacock and Chodrow (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 21.88%&lt;br /&gt;
| 13.91–33.21%&lt;br /&gt;
|-&lt;br /&gt;
| LIN&lt;br /&gt;
| Lin (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 24.06%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| Random&lt;br /&gt;
| Random guessing&lt;br /&gt;
| 1 / 4 = 25.00%&lt;br /&gt;
| Random&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| JC&lt;br /&gt;
| Jiang and Conrath (1997)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 25.00%&lt;br /&gt;
| 15.99–35.94%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 64.38%&lt;br /&gt;
| 52.90–74.80%&lt;br /&gt;
|-&lt;br /&gt;
| Human&lt;br /&gt;
| Average non-English US college applicant&lt;br /&gt;
| Landauer and Dumais (1997)&lt;br /&gt;
| Human&lt;br /&gt;
| 64.50%&lt;br /&gt;
| 53.01–74.88%&lt;br /&gt;
|-&lt;br /&gt;
| RI&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Karlgren and Sahlgren (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 72.50%&lt;br /&gt;
| 61.38-81.90%&lt;br /&gt;
|-&lt;br /&gt;
| DS&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Pado and Lapata (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.00%&lt;br /&gt;
| 62.72-82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Turney (2001)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 73.75%&lt;br /&gt;
| 62.72–82.96%&lt;br /&gt;
|-&lt;br /&gt;
| PairClass&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Turney (2008)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 76.25%&lt;br /&gt;
| 65.42-85.06%&lt;br /&gt;
|-&lt;br /&gt;
| HSO&lt;br /&gt;
| Hirst and St.-Onge (1998)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 77.91%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| JS&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Jarmasz and Szpakowicz (2003)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 78.75%&lt;br /&gt;
| 68.17–87.11%&lt;br /&gt;
|-&lt;br /&gt;
| PMI-IR&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Terra and Clarke (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| LC-IR&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Higgins (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 81.25%&lt;br /&gt;
| 70.97–89.11%&lt;br /&gt;
|-&lt;br /&gt;
| CWO&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Ruiz-Casado et al. (2005)&lt;br /&gt;
| Web-based&lt;br /&gt;
| 82.55%&lt;br /&gt;
| 72.38–90.09%&lt;br /&gt;
|-&lt;br /&gt;
| PPMIC&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Bullinaria and Levy (2007)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 85.00%&lt;br /&gt;
| 75.26-92.00%&lt;br /&gt;
|-&lt;br /&gt;
| GLSA&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Matveeva et al. (2005)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 86.25%&lt;br /&gt;
| 76.73-92.93%&lt;br /&gt;
|-&lt;br /&gt;
| SR&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Tsatsaronis et al. (2010)&lt;br /&gt;
| Lexicon-based&lt;br /&gt;
| 87.50%&lt;br /&gt;
| 78.21-93.84%&lt;br /&gt;
|-&lt;br /&gt;
| bnc-parsed-num-cos-loglh+enwiki-parsed-num-cos-pmi&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Dobó and Csirik (2013)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 88.75%&lt;br /&gt;
| 79.72-94.72%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Rapp (2003)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 92.50%&lt;br /&gt;
| 84.39-97.20%&lt;br /&gt;
|-&lt;br /&gt;
| LSA&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Han (2014)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 95.0%&lt;br /&gt;
| 87.69-98.62%&lt;br /&gt;
|-&lt;br /&gt;
| ADW&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| Pilehvar et al. (2013)&lt;br /&gt;
| WordNet graph-based (unsupervised)&lt;br /&gt;
| 96.25%&lt;br /&gt;
| 89.43-99.22%&lt;br /&gt;
|-&lt;br /&gt;
| PR&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Turney et al. (2003)&lt;br /&gt;
| Hybrid&lt;br /&gt;
| 97.50%&lt;br /&gt;
| 91.26–99.70%&lt;br /&gt;
|-&lt;br /&gt;
| PCCP&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Bullinaria and Levy (2012)&lt;br /&gt;
| Corpus-based&lt;br /&gt;
| 100.00%&lt;br /&gt;
| 96.32-100.00%&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Explanation of table ==&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Algorithm&#039;&#039;&#039; = name of algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for algorithm&#039;&#039;&#039; = where to find out more about given algorithm&lt;br /&gt;
* &#039;&#039;&#039;Reference for experiment&#039;&#039;&#039; = where to find out more about evaluation of given algorithm with TOEFL questions&lt;br /&gt;
* &#039;&#039;&#039;Type&#039;&#039;&#039; = general type of algorithm: corpus-based, lexicon-based, hybrid&lt;br /&gt;
* &#039;&#039;&#039;Correct&#039;&#039;&#039; = percent of 80 questions that given algorithm answered correctly&lt;br /&gt;
* &#039;&#039;&#039;95% confidence&#039;&#039;&#039; = confidence interval calculated using the [[Statistical calculators|Binomial Exact Test]]&lt;br /&gt;
* table rows sorted in order of increasing percent correct&lt;br /&gt;
* several WordNet-based similarity measures are implemented in [http://www.d.umn.edu/~tpederse/ Ted Pedersen]&#039;s [http://www.d.umn.edu/~tpederse/similarity.html WordNet::Similarity] package&lt;br /&gt;
* LSA = Latent Semantic Analysis&lt;br /&gt;
* PCCP = Principal Component vectors with Caron P&lt;br /&gt;
* PMI-IR = Pointwise Mutual Information - Information Retrieval&lt;br /&gt;
* PR = Product Rule&lt;br /&gt;
* PPMIC = Positive Pointwise Mutual Information with Cosine&lt;br /&gt;
* GLSA = Generalized Latent Semantic Analysis&lt;br /&gt;
* CWO = Context Window Overlapping&lt;br /&gt;
* DS = Dependency Space&lt;br /&gt;
* RI = Random Indexing&lt;br /&gt;
&lt;br /&gt;
== Notes ==&lt;br /&gt;
&lt;br /&gt;
* the performance of a corpus-based algorithm depends on the corpus, so the difference in performance between two corpus-based systems may be due to the different corpora, rather than the different algorithms&lt;br /&gt;
* the TOEFL questions include nouns, verbs, and adjectives, but some of the WordNet-based algorithms were only designed to work with nouns; this explains some of the lower scores&lt;br /&gt;
* some of the algorithms may have been tuned on the TOEFL questions; read the references for details&lt;br /&gt;
* Landauer and Dumais (1997) report scores that were corrected for guessing by subtracting a penalty of 1/3 for each incorrect answer; they report a score of 52.5% when this penalty is applied; when the penalty is removed, their performance is 64.4% correct&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2007). [http://www.cs.bham.ac.uk/~jxb/PUBS/BRM.pdf Extracting semantic representations from word co-occurrence statistics: A computational study]. &#039;&#039;Behavior Research Methods&#039;&#039;, 39(3), 510-526.&lt;br /&gt;
&lt;br /&gt;
Bullinaria, J.A., and Levy, J.P. (2012). [http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.228.9582&amp;amp;rep=rep1&amp;amp;type=pdf Extracting semantic representations from word co-occurrence statistics: stop-lists, stemming, and SVD]. &#039;&#039;Behavior Research Methods&#039;&#039;,  44(3):890-907.&lt;br /&gt;
&lt;br /&gt;
Dobó, A., and Csirik, J. (2013). [http://link.springer.com/chapter/10.1007/978-3-642-35843-2_42 Computing semantic similarity using large static corpora]. In: van Emde Boas, P. et al. (eds.) &#039;&#039;SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741&#039;&#039;. Springer-Verlag, Berlin Heidelberg, pp. 491-502&lt;br /&gt;
&lt;br /&gt;
Lushan Han. (2014). [http://ebiquity.umbc.edu/paper/html/id/658/Schema-Free-Querying-of-Semantic-Data Schema Free Querying of Semantic Data], Ph.D. dissertation, University of Maryland, Baltimore County, Baltimore, MD USA.&lt;br /&gt;
&lt;br /&gt;
Higgins, D. (2005). [http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.329.1517 Which Statistics Reflect Semantics? Rethinking Synonymy and Word Similarity.] In: Kepser, S., Reis, M. (eds.) &#039;&#039;Linguistic Evidence: Empirical, Theoretical and Computational Perspectives&#039;&#039;. Mouton de Gruyter, Berlin, pp. 265–284.&lt;br /&gt;
&lt;br /&gt;
Hirst, G., and St-Onge, D. (1998). [http://mirror.eacoss.org/documentation/ITLibrary/IRIS/Data/1997/Hirst/Lexical/1997-Hirst-Lexical.pdf Lexical chains as representation of context for the detection and correction of malapropisms]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, 305-332.&lt;br /&gt;
&lt;br /&gt;
Jarmasz, M., and Szpakowicz, S. (2003). [http://www.csi.uottawa.ca/~szpak/recent_papers/TR-2003-01.pdf Roget’s thesaurus and semantic similarity], &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, September, pp. 212-219.&lt;br /&gt;
&lt;br /&gt;
Jiang, J.J., and Conrath, D.W. (1997). [http://wortschatz.uni-leipzig.de/~sbordag/aalw05/Referate/03_Assoziationen_BudanitskyResnik/Jiang_Conrath_97.pdf Semantic similarity based on corpus statistics and lexical taxonomy]. &#039;&#039;Proceedings of the International Conference on Research in Computational Linguistics&#039;&#039;, Taiwan.&lt;br /&gt;
&lt;br /&gt;
Karlgren, J. and Sahlgren, M. (2001). [http://www.sics.se/~jussi/Artiklar/2001_RWIbook/KarlgrenSahlgren2001.pdf From Words to Understanding]. In Uesaka, Y., Kanerva, P., &amp;amp; Asoh, H. (Eds.), &#039;&#039;Foundations of Real-World Intelligence&#039;&#039;, Stanford: CSLI Publications, pp. 294–308. &lt;br /&gt;
&lt;br /&gt;
Landauer, T.K., and Dumais, S.T. (1997). [http://lsa.colorado.edu/papers/plato/plato.annote.html A solution to Plato&#039;s problem: The latent semantic analysis theory of the acquisition, induction, and representation of knowledge]. &#039;&#039;Psychological Review&#039;&#039;, 104(2):211–240.&lt;br /&gt;
&lt;br /&gt;
Leacock, C., and Chodorow, M. (1998). [http://books.google.ca/books?id=Rehu8OOzMIMC&amp;amp;lpg=PA265&amp;amp;ots=IpnaLkZUec&amp;amp;lr&amp;amp;pg=PA265#v=onepage&amp;amp;q&amp;amp;f=false Combining local context and WordNet similarity for word sense identification]. In C. Fellbaum (ed.), &#039;&#039;WordNet: An Electronic Lexical Database&#039;&#039;. Cambridge: MIT Press, pp. 265-283.&lt;br /&gt;
&lt;br /&gt;
Lin, D. (1998). [http://www.cs.ualberta.ca/~lindek/papers/sim.pdf An information-theoretic definition of similarity]. &#039;&#039;Proceedings of the 15th International Conference on Machine Learning (ICML-98)&#039;&#039;, Madison, WI, pp. 296-304.&lt;br /&gt;
&lt;br /&gt;
Matveeva, I., Levow, G., Farahat, A., and Royer, C. (2005). [http://people.cs.uchicago.edu/~matveeva/SynGLSA_ranlp_final.pdf Generalized latent semantic analysis for term representation]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-05)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Pado, S., and Lapata, M. (2007). [http://www.nlpado.de/~sebastian/pub/papers/cl07_pado.pdf Dependency-based construction of semantic space models]. &#039;&#039;Computational Linguistics&#039;&#039;, 33(2), 161-199.&lt;br /&gt;
&lt;br /&gt;
Pilehvar, M.T., Jurgens D., and Navigli R. (2013). [http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2013_Pilehvar_Jurgens_Navigli.pdf Align, disambiguate and walk: A unified approach for measuring semantic similarity]. &#039;&#039;Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (ACL 2013),&#039;&#039; Sofia, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Rapp, R. (2003). [http://www.amtaweb.org/summit/MTSummit/FinalPapers/19-Rapp-final.pdf Word sense discovery based on sense descriptor dissimilarity]. &#039;&#039;Proceedings of the Ninth Machine Translation Summit&#039;&#039;, pp. 315-322.&lt;br /&gt;
&lt;br /&gt;
Resnik, P. (1995). [http://citeseer.ist.psu.edu/resnik95using.html Using information content to evaluate semantic similarity]. &#039;&#039;Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI-95)&#039;&#039;, Montreal, pp. 448-453.&lt;br /&gt;
&lt;br /&gt;
Ruiz-Casado, M., Alfonseca, E. and Castells, P. (2005) [http://alfonseca.org/pubs/2005-ranlp1.pdf Using context-window overlapping in Synonym Discovery and Ontology Extension]. &#039;&#039;Proceedings of the International Conference Recent Advances in Natural Language Processing (RANLP-2005)&#039;&#039;, Borovets, Bulgaria.&lt;br /&gt;
&lt;br /&gt;
Terra, E., and Clarke, C.L.A. (2003). [http://acl.ldc.upenn.edu/N/N03/N03-1032.pdf Frequency estimates for statistical word similarity measures]. &#039;&#039;Proceedings of the Human Language Technology and North American Chapter of Association of Computational Linguistics Conference 2003 (HLT/NAACL 2003)&#039;&#039;, pp. 244–251.&lt;br /&gt;
&lt;br /&gt;
Tsatsaronis, G., Varlamis, I., and Vazirgiannis, M. (2010). [http://arxiv.org/abs/1401.5699 Text Relatedness Based on a Word Thesaurus]. &#039;&#039;Journal of Artificial Intelligence Research&#039;&#039; 37, 1–39&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2001). [http://arxiv.org/abs/cs.LG/0212033 Mining the Web for synonyms: PMI-IR versus LSA on TOEFL]. &#039;&#039;Proceedings of the Twelfth European Conference on Machine Learning (ECML-2001)&#039;&#039;, Freiburg, Germany, pp. 491-502.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D., Littman, M.L., Bigham, J., and Shnayder, V. (2003). [http://arxiv.org/abs/cs.CL/0309035 Combining independent modules to solve multiple-choice synonym and analogy problems]. &#039;&#039;Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP-03)&#039;&#039;, Borovets, Bulgaria, pp. 482-489.&lt;br /&gt;
&lt;br /&gt;
Turney, P.D. (2008). [http://arxiv.org/abs/0809.0124 A uniform approach to analogies, synonyms, antonyms, and associations]. &#039;&#039;Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)&#039;&#039;, Manchester, UK, pp. 905-912.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:State of the art]]&lt;br /&gt;
[[Category:Similarity]]&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11270</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11270"/>
		<updated>2015-10-29T15:49:24Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;András Dobó&lt;br /&gt;
PhD Student&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 Room 222, 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
&lt;br /&gt;
2013/2014 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2013/2014 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2011/2012 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Databases tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Introduction to informatics tutorials&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	PhD in Computer Science&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
	Research topic: Automatic interpretation of English and Hungarian noun&lt;br /&gt;
	compounds&lt;br /&gt;
&lt;br /&gt;
2009-2010	Master of Science in Computer Science&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	Bachelor of Science in Computer Program Designer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
2010	English, level C2, Cambridge ESOL&lt;br /&gt;
&lt;br /&gt;
2005	German, level B2, Goethe Institut&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: Information Extraction from Hungarian, English and German CVs for a Career Portal. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. Springer International Publishing, Switzerland (2014) 333-341&lt;br /&gt;
&lt;br /&gt;
2.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2014) 359-360&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: Computing semantic similarity using large static corpora. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. Springer-Verlag, Berlin Heidelberg (2013) 491-502&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 213-224&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A., Pulman, S.G.: Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 35-46&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A., Pulman, S.G.: Interpreting noun compounds using paraphrases. Procesamiento del Lenguaje Natural, Vol. 46 (2011) 59-66&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
	<entry>
		<id>https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11269</id>
		<title>User:Doboandris</title>
		<link rel="alternate" type="text/html" href="https://www.aclweb.org/aclwiki/index.php?title=User:Doboandris&amp;diff=11269"/>
		<updated>2015-10-29T15:49:00Z</updated>

		<summary type="html">&lt;p&gt;Doboandris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;András Dobó&lt;br /&gt;
PhD Student&lt;br /&gt;
PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Contact information&lt;br /&gt;
Address: University of Szeged, Institute of Informatics&lt;br /&gt;
	 Room 222, 2 Árpád tér, Szeged, 6720, Hungary&lt;br /&gt;
Email:	 dobo@inf.u-szeged.hu&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Professional experience&lt;br /&gt;
2012-2014	Research mathematician&lt;br /&gt;
	nexum Magyarország Kft.	&lt;br /&gt;
&lt;br /&gt;
2010-2011	Software developer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary	&lt;br /&gt;
&lt;br /&gt;
2009		Software developer&lt;br /&gt;
	Biological Research Centre, Hungarian Academy of Sciences, Hungary	&lt;br /&gt;
&lt;br /&gt;
2008-2009	Software developer&lt;br /&gt;
	Szeged és Környéke Vízgazdálkodási Társulat, Szeged, Hungary&lt;br /&gt;
	&lt;br /&gt;
&lt;br /&gt;
Teaching&lt;br /&gt;
2013/2014 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2013/2014 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2012/2013 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2011/2012 I. semester	Artificial intelligence I. tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 II. semester	Formal languages tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Databases tutorials&lt;br /&gt;
&lt;br /&gt;
2008/2009 I. semester	Introduction to informatics tutorials&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Education&lt;br /&gt;
2012	Guest student (1 semester)&lt;br /&gt;
	Georg-August Universität Göttingen&lt;br /&gt;
&lt;br /&gt;
2011-	PhD in Computer Science&lt;br /&gt;
	PhD School in Computer Science, University of Szeged, Hungary&lt;br /&gt;
	Research topic: Automatic interpretation of English and Hungarian noun&lt;br /&gt;
	compounds&lt;br /&gt;
&lt;br /&gt;
2009-2010	Master of Science in Computer Science&lt;br /&gt;
	Computing Laboratory, University of Oxford, UK&lt;br /&gt;
&lt;br /&gt;
2006-2009	Bachelor of Science in Computer Program Designer&lt;br /&gt;
	Institute of Informatics, University of Szeged, Hungary&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Language exams&lt;br /&gt;
&lt;br /&gt;
2010	English, level C2, Cambridge ESOL&lt;br /&gt;
&lt;br /&gt;
2005	German, level B2, Goethe Institut&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Publications&lt;br /&gt;
&lt;br /&gt;
1.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Nagy, Á., Vincze, V. and Zsibrita, J.: Information Extraction from Hungarian, English and German CVs for a Career Portal. In: Prasath, R. et al. (eds.) Mining Intelligence and Knowledge Exploration. LNAI, Vol. 8891. Springer International Publishing, Switzerland (2014) 333-341&lt;br /&gt;
&lt;br /&gt;
2.	Farkas, R., Dobó, A., Kurai, Z., Miklós, I., Miszori, A., Nagy, Á., Vincze, V. and Zsibrita, J.: Információkinyerés magyar nyelvű önéletrajzokból a nexum Karrierportálhoz. In: Tanács, A. et al. (eds.) X. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2014) 359-360&lt;br /&gt;
&lt;br /&gt;
3.	Dobó, A., Csirik, J.: Computing semantic similarity using large static corpora. In: van Emde Boas, P. et al. (eds.) SOFSEM 2013: Theory and Practice of Computer Science. LNCS, Vol. 7741. Springer-Verlag, Berlin Heidelberg (2013) 491-502&lt;br /&gt;
&lt;br /&gt;
4.	Dobó, A., Csirik, J.: Magyar és angol szavak szemantikai hasonlóságának automatikus kiszámítása. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 213-224&lt;br /&gt;
&lt;br /&gt;
5.	Dobó, A., Pulman, S.G.: Angol nyelvű összetett főnevek értelmezése parafrázisok segítségével. In: Tanács, A. and Vincze, V. (eds.) IX. Magyar Számítógépes Nyelvészeti Konferencia. SZTE Informatikai Tanszékcsoport, Szeged (2012) 35-46&lt;br /&gt;
&lt;br /&gt;
6.	Dobó, A., Pulman, S.G.: Interpreting noun compounds using paraphrases. Procesamiento del Lenguaje Natural, Vol. 46 (2011) 59-66&lt;/div&gt;</summary>
		<author><name>Doboandris</name></author>
	</entry>
</feed>