Difference between revisions of "SimLex-999 (State of the art)"

From ACL Wiki
Jump to navigation Jump to search
(Created page with "[http://www.cl.cam.ac.uk/~fh295/simlex.html SimLex-999] aims at a cleaner benchmark of similarity (but not relatedness). Pairs of words were chosen to represent different rang...")
 
m (Hill et al. (2014) --> 2014a & 2014b)
Line 7: Line 7:
 
| Neural language model
 
| Neural language model
 
| Collobert & Weston (2008)<ref>R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.</ref>
 
| Collobert & Weston (2008)<ref>R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.</ref>
| Hill et al. (2014)<ref name=simlex/>  
+
| Hill et al. (2014a)<ref name=simlex/>  
 
| Distributional || 0.268 || - || Trained on Wikipedia
 
| Distributional || 0.268 || - || Trained on Wikipedia
 
|-
 
|-
 
| Neural language model with global context
 
| Neural language model with global context
 
| Huang et al. (2012)<ref>Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.</ref>
 
| Huang et al. (2012)<ref>Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.</ref>
| Hill et al. (2014)<ref name=simlex/>  
+
| Hill et al. (2014a)<ref name=simlex/>  
 
| Distributional || 0.098 || - || Trained on Wikipedia
 
| Distributional || 0.098 || - || Trained on Wikipedia
 
|-
 
|-
 
| Word2vec
 
| Word2vec
 
| Mikolov et al. (2013)<ref>Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.</ref>
 
| Mikolov et al. (2013)<ref>Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.</ref>
| Hill et al. (2014)<ref name=simlex>Hill, F., Reichart, R., & Korhonen, A. (2014). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.</ref>
+
| Hill et al. (2014a)<ref name=simlex>Hill, F., Reichart, R., & Korhonen, A. (2014a). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.</ref>
 
| Distributional || 0.414 || - || Trained on Wikipedia
 
| Distributional || 0.414 || - || Trained on Wikipedia
 
|-
 
|-
Line 41: Line 41:
 
|-
 
|-
 
| RNNenc
 
| RNNenc
| Hill et al. (2014)<ref name=rnnenc>Hill, F., Cho, K., Jean, S., Devin, C., & Bengio, Y. (2014). Not All Neural Embeddings are Born Equal, 1–5.</ref>
+
| Hill et al. (2014b)<ref name=rnnenc>Hill, F., Cho, K., Jean, S., Devin, C., & Bengio, Y. (2014b). Not All Neural Embeddings are Born Equal, 1–5.</ref>
| Hill et al. (2014)<ref name=rnnenc/>
+
| Hill et al. (2014b)<ref name=rnnenc/>
 
| Distributional, multilingual || 0.52 || -
 
| Distributional, multilingual || 0.52 || -
 
|}
 
|}

Revision as of 06:58, 25 June 2015

SimLex-999 aims at a cleaner benchmark of similarity (but not relatedness). Pairs of words were chosen to represent different ranges of similarity and with either high or low association. Subjects were instructed to differentiate between similarity and relatedness and rate regarding the former only.

Algorithm Reference for algorithm Reference for reported results Type Spearman's rho Pearson's r Notes
Neural language model Collobert & Weston (2008)[1] Hill et al. (2014a)[2] Distributional 0.268 - Trained on Wikipedia
Neural language model with global context Huang et al. (2012)[3] Hill et al. (2014a)[2] Distributional 0.098 - Trained on Wikipedia
Word2vec Mikolov et al. (2013)[4] Hill et al. (2014a)[2] Distributional 0.414 - Trained on Wikipedia
Lesk Banjade et al. (2015)[5] 0.404 0.347
UMBC Han et al. (2013)[6] Banjade et al. (2015)[5] 0.558 0.557 without using POS information
SVR4 Banjade et al. (2015)[5] Banjade et al. (2015)[5] Combined 0.642 0.658
ESA Banjade et al. (2015)[5] 0.271 0.145
RNNenc Hill et al. (2014b)[7] Hill et al. (2014b)[7] Distributional, multilingual 0.52 -

References

  1. R. Collobert and J. Weston. 2008. A unified architecture for natural language pro- cessing: Deep neural networks with multitask learning. In International Conference on Machine Learn- ing, ICML.
  2. 2.0 2.1 2.2 Hill, F., Reichart, R., & Korhonen, A. (2014a). SimLex-999: Evaluating Semantic Models with (Genuine) Similarity Estimation. Computation and Language.
  3. Eric H Huang, Richard Socher, Christopher D Manning, and Andrew Y Ng. 2012. Improving word representations via global context and multiple word prototypes. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Long Papers-Volume 1, pages 873–882. Association for Computational Linguistics.
  4. Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of International Conference of Learning Representations, Scottsdale, Arizona, USA.
  5. 5.0 5.1 5.2 5.3 5.4 Banjade, R., Maharjan, N., Niraula, N., Rus, V., & Gautam, D. (2015). Lemon and Tea Are Not Similar: Measuring Word-to-Word Similarity by Combining Different Methods. Computational Linguistics and Intelligent Text Processing, 9041, 335–346. doi:10.1007/978-3-319-18111-0_25
  6. Han, L., Kashyap, A., Finin, T., Mayfield, J., Weese, J.: UMBC EBIQUITY-CORE: Semantic textual similarity systems. In: Proceedings of the Second Joint Conference on Lexical and Computational Semantics, vol. 1, pp. 44–52 (2013)
  7. 7.0 7.1 Hill, F., Cho, K., Jean, S., Devin, C., & Bengio, Y. (2014b). Not All Neural Embeddings are Born Equal, 1–5.