July 15, 2010 |
09:00–10:40
|
Task description papers
09:00–09:20 |
SemEval-2010 Task 1: Coreference Resolution in Multiple Languages
Marta Recasens, Lluís Màrquez, Emili Sapena, M. Antònia Martí, Mariona Taulé, Véronique Hoste, Massimo Poesio and Yannick Versley
show abstracthide abstractThis paper presents the SemEval-2010 task on "Coreference Resolution in Multiple Languages." The goal was to evaluate and compare automatic coreference resolution systems for six different languages (Catalan, Dutch, English, German, Italian, and Spanish) in four evaluation settings and using four different metrics. Such a rich scenario had the potential to provide insight into key issues concerning coreference resolution: (i) the portability of systems across languages, (ii) the relevance of different levels of linguistic information, and (iii) the behavior of scoring metrics.
|
09:20–09:40 |
SemEval-2010 Task 2: Cross-Lingual Lexical Substitution
Rada Mihalcea, Ravi Sinha and Diana McCarthy
show abstracthide abstractIn this paper we describe the SemEval-2010 Cross-Lingual Lexical Substitution task, where given an English target word in context, participating systems had to find an alternative substitute word or phrase in Spanish. The task is based on the English Lexical Substitution task run at SemEval-2007. In this paper we provide background and motivation for the task, we describe the data annotation process and the scoring system, and present the results of the participating systems.
|
09:40–10:00 |
SemEval-2010 Task 3: Cross-Lingual Word Sense Disambiguation
Els Lefever and Véronique Hoste
show abstracthide abstractThe goal of this task is to evaluate the feasibility of multilingual WSD on a newly developed multilingual lexical sample data set. Participants were asked to automatically determine the contextually appropriate translation of a given English noun in five languages, viz. Dutch, German, Italian, Spanish and French. This paper reports on the sixteen submissions from the five different participating teams.
|
10:00–10:20 |
SemEval-2010 Task 5 : Automatic Keyphrase Extraction from Scientific Articles
Su Nam Kim, Olena Medelyan, Min-Yen Kan and Timothy Baldwin
show abstracthide abstractThis paper describes Task 5 of the Workshop on Semantic Evaluation 2010 (SemEval-2010). Systems are to automatically assign keyphrases or keywords to given scientific articles. The participating systems were evaluated by matching their extracted keyphrases against manually assigned ones. We present the overall ranking of the submitted systems and discuss our findings to suggest future directions for this task.
|
10:20–10:40 |
SemEval-2010 Task 7: Argument Selection and Coercion
James Pustejovsky, Anna Rumshisky, Alex Plotnick, Elisabetta Jezek, Olga Batiukova and Valeria Quochi
show abstracthide abstractWe describe the argument selection and coercion task for the SemEval-2010 evaluation exercise. This task involves characterizing the type of compositional operation that exists between a predicate and the arguments it selects. Specifically, the goal is to identify whether the type that a verb selects is satisfied directly by the argument, or whether the argument must change type to satisfy the verb typing. We discuss the problem in detail, describe the data preparation for the task, and analyze the results of the submissions.
|
|
10:40–11:00
|
Coffee/Tea Break
|
11:00–12:40
|
Task description papers
11:00–11:20 |
SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals
Iris Hendrickx, Su Nam Kim, Zornitsa Kozareva, Preslav Nakov, Diarmuid Ó Séaghdha, Sebastian Pado, Marco Pennacchiotti, Lorenza Romano and Stan Szpakowicz
show abstracthide abstractSemEval-2 Task 8 focuses on Multi-way classification of semantic relations between pairs of nominals. The task was designed to compare different approaches to semantic relation classification and to provide a standard testbed for future research. This paper defines the task, describes the training and test data and the process of their creation, lists the participating systems (10 teams, 28 runs), and discusses their results.
|
11:20–11:40 |
SemEval-2 Task 9: The Interpretation of Noun Compounds Using Paraphrasing Verbs and Prepositions
Cristina Butnariu, Su Nam Kim, Preslav Nakov, Diarmuid Ó Séaghdha, Stan Szpakowicz and Tony Veale
show abstracthide abstractPrevious research has shown that the meaning of many noun-noun compounds "N1 N2" can be approximated reasonably well by paraphrasing clauses of the form "N2 that ... N1", where "..." stands for a verb with or without a preposition. For example, "malaria mosquito" is a "mosquito that carries malaria". Evaluating the quality of such paraphrases is the theme of Task 9 at SemEval-2. This paper describes some background, the task definition, the process of data collection and the task results. We also venture a few general conclusions before the participating teams present their systems at the SemEval-2 workshop. There were 5 teams who submitted 7 systems.
|
11:40–12:00 |
SemEval-2010 Task 10: Linking Events and Their Participants in Discourse
Josef Ruppenhofer, Caroline Sporleder, Roser Morante, Collin Baker and Martha Palmer
show abstracthide abstractWe describe the SemEval-2010 shared task on “Linking Events and Their Participants in Discourse”. This task is an extension to the classical semantic role labeling task. While semantic role labeling is traditionally viewed as a sentenceinternal task, it is clear that local semantic argument structures also interact with each other in a larger context, e.g., by sharing references to specific discourse entities or events. In the shared task we looked at one particular aspect of cross-sentence links between argument structures, namely linking locally uninstantiated roles to their coreferents in the wider discourse context (if such co-referents exist). This task is potentially beneficial for a number of NLP applications, such as information extraction, question answering or text summarization.
|
12:00–12:20 |
SemEval-2010 Task 12: Parser Evaluation using Textual Entailments
Deniz Yuret, Aydin Han and Zehra Turgut
show abstracthide abstractParser Evaluation using Textual Entailments (PETE) is a shared task in the SemEval-2010 Evaluation Exercises on Semantic Evaluation. The task involves recognizing textual entailments based on syntactic information alone. PETE introduces a new parser evaluation scheme that is formalism independent, less prone to annotation error, and focused on semantically relevant distinctions.
|
12:20–12:40 |
SemEval-2010 Task 13: TempEval-2
Marc Verhagen, Roser Sauri, Tommaso Caselli and James Pustejovsky
show abstracthide abstractTempeval-2 comprises evaluation tasks for time expressions, events and temporal relations, the latter of which was split up in four sub tasks, motivated by the notion that smaller subtasks would make both data preparation and temporal relation extraction easier. Manually annotated data were provided for six languages: Chinese, English, French, Italian, Korean and Spanish.
|
|
12:40–14:00
|
Lunch
|
14:00–15:20
|
Task description papers
14:00–14:20 |
SemEval-2010 Task 14: Word Sense Induction & Disambiguation
Suresh Manandhar, Ioannis Klapaftis, Dmitriy Dligach and Sameer Pradhan
show abstracthide abstractThis paper presents the description and evaluation framework of SemEval-2010 Word Sense Induction & Disambiguation task, as well as the evaluation results of 26 participating systems. In this task, participants were required to induce the senses of 100 target words using a training set, and then disambiguate unseen instances of the same words using the induced senses. Systems’ answers were evaluated in: (1) an unsupervised manner by using two clustering evaluation measures, and (2) a supervised manner, i.e. in a WSD task.
|
14:20–14:40 |
SemEval-2010 Task: Japanese WSD
Manabu Okumura, Kiyoaki Shirai, Kanako Komiya and Hikaru Yokono
show abstracthide abstractAn overview of the SemEval-2 Japanese WSD task is presented. It is a lexical sample task, and word senses are defined according to a Japanese dictionary, the Iwanami Kokugo Jiten. This dictionary and a training corpus were distributed to participants. The number of target words was 50, with 22 nouns, 23 verbs, and 5 adjectives. Fifty instances of each target word were provided, consisting of a total of 2,500 instances for the evaluation. Nine systems from four organizations participated in the task.
|
14:40–15:00 |
SemEval-2010 Task 17: All-words Word Sense Disambiguation on a Specific Domain
Eneko Agirre, Oier Lopez de Lacalle, Christiane Fellbaum, Shu-Kai Hsieh, Maurizio Tesconi, Monica Monachini, Piek Vossen and Roxanne Segers
show abstracthide abstractDomain portability and adaptation of NLP components and Word Sense Disambiguation systems present new challenges. The difficulties found by supervised systems to adapt might change the way we assess the strengths and weaknesses of supervised and knowledge-based WSD systems. Unfortunately, all existing evaluation datasets for specific domains are lexical-sample corpora. This task presented all-words datasets on the environment domain for WSD in four languages (Chinese, Dutch, English, Italian). 11 teams participated, with supervised and knowledge-based systems, mainly in the English dataset. The results show that in all languages the participants where able to beat the most frequent sense heuristic as estimated from general corpora. The most successful approaches used some sort of supervision in the form of hand-tagged examples from the domain.
|
15:00–15:20 |
SemEval-2010 Task 18: Disambiguating Sentiment Ambiguous Adjectives
Yunfang Wu and Peng Jin
show abstracthide abstractSentiment ambiguous adjectives cause major difficulties for existing algorithms of sentiment analysis. We present an evaluation task designed to provide a framework for comparing different approaches in this problem. We define the task, describe the data creation, list the participating systems and discuss their results. There are 8 teams and 16 systems.
|
|
15:20–16:00
|
Coffee/Tea Break
|
16:00–17:30
|
Poster Session
101 |
RelaxCor: A Global Relaxation Labeling Approach to Coreference Resolution
Emili Sapena, Lluís Padró and Jordi Turmo
show abstracthide abstractThis paper describes the participation of RelaxCor in the Semeval-2010 task number 1: "Coreference Resolution in Multiple Languages". RelaxCor is a constraint-based graph partitioning approach to coreference resolution solved by relaxation labeling. The approach combines the strengths of groupwise classifiers and chain formation methods in one global method.
|
102 |
SUCRE: A Modular System for Coreference Resolution
Hamidreza Kobdani and Hinrich Schütze
show abstracthide abstractThis paper presents SUCRE, a new software tool for coreference resolution and its feature engineering. It is able to separately do noun, pronoun and full coreference resolution. SUCRE introduces a new approach to the feature engineering of coreference resolution based on a relational database model and a regular feature definition language. SUCRE successfully participated in SemEval-2010 Task 1 on Coreference Resolution in Multiple Languages for gold and regular closed annotation tracks of six languages. It obtained the best results in several categories, including the regular closed annotation tracks of English and German.
|
103 |
UBIU: A Language-Independent System for Coreference Resolution
Desislava Zhekova and Sandra Kübler
show abstracthide abstractWe present UBIU, a language independent system for detecting full coreference chains, composed of named entities, pronouns, and full noun phrases which makes use of memory based learning and a feature model following Rahman and Ng (2009). UBIU is evaluated on the task "Coreference Resolution in Multiple Languages" (SemEval Task 1 (Recasens et al., 2010)) in the context of the 5th International Workshop on Semantic Evaluation.
|
104 |
Corry: a System for Coreference Resolution
Olga Uryupina
show abstracthide abstractCorry is a system for coreference resolution in English. It supports both local and global (ILP) models of coreference. The backbone of the system is a family of SVM classifiers for pairs of mentions: each mention type receives its own classifier. A separate anaphoricity classifier is learned for the ILP setting. Corry relies on a rich linguistically motivated feature set, which has, however, been manually reduced to 64 features for efficiency reasons. The system uses the Stanford NLP toolkit for parsing and NE-tagging, Wordnet for semantic classes and the U.S. census data for assigning gender values to person names. Three runs have been submitted for the SemEval task 1, optimizing Corry’s performance for BLANC, MUC and CEAF. The runs differ with respect to the model (local for BLANC, global for MUC and CEAF) and the definition of mention types. Corry runs have shown the best performance level among all the system in their track for the corresponding metric.
|
105 |
BART: A Multilingual Anaphora Resolution System
Samuel Broscheit, Massimo Poesio, Simone Paolo Ponzetto, Kepa Joseba Rodriguez, Lorenza Romano, Olga Uryupina, Yannick Versley and Roberto Zanoli
show abstracthide abstractBART is a highly modular toolkit for coreference resolution that supports state-of-the-art statistical approaches to the task and enables efficient feature engineering. BART has originally been created and tested for English, but its flexible architecture ensures its portability to other languages and domains. At the SemEval task 1 on Coreference Resolution, BART runs have been submitted for German, English, and Italian. BART relies on a maximum enthropy-based classifier for pairs of mentions. A novel entity-mention approach based on Semantic Trees is at the moment only supported for English. For German and English, BART relies on Wordnet/Germanet for determining semantic classes and a list of names pre-classified for gender (extracted from Wikipedia). Mention boundaries are derived from parse trees. For Italian, mention boundaries and semantic types are provided by our mention tagger – it relies on Wikipedia and a gazetteer extracted from the ICab dataset.
|
106 |
TANL-1: Coreference Resolution by Parse Analysis and Similarity Clustering
Giuseppe Attardi, Maria Simi and Stefano Dei Rossi
show abstracthide abstractThis paper describes our submission to the Semeval 2010 task on coreference resolution in multiple languages. The system uses a binary classifier, based on Maximum Entropy, to decide whether or not there is a relationship between each pair of mentions extracted from a textual document. Mention detection is based on the analysis of the dependency parse tree.
|
107 |
FCC: Modeling Probabilities with GIZA++ for Task #2 and #3 of SemEval-2
Darnes Vilariño Ayala, Carlos Balderas Posada, David Eduardo Pinto Avendaño, Miguel Rodríguez Hernández and Saul León Silverio
show abstracthide abstractIn this paper we present a naïve approach to tackle the problem of cross-lingual WSD and cross-lingual lexical substitution which correspond to the Task #2 and #3 of the SemEval-2 competition. We used a bilingual statistical dictionary, which is calculated with Giza++ by using the EUROPARL parallel corpus, in order to calculate the probability of a source word to be translated to a target word (which is assumed to be the correct sense of the source word but in a different language). Two versions of the probabilistic model are tested: unweighted and weighted. The obtained values show that the unweighted version performs better thant the weighted one.
|
108 |
Combining Dictionaries and Contextual Information for Cross-Lingual Lexical Substitution
Wilker Aziz and Lucia Specia
show abstracthide abstractWe describe two systems participating in Semeval-2010’s Cross-Lingual Lexical Substitution task: USPwlv and WLVusp. Both systems are based on two main components: (i) a dictionary to provide a number of possible translations for each source word, and (ii) a contextual model to select the best translation according to the context where the source word occurs. These components and the way they are integrated are different in the two systems: they exploit corpus-based and linguistic resources, and supervised and unsupervised learning methods. Among the 14 participants in the subtask to identify the best translation, our systems were ranked 2nd and 4th in terms of recall, 3rd and 4th in terms of precision. used.
|
110 |
COLEPL and COLSLM: An Unsupervised WSD Approach to Multilingual Lexical Substitution, Tasks 2 and 3 SemEval 2010
Weiwei Guo and Mona Diab
show abstracthide abstractIn this paper, we present a word sense disambiguation (WSD) based system for multilingual lexical substitution. Our method depends on having a WSD system for English and an automatic word alignment method. Crucially the approach relies on having parallel corpora. For Task 2 we apply a supervised WSD system to derive the English word senses. For Task 3, we apply an unsupervised approach to the training and test data. Both of our systems that participated in Task 2 achieve a decent ranking among the participating systems. For Task 3 we achieve the highest ranking on several of the language pairs: French, German and Italian.
|
111 |
UHD: Cross-Lingual Word Sense Disambiguation Using Multilingual Co-occurrence Graphs
Carina Silberer and Simone Paolo Ponzetto
show abstracthide abstractWe describe the University of Heidelberg (UHD) system for the Cross-Lingual Word Sense Disambiguation SemEval-2010 task (CL-WSD). The system performs CL-WSD by applying graph algorithms previously developed for monolingual Word Sense Disambiguation to multilingual co-occurrence graphs. UHD has participated in the Best and out-of-five (OOF) evaluations and ranked among the most competitive systems for this task, thus indicating that graph-based approaches represent a powerful alternative for this task.
|
112 |
OWNS: Cross-lingual Word Sense Disambiguation Using Weighted Overlap Counts and Wordnet Based Similarity Measures
Lipta Mahapatra, Meera Mohan, Mitesh Khapra and Pushpak Bhattacharyya
show abstracthide abstractWe report here our work on English French Cross-lingual Word Sense Disambiguation where the task is to find the best French translation for a target English word depending on the context in which it is used. Our approach relies on identifying the nearest neighbors of the test sentence from the training data using a pairwise similarity measure. The proposed measure finds the affinity between two sentences by calculating a weighted sum of the word overlap and the semantic overlap between them. The semantic overlap is calculated using standard Wordnet Similarity measures. Once the nearest neighbors have been identified, the best translation is found by taking a majority vote over the French translations of the nearest neighbors.
|
113 |
273. Task 5. Keyphrase Extraction Based on Core Word Identification and Word Expansion
You Ouyang, Wenjie Li and Renxian Zhang
show abstracthide abstractThis paper provides a description of the Hong Kong Polytechnic University (PolyU) System that participated in the task #5 of SemEval-2, i.e., the Automatic Keyphrase Extraction from Scientific Articles task. We followed a novel framework to develop our keyphrase extraction system, motivated by differentiating the roles of the words in a keyphrase. We first identified the core words which are defined as the most essential words in the article, and then expanded the identified core words to the target keyphrases by a word expansion approach.
|
114 |
DERIUNLP: A Context Based Approach to Automatic Keyphrase Extraction
Georgeta Bordea and Paul Buitelaar
show abstracthide abstractThe DERI UNLP team participated in the SemEval 2010 Task #5 with an unsupervised system that automatically extracts keyphrases from scientific articles. Our approach does not consider only a general description of a term to select keyphrase candidates but also context information in the form of "skill types". Even though our system analysed a restricted list of candidates, our team was able to outperform baseline unsupervised and supervised approaches.
|
115 |
DFKI KeyWE: Ranking keyphrases extracted from scientific articles
Kathrin Eichler and Günter Neumann
show abstracthide abstractA central issue for making the content of a scientific document quickly accessible to a potential reader is the extraction of keyphrases, which capture the main topic of the document. Keyphrases can be extracted automatically by generating a list of keyphrase candidates, ranking these candidates, and selecting the top-ranked candidates as keyphrases. We present the KeyWE system, which uses an adapted nominal group chunker for candidate extraction and a supervised ranking algorithm based on support vector machines for ranking the extracted candidates. The system was evaluated on data provided for the SemEval 2010 Shared Task on Keyphrase Extraction.
|
116 |
Single Document Keyphrase Extraction Using Sentence Clustering and Latent Dirichlet Allocation
Claude Pasquier
show abstracthide abstractThis paper describes the design of a system for extracting keyphrases from a single document The principle of the algorithm is to cluster sentences of the documents in order to highlight parts of text that are semantically related. The clusters of sentences, that reflect the themes of the document, are then analyzed to find the main topics of the text. Finally, the most important words, or groups of words, from these topics are proposed as keyphrases. This method is evaluated on task number 5 (Automatic Keyphrase Extraction from Scientific Articles) of SemEval-2010: the 5th International Workshop on Semantic Evaluations.
|
117 |
SJTULTLAB: Chunk Based Method for Keyphrase Extraction
Letian Wang and Fang Li
show abstracthide abstractIn this paper we present a chunk based keyphrase extraction method for scientific articles. Different from most previous systems, supervised machine learning algorithms are not used in our system. Instead, document structure information is used to remove unimportant contents; Chunk extraction and filtering is used to reduce the quantity of candidates; Keywords are used to filter the candidates before generating final keyphrases. Our experimental results on test data show that the method works better than the baseline systems and is comparable with other known algorithms.
|
118 |
Likey: Unsupervised Language-independent Keyphrase Extraction
Mari-Sanna Paukkeri and Timo Honkela
show abstracthide abstractLikey is an unsupervised statistical approach for keyphrase extraction. The method is language-independent and the only language-dependent component is the reference corpus with which the documents to be analyzed are compared. In this study, we have also used another language-dependent component: an English-specific Porter stemmer as a preprocessing step. In our experiments of keyphrase extraction from scientific articles, the Likey method outperforms both supervised and unsupervised baseline methods.
|
119 |
WINGNUS: Keyphrase Extraction Utilizing Document Logical Structure
Thuy Dung Nguyen and Minh-Thang Luong
show abstracthide abstractWe present a system description of the WINGNUS team work for the SemEval-2010 task #5 Automatic Keyphrase Extraction from Scientific Articles. A key feature of our system is that it utilizes an inferred document logical structure in our candidate identification process, to limit the number of phrases in the candidate list, while maintaining its coverage of important phrases. Our top performing system achieves an F1 of 25.22% for the combined keyphrases (author and reader assigned) in the final test data. We note that method we report here is novel and orthogonal from other systems, so it can be combined with other techniques to potentially achieve higher performance.
|
120 |
KX: A flexible system for Keyphrase eXtraction
Emanuele Pianta and Sara Tonelli
show abstracthide abstractIn this paper we present KX, a system for keyphrase extraction developed at FBK-IRST, which exploits basic linguistic annotation combined with simple statistical measures to select a list of weighted keywords from a document. The system is flexible in that it offers to the user the possibility of setting parameters such as frequency thresholds for collocation extraction and indicators for keyphrase relevance, as well as it allows for domain adaptation exploiting a corpus of documents in an unsupervised way. KX is also easily adaptable to new languages in that it requires only a PoS-Tagger to derive lexical patterns. In the SemEval task 5 “Automatic Keyphrase Extraction from Scientific Articles”, KX performance achieved satisfactory results both in finding reader-assigned keywords and in the combined keywords subtask.
|
121 |
BUAP: An Unsupervised Approach to Automatic Keyphrase Extraction from Scientific Articles
Roberto Ortiz, David Pinto, Mireya Tovar and Héctor Jiménez-Salazar
show abstracthide abstractIn this paper, it is presented an unsupervised approach to automatically discover the latent keyphrases contained in scientific articles. The proposed technique is constructed on the basis of the combination of two techniques: maximal frequent sequences and pageranking. We evaluated the obtained results by using micro-averaged precision, recall and Fscores with respect to two different gold standards: 1) reader’s keyphrases, and 2) a combined set of author’s and reader’s keyphrases. The obtained results were also compared against three different baselines: one unsupervised (TF-IDF based) and two supervised (Na¨ıve Bayes and Maximum Entropy).
|
122 |
UNPMC: Naive Approach to Extract Keyphrases from Scientific Articles
Jungyeul Park, Jong Gun Lee and Béatrice Daille
show abstracthide abstractWe describe our method for extracting keyphrases from scientific articles which we participate in the shared task of SemEval-2 Evaluation Exercise. Even though general-purpose term extractors along with linguistically-motivated analysis allow us to extract elaborated morpho-syntactic variation forms of terms, a naive statistic approach proposed in this paper is very simple and quite efficient for extracting keyphrases especially from well-structured scientific articles. Based on the characteristics of keyphrases with section information, we obtain 18.34% for f-measure using top 15 candidates. We also show further improvement without any complications and we discuss this at the end of the paper.
|
123 |
SEERLAB: A System for Extracting Keyphrases from Scholarly Documents
Pucktada Treeratpituk, Pradeep Teregowda, Jian Huang and C. Lee Giles
show abstracthide abstractWe describe the SEERLAB system that participated in the SemEval 2010’s Keyphrase Extraction Task. SEERLAB utilizes the DBLP corpus for generating a set of candidate keyphrases from a document. Random Forest, a supervised ensemble classifier, is then used to select the top keyphrases from the candidate set. SEERLAB achieved a 0.24 F-score in generating the top 15 keyphrases, which places it sixth among 19 participating sys- tems. Additionally, SEERLAB performed particularly well in generating the top 5 keyphrases with an F-score that ranked third.
|
124 |
SZTERGAK : Feature Engineering for Keyphrase Extraction
Gábor Berend and Richárd Farkas
show abstracthide abstractAutomatically assigning keyphrases to documents has a great variety of applications. Here we focus on the keyphrase extraction of scientific publications and present a novel set of features for the supervised learning of keyphraseness. Although these features are intended for extracting keyphrases from scientific papers, because of their generality and robustness, they should have uses in other domains as well. With the help of these features SZTERGAK achieved top results on the SemEval-2 shared task on Automatic Keyphrase Extraction from Scientific Articles and exceeded its baseline by 10%.
|
125 |
KP-Miner: Participation in SemEval-2
Samhaa R. El-Beltagy and Ahmed Rafea
show abstracthide abstractThis paper briefly describes the KP-Miner sys-tem which is a system developed for the extraction of keyphrases from English and Arabic documents, irrespective of their nature. The paper also outlines the performance of the system in the “Automatic Keyphrase Extraction from Scientific Articles” task which is part of SemEval-2.
|
126 |
UvT: The UvT Term Extraction System in the Keyphrase Extraction task
Kalliopi Zervanou
show abstracthide abstractThe UvT system is based on a hybrid, linguistic and statistical approach, originally proposed for the recognition of multi-word terminological phrases, the C-value method (Frantzi et al., 2000). In the UvT implementation, we use an extended noun phrase rule set and take into consideration orthographic and morphological variation, term abbreviations and acronyms, and basic document structure information.
|
127 |
UNITN: Part-Of-Speech Counting in Relation Extraction
Fabio Celli
show abstracthide abstractThis report describes the UNITN system, a Part-Of-Speech Context Counter, that participated at Semeval 2010 Task 8: Multi- Way Classification of Semantic Relations Between Pairs of Nominals. Given a text annotated with Part-of-Speech, the system outputs a vector representation of a sentence containing 20 features in total. There are three steps in the system’s pipeline: first the system produces an estimation of the entities’ position in the relation, then an estimation of the semantic relation type by means of decision trees and finally it gives a predicition of semantic relation plus entities’ position. The system obtained good results in the estimation of entities’ position (F1=98.3%) but a critically poor performance in relation classification (F1=26.6%), indicating that lexical and semantic information is essential in relation extraction. The system can be used as an integration for other systems or for purposes different from relation extraction.
|
128 |
FBK_NK: a WordNet-based System for Multi-Way Classification of Semantic Relations
Matteo Negri and Milen Kouylekov
show abstracthide abstractWe describe a WordNet-based system for the extraction of semantic relations between pairs of nominals appearing in English texts. The system adopts a lightweight approach, based on training a Bayesian Network classifier using large sets of binary features. Our features consider: i) the context surrounding the nominals involved in the relation, and ii) different types of knowledge extracted from WordNet, including direct and explicit relations between the annotated nominals, and more general and implicit evidence (e.g. semantic boundary collocations). The system achieved a Macro-averaged F1 of 68.02% on the “Multi-Way Classification of Semantic Relations Between Pairs of Nominals” task (Task #8) at SemEval-2010.
|
129 |
JU: A Supervised Approach to Identify Semantic Relations from Paired Nominals
Santanu Pal, Partha Pakray, Dipankar Das and Sivaji Bandyopadhyay
show abstracthide abstractThis article presents the experiments carried out at Jadavpur University as part of the participation in Multi-Way Classification of Semantic Relations between Pairs of Nomi-nals in the SemEval 2010 exercise. Separate rules for each type of the relations are iden-tified in the baseline model based on the verbs and prepositions present in the seg-ment between each pair of nominals. Inclu-sion of WordNet features associated with the paired nominals play an important role in distinguishing the relations from each other. The Conditional Random Field (CRF) based machine-learning framework is adopted for classifying the pair of nominals. Application of dependency relations, Named Entities (NE) and various types of WordNet features along with several com-binations of these features help to improve the performance of the system. Error analy-sis suggests that the performance can be im-proved by applying suitable strategies to differentiate each paired nominal in an al-ready identified relation. Evaluation result gives an overall macro-averaged F1 score of 52.16%.
|
131 |
FBK-IRST: Semantic Relation Extraction using Cyc
Kateryna Tymoshenko and Claudio Giuliano
show abstracthide abstractWe present an approach for semantic relation extraction between nominals that combines semantic information with shallow syntactic processing. We propose to use the ResearchCyc knowledge base as a source of semantic information about nominals. Each kind of information is represented by kernel functions. The experiments were carried out using support vector machines as a classifier. The system achieves an overall F1 of 77.62 on the "Multi-Way Classification of Semantic Relations Between Pairs of Nominals" task at SemEval-2010.
|
132 |
ISTI@SemEval-2 Task #8: Boosting-Based Multiway Relation Classification
Andrea Esuli, Diego Marcheggiani and Fabrizio Sebastiani
show abstracthide abstractWe describe a boosting-based supervised learning approach to the “Multi-Way Classification of Semantic Relations between Pairs of Nominals” task #8 of SemEval-2. Participants were asked to determine which relation, from a set of nine relations plus “Other”, exists between two nominals, and also to determine the roles of the two nominals in the relation. Our participation has focused, rather than on the choice of a rich set of features, on the classification model adopted to determine the correct assignment of relation and roles.
|
133 |
ISI: Automatic Classification of Relations Between Nominals Using a Maximum Entropy Classifier
Stephen Tratz and Eduard Hovy
show abstracthide abstractThe automatic interpretation of semantic relations between nominals is an important subproblem within natural language understanding applications and is an area of increasing interest. In this paper, we present the system we used to participate in the SemEval 2010 Task 8 Multi-Way Classification of Semantic Relations between Pairs of Nominals. Our system, based upon a Maximum Entropy classifier trained using a large number of boolean features, received the third highest score.
|
134 |
ECNU: Effective Semantic Relations Classification without Complicated Features or Multiple External Corpora
Yuan Chen, Man Lan, Jian Su, Zhi Min Zhou and Yu Xu
show abstracthide abstractThis paper describes our approach to the automatic identification of semantic relations between nominals in English sentences. The basic idea of our strategy is to develop machine-learning classifiers which:(1) make use of class-independent features and classifier; (2) make use of a simple and effective feature set without high computational cost; (3) make no use of external annotated or unannotated corpus at all. At SemEval 2010 Task 8 our system achieved an F-measure of 75.43% and an accuracy of 70.22%.
|
135 |
UCD-Goggle: A Hybrid System for Noun Compound Paraphrasing
Guofu Li, Alejandra Lopez-Fernandez and Tony Veale
show abstracthide abstractThis paper addresses the problem of ranking a list of paraphrases associated with a noun-noun compound as closely as possible to the judgments of human raters. UCD-Goggle tackles this task using semantic knowledge learnt from the Google n-grams together with human-preferences for paraphrases mined from training data. Empirical evaluation shows that UCD-Goggle achieves 0.432 Spearman correlation with human judgments.
|
136 |
UCD-PN: Selecting General Paraphrases Using Conditional Probability
Paul Nulty and Fintan Costello
show abstracthide abstractWe describe a system which ranks human-provided paraphrases of noun compounds, where the frequency with which a given paraphrase was provided by human volunteers is the gold standard for ranking. Our system assigns a score to a paraphrase of a given compound according to the number of times it has co-occurred with other paraphrases given in the rest of the dataset. We use these co-occurrence statistics to compute conditional probabilities which cluster together paraphrases which have similar meanings and also favour frequent, general paraphrases rather than infrequent paraphrases with more specific meanings.
|
|
July 16, 2010 |
09:00–10:30
|
System papers
09:00–09:15 |
COLEPL and COLSLM: An Unsupervised WSD Approach to Multilingual Lexical Substitution, Tasks 2 and 3 SemEval 2010
Weiwei Guo and Mona Diab
show abstracthide abstractIn this paper, we present a word sense disambiguation (WSD) based system for multilingual lexical substitution. Our method depends on having a WSD system for English and an automatic word alignment method. Crucially the approach relies on having parallel corpora. For Task 2 we apply a supervised WSD system to derive the English word senses. For Task 3, we apply an unsupervised approach to the training and test data. Both of our systems that participated in Task 2 achieve a decent ranking among the participating systems. For Task 3 we achieve the highest ranking on several of the language pairs: French, German and Italian.
|
09:15–09:30 |
UBA: Using Automatic Translation and Wikipedia for Cross-Lingual Lexical Substitution
Pierpaolo Basile and Giovanni Semeraro
show abstracthide abstractThis paper presents the participation of the University of Bari (UBA) at the SemEval-2010 Cross-Lingual Lexical Substitution Task. The goal of the task is to substitute a word in a language Ls, which occurs in a particular context, by providing the best synonyms in a different language Lt which fit in that context. This task has a strict relation with the task of automatic machine translation, but there are some differences: Cross-lingual lexical substitution targets one word at a time and the main goal is to find as many good translations as possible for the given target word. Moreover, there are some connections with Word Sense Disambiguation (WSD) algorithms. Indeed, understanding the meaning of the target word is necessary to find the best substitutions. An important aspect of this kind of task is the possibility of finding synonyms without using a particular sense inventory or a specific parallel corpus, thus allowing the participation of unsupervised approaches. UBA proposes two systems: the former is based on an automatic translation system which exploits Google Translator, the latter is based on a parallel corpus approach which relies on Wikipedia in order to find the best substitutions.
|
09:30–09:45 |
HUMB: Automatic Key Term Extraction from Scientific Articles in GROBID
Patrice Lopez and Laurent Romary
show abstracthide abstractThe Semeval task 5 was an opportunity for experimenting with the key term extraction module of GROBID, a system for extracting and generating bibliographical information from technical and scientific documents. The tool first uses GROBID’s facilities for analyzing the structure of scientific articles, resulting in a first set of structural features. A second set of features captures content properties based on phraseness, informativeness and keywordness measures. Two knowledge bases, GRISP and Wikipedia, are then exploited for producing a last set of lexical/semantic features. Bagged decision trees appeared to be the most efficient machine learning algorithm for generating a list of ranked key term candidates. Finally a post ranking was realized based on statistics of co-usage of keywords in HAL, a large Open Access publication repository.
|
09:45–10:00 |
UTDMet: Combining WordNet and Corpus Data for Argument Coercion Detection
Kirk Roberts and Sanda Harabagiu
show abstracthide abstractThis paper describes our system for the classification of argument coercion for SemEval-2010 Task 7. We present two approaches to classifying an argument’s semantic class, which is then compared to the predicate’s expected semantic class to detect coercions. The first approach is based on learning the members of an arbitrary semantic class using WordNet’s hypernymy structure. The second approach leverages automatically extracted semantic parse information from a large corpus to identify similar arguments by the predicates that select them. We show the results these approaches obtain on the task as well as how they can improve a traditional feature-based approach.
|
10:00–10:15 |
UTD: Classifying Semantic Relations by Combining Lexical and Semantic Resources
Bryan Rink and Sanda Harabagiu
show abstracthide abstractThis paper describes our system for SemEval-2010 Task 8 on multi-way classification of semantic relations between nominals. First, the type of semantic relation is classified. Then a relation type-specific classifier determines the relation direction. Classification is performed using SVM classifiers and a number of features that capture the context, semantic role affiliation, and possible pre-existing relations of the nominals. This approach achieved an F1 score of 82.19% and an accuracy of 77.92%.
|
10:15–10:30 |
UvT: Memory-based pairwise ranking of paraphrasing verbs
Sander Wubben
show abstracthide abstractIn this paper we describe Mephisto, our system for Task 9 of the SemEval-2 workshop. Our approach to this task is to develop a machine learning classifier which determines for each verb pair describing a noun compound which verb should be ranked higher. These classifications are then combined into one ranking. Our classifier uses features from the Google N-gram Corpus, WordNet and the provided training data.
|
|
10:40–11:00
|
Coffee/Tea Break
|
11:00–12:30
|
System papers
11:00–11:15 |
SEMAFOR: Frame Argument Resolution with Log-Linear Models
Desai Chen, Nathan Schneider, Dipanjan Das and Noah A. Smith
show abstracthide abstractThis paper describes the SEMAFOR system’s performance in the SemEval 2010 task on linking events and their participants in discourse. Our entry is based upon SEMAFOR 1.0 (Das et al., 2010), a frame-semantic probabilistic parser built from log-linear models. The extended system models null instantiations, including non-local argument reference. Performance is evaluated on the task data with and without gold-standard overt arguments. In both settings, it fares the best of the submitted systems with respect to recall and F1.
|
11:15–11:30 |
Cambridge: Parser Evaluation using Textual Entailment by Grammatical Relation Comparison
Laura Rimell and Stephen Clark
show abstracthide abstractThis paper describes the Cambridge submission to the SemEval-2010 Parser Evaluation using Textual Entailment (PETE) task. We used a simple definition of entailment, parsing both T and H with the C&C parser and checking whether the core grammatical relations (subject and object) produced for H were a subset of those for T. This simple system achieved the top score for the task out of those systems submitted. We analyze the errors made by the system and the potential role of the task in parser evaluation.
|
11:30–11:45 |
MARS: A Specialized RTE System for Parser Evaluation
Rui Wang and Yi Zhang
show abstracthide abstractThis paper describes our participation in the the SemEval-2010 Task #12, Parser Evaluation using Textual Entailment. Our system incorporated two dependency parsers, one semantic role labeler, and a deep parser based on hand-crafted grammars. The shortest path algorithm is applied on the graph representation of the parser outputs. Then, different types of features are extracted and the entailment recognition is casted into a machine-learning-based classification task. The best setting of the system achieves 66.78% of accuracy, which ranks the 3rd place.
|
11:45–12:00 |
TRIPS and TRIOS System for TempEval-2: Extracting Temporal Information from Text
Naushad UzZaman and James Allen
show abstracthide abstractExtracting temporal information from raw text is fundamental for deep language understanding, and key to many applications like question answering, information extraction, and document summarization. In this paper, we describe two systems we submitted to the TempEval 2 challenge, for extracting temporal information from raw text. The systems use a combination of deep semantic parsing, Markov Logic Networks and Conditional Random Field classifiers. Our two submitted systems, TRIPS and TRIOS, approached all tasks and outperformed all teams in two tasks. Furthermore, TRIOS mostly had second-best performances in other tasks. TRIOS also outperformed the other teams that attempted all the tasks. Our system is notable in that for tasks C – F, they operated on raw text while all other systems used tagged events and temporal expressions in the corpus as input.
|
12:00–12:15 |
TIPSem (English and Spanish): Evaluating CRFs and Semantic Roles in TempEval-2
Hector Llorens, Estela Saquete Boro and Borja Navarro
show abstracthide abstractThis paper presents TIPSem, a system to extract temporal information from natural language texts for English and Spanish. TIPSem, learns CRF models from training data. Although the used features include different language analysis levels, the approach is focused on semantic information. For Spanish, TIPSem achieved the best F1 score in all the tasks. For English, it obtained the best F1 in tasks B (events) and D (event-dct links); and was among the best systems in the rest.
|
12:15–12:30 |
CityU-DAC: Disambiguating Sentiment-Ambiguous Adjectives within Context
Bin Lu and Benjamin K. Tsou
show abstracthide abstractThis paper describes our system participating in task 18 of SemEval-2010, i.e. disambiguating Sentiment-Ambiguous Adjectives (SAAs). To disambiguating SAAs, we compare the machine learning-based and lexicon-based methods in our submissions: 1) Maximum entropy is used to train classifiers based on the annotated Chinese data from the NTCIR opinion analysis tasks, and the clause-level and sentence-level classifiers are compared; 2) For the lexicon-based method, we first classify the adjectives into two classes: intensifiers (i.e. adjectives intensifying the intensity of context) and suppressors (i.e. adjectives decreasing the intensity of context), and then use the polarity of context to get the SAAs’ contextual polarity based on a sentiment lexicon. The results show that the performance of maximum entropy is not quite high due to little training data; on the other hand, the lexicon-based method could improve the precision by considering the polarity of context.
|
|
12:30–14:00
|
Lunch
|
14:00–15:30
|
Panel
|
15:30–16:00
|
Coffee/Tea Break
|
16:00–17:30
|
Posters Session
101 |
VENSES++: Adapting a deep semantic processing system to the identification of null instantiations
Sara Tonelli and Rodolfo Delmonte
show abstracthide abstractIn this paper we present VENSES++, a system to spot null instantiations and their antecedents, if available, as required by the "NIs-only" subtask of the SemEval 2010 Task 10 "Linking events and their participants in discourse". Our application is an adaptation of VENSES, a system for semantic evaluation that has been used for RTE challenges in the last 6 years. The new version exploits three modules of VENSES, namely the lexico-semantic module, the anaphora resolution module and the semantic module, in order to represent and analyse the document information. Then, two further procedures have been added: one identifies null instantiated roles of verbal predicates, while the other deals with nominal predicates. The first is based on the valence patterns extracted for every verbal lexical unit from FrameNet v. 1.4 and from the training data. The second procedure, instead, relies on a History List created by VENSES containing all events, spatial and temporal locations and body parts found in the document. Another useful resource employed to find antecedents is ConceptNet 2.0. Even if the preliminary results are far from satisfactory, we were able to devise a robust, knowledge-based system and a general strategy for dealing with the task.
|
102 |
CLR: Linking Events and Their Participants in Discourse Using a Comprehensive FrameNet Dictionary
Ken Litkowski
show abstracthide abstractThe CL Research system for SemEval-2 Task 10 for linking events and their participants in discourse is an exploration of the use of a specially created FrameNet dictionary that cap-tures all FrameNet information about frames, lexical units, and frame-to-frame relations. This system is embedded in a specially designed interface, the Linguistic Task Analyzer. The implementation of this system was quite minimal at the time of submission, allowing only an initial completion of the role recognition and labeling task, with recall of 0.112, precision of 0.670, and F-score of 0.192. We describe the design of the system and the continuing efforts to determine how much of this task can be performed with the available lexical resources. Changes since the official submission have improved the F-score to 0.266.
|
103 |
PKU_HIT: An Event Detection System Based on Instances Expansion and Rich Syntactic Features
Shiqi Li, Peng-Yuan Liu, Tiejun Zhao, Qin Lu and Hanjing Li
show abstracthide abstractThis paper describes the PKU_HIT system on event detection in the SemEval-2010 Task. We construct three modules for the three sub-tasks of this evaluation. For target verb WSD, we build a Naïve Bayesian classifier which uses additional training instances expanded from an untagged Chinese corpus automatically. For sentence SRL and event detection, we use a feature-based machine learning method which makes combined use of both consti-tuent-based and dependency-based features. Experimental results show that the Macro Accuracy of the WSD module reaches 83.81% and F-Score of the SRL module is 55.71%.
|
104 |
372:Comparing the Benefit of Different Dependency Parsers for Textual Entailment Using Syntactic Constraints Only
Alexander Volokh and Günter Neumann
show abstracthide abstractWe compare several state of the art dependency parsers with our own parser based on a linear classification technique. Our primary goal is therefore to use syntactic information only, in order to keep the comparison of the parsers as fair as possible. We demonstrate, that despite the inferior result using the standard evaluation metrics for parsers like UAS or LAS on standard test data, our system achieves comparable results when used in an application, such as the PETE shared task. Our submission achieved the 4th position out of 19 participating systems. However, since it only uses a linear classifier it works 17-20 times faster than other state of the parsers, as for instance MaltParser or Stanford Parser.
|
105 |
SCHWA: PETE using CCG Dependencies with the C&C Parser
Dominick Ng, James W.D. Constable, Matthew Honnibal and James R. Curran
show abstracthide abstractThis paper describes the SCHWA system entered by the University of Sydney in SemEval 2010 Task 12 – Parser Evaluation using Textual Entailments (Yuret et al., 2010). Our system achieved an overall accuracy of 70% in the task evaluation. We used the C&C parser to build CCG dependency parses of the truth and hypothesis sentences. We then used partial match heuristics to determine whether the system should predict entailment. Heuristics were used because the dependencies generated by the parser are construction specific, making full compatibility unlikely. We also manually annotated the development set with CCG analyses, establishing an upper bound for our entailment system of 87%.
|
106 |
ID 392:TERSEO + T2T3 Transducer. A systems for recognizing and normalizing TIMEX3
Estela Saquete Boro
show abstracthide abstractThe system described in this paper has participated in the Tempeval 2 competition, specifically in the Task A, which aim is to determine the extent of the time expressions in a text as defined by the TimeML TIMEX3 tag, and the value of the features type and val. For this purpose, a combination of TERSEO system and the T2T3 Transducer was used. TERSEO system is able to annotate text with TIDES TIMEX2 tags, and T2T3 transducer performs the translation from this TIMEX2 tags to TIMEX3 tags.
|
107 |
HeidelTime: High Quality Rule-based Extraction and Normalization of Temporal Expressions
Jannik Strötgen and Michael Gertz
show abstracthide abstractIn this paper, we describe HeidelTime, a system for the extraction and normalization of temporal expressions. HeidelTime is a rule-based system mainly using regular expression patterns for the extraction of temporal expressions and knowledge resources as well as linguistic clues for their normalization. In the TempEval-2 challenge, HeidelTime achieved the highest F-Score (86%) for the extraction and the best results in assigning the correct value attribute, i.e., in understanding the semantics of the temporal expressions.
|
108 |
KUL: Recognition and Normalization of Temporal Expressions
Oleksandr Kolomiyets and Marie-Francine Moens
show abstracthide abstractIn this paper we describe a system for the recognition and normalization of temporal expressions (Task 13: TempEval-2, Task A). The recognition task is approached as a classification problem of sentence constituents and the normalization is implemented in a rule-based manner. One of the system features is extending positive annotations in the corpus by semantically similar words automatically obtained from a large unannotated textual corpus. The best results obtained by the system are 0.85 and 0.84 for precision and recall respectively for recognition of temporal expressions; the accuracy values of 0.91 and 0.55 were obtained for the feature values TYPE and VAL respectively.
|
109 |
UC3M system: Determining the Extent, Type and Value of Time Expressions in TempEval-2
María Teresa Vicente-Díez, Julián Moreno-Schneider and Paloma Martínez
show abstracthide abstractThis paper describes the participation of Universidad Carlos III de Madrid in Task A of the TempEval-2 evaluation. The UC3M system was originally developed for the temporal expressions recognition and normalization (TERN task) in Spanish texts, according to the TIDES standard. Current version supposes an almost-total refactoring of the earliest system. Additionally, it has been adapted to the TimeML annotation schema and a considerable effort has been done with the aim of increasing its coverage. It takes a rule-based design both in the identification and the resolution phases. It adopts an inductive approach based on the empirical study of frequency of temporal expressions in Spanish corpora. Detecting the extent of the temporal expressions the system achieved a Precision/Recall of 0.90/0.87 whereas, in determining the TYPE and VALUE of those expressions, system results were 0.91 and 0.83, respectively.
|
110 |
Edinburgh-LTG: TempEval-2 System Description
Claire Grover, Richard Tobin, Beatrice Alex and Kate Byrne
show abstracthide abstractWe describe the Edinburgh information extraction system which we are currently adapting for analysis of newspaper text as part of the SYNC3 project. Our most recent focus is geospatial and temporal grounding of entities and it has been useful to participate in TempEval-2 to measure the performance of our system and to guide further development. We took part in Tasks A and B for English.
|
111 |
USFD2: Annotating Temporal Expresions and TLINKs for TempEval-2
Leon Derczynski and Robert Gaizauskas
show abstracthide abstractWe describe the University of Sheffield system used in the TempEval-2 challenge, USFD2. The challenge requires the automatic identification of temporal entities and relations in text. USFD2 identifies and anchors temporal expressions, and also attempts two of the four temporal relation assignment tasks. A rule-based system picks out and anchors temporal expressions, and a maximum entropy classifier assigns temporal link labels, based on features that including descriptions of associated temporal signal words. USFD2 identified temporal expressions successfully, and correctly classified their type in 90% of cases. Determining the relation between an event and time expression in the same sentence was performed at 63% accuracy, the second highest score in this part of the challenge.
|
112 |
NCSU: Modeling Temporal Relations with Markov Logic and Lexical Ontology
Eun Ha, Alok Baikadi, Carlyle Licata and James Lester
show abstracthide abstractAs a participant in TempEval-2, we address the temporal relations task consisting of four related subtasks. We take a supervised machine-learning technique using Markov Logic in combination with rich lexical relations beyond basic and syntactic features. One of our two submitted systems achieved the highest score for the Task F (66% precision), untied, and the second highest score (63% precision) for the Task C, which tied with three other systems.
|
113 |
JU_CSE_TEMP: A First Step towards Evaluating Events, Time Expressions and Temporal Relations
Anup Kumar Kolya, Asif Ekbal and Sivaji Bandyopadhyay
show abstracthide abstractTemporal information extraction is a popular and interesting research field in the area of Natural Language Processing (NLP). In this paper, we report our works on TempEval-2 shared task. This is our first participation and we participated in Tasks A, B, C, D, E and F. We develop the rule-based systems for Tasks A and B, whereas the remaining tasks are based on a machine learning approach, namely Conditional Random Field (CRF). All our systems are still in their development stages, and we report the very ini-tial results. Evaluation results on the shared task English datasets yield the precision, recall and F-measure values of 55%, 17% and 26%, respec-tively for Task A and 48%, 56% and 52%, re-spectively for Task B (event recognition). The rest of tasks, namely C, D, E and F were eva-luated with a relatively simpler metric: the num-ber of correct answers divided by the number of answers. Experiments on the English datasets yield the accuracies of 63%, 80%, 56% and 56% for tasks C, D, E and F, respectively.
|
114 |
KCDC: Word Sense Induction by Using Grammatical Dependencies and Sentence Phrase Structure
Roman Kern, Markus Muhr and Michael Granitzer
show abstracthide abstractWord sense induction and discrimination (WSID) identifies the senses of an ambiguous word and assigns instances of this word to one of these senses. We have build a WSID system that exploits syntactic and semantic features based on the results of a natural language parser component. To achieve high robustness and good generalization capabilities, we designed our system to work on a restricted, but grammatically rich set of features. Based on the results of the evaluations our system provides a promising performance and robustness.
|
115 |
UoY: Graphs of Unambiguous Vertices for Word Sense Induction and Disambiguation
Ioannis Korkontzelos and Suresh Manandhar
show abstracthide abstractThis paper presents an unsupervised graph-based method for automatic word sense induction and disambiguation. The innovative part of our method is the assignment of either a word or a word pair to each vertex of the constructed graph. Word senses are induced by clustering the constructed graph. In the disambiguation stage, each induced cluster is scored according to the number of its vertices found in the context of the target word. Our system participated in SemEval-2010 word sense induction and disambiguation task.
|
116 |
HERMIT: Flexible Clustering for the SemEval-2 WSI Task
David Jurgens and Keith Stevens
show abstracthide abstractA single word may have multiple unspecified meanings in a corpus. Word sense induction aims to discover these different meanings through word use, and knowledge-poor algorithms attempt this without using external lexical resources.We propose a new method for identifying the different senses that uses a flexible clustering strategy to automatically determine the number of senses, rather than predefining it. We demonstrate the effectiveness using the SemEval-2 WSI task, achieving competitive scores on both the V-Measure and Recall metrics, depending on the parameter configuration.
|
117 |
Duluth-WSI: SenseClusters Applied to the Sense Induction Task of SemEval-2
Ted Pedersen
show abstracthide abstractThe Duluth-WSI systems in SemEval-2 built word co–occurrence matrices from the task test data to create a second order co–occurrence representation of those test instances. The senses of words were induced by clustering these instances, where the number of clusters was automatically predicted. The Duluth-Mix system was a variation of WSI that used the combination of training and test data to create the co-occurrence matrix. The Duluth-R system was a series of random baselines.
|
118 |
KSU KDD: Word Sense Induction by Clustering in Topic Space
Wesam Elshamy, Doina Caragea and William Hsu
show abstracthide abstractWe describe our language-independent unsupervised word sense induction system. This system only uses topic features to cluster different word senses in their global context topic space. Using unlabeled data, this system trains a latent Dirichlet allocation (LDA) topic model then uses it to infer the topics distribution of the test instances. By clustering these topics distributions in their topic space we cluster them into different senses. Our hypothesis is that closeness in topic space reflects similarity between different word senses. This system participated in SemEval-2 word sense induction and disambiguation task and achieved the second highest V-measure score among all other systems.
|
119 |
PengYuan@PKU: Extracting Infrequent Sense Instance with the Same N-gram Pattern for the SemEval-2010 Task 15
Peng-Yuan Liu, Shi-Wen Yu, Shui Liu and Tiejun Zhao
show abstracthide abstractThis paper describes our infrequent sense identification system participating in the SemEval-2010 task 15 on Infrequent Sense Identification for Mandarin Text to Speech Systems. The core system is a supervised system based on the ensembles of Naïve Bayesian classifiers. In order to solve the problem of unbalanced sense distribution, we intentionally extract only instances of infrequent sense with the same N-gram pattern as the complemental training data from an untagged Chinese corpus – People’s Daily of the year 2001. At the same time, we adjusted the prior probability to adapt to the distribution of the test data and tuned the smoothing coefficient to take the data sparseness into account. Official result shows that, our system ranked the first with the best Macro Accuracy 0.952. We briefly describe this system, its configuration options and the features used for this task and present some discussion of the results.
|
120 |
RALI: Automatic weighting of text window distances
Bernard Brosseau-Villeneuve, Noriko Kando and Jian-Yun Nie
show abstracthide abstractSystems using text windows to model word contexts have mostly been using fixed-sized windows and uniform weights. The window size is often selected by trial and error to maximize task results. We propose a non-supervised method for selecting weights for each window distance, effectively removing the need to limit window sizes, by maximizing the mutual generation of two sets of samples of the same word. Experiments on Semeval Word Sense Disambiguation tasks showed considerable improvements.
|
121 |
JAIST: Clustering and Classification based Approaches for Japanese WSD
Kiyoaki Shirai and Makoto Nakamura
show abstracthide abstractThis paper reports about our three participating systems in SemEval-2 Japanese WSD task. The first one is a clustering based method, which chooses a sense for, not individual instances, but automatically constructed clusters of instances. The second one is a classification method, which is an ordinary SVM classifier with simple domain adaptation techniques. The last is an ensemble of these two systems. Results of the formal run shows the second system is the best. Its precision is 0.7476.
|
122 |
MSS: Investigating the Effectiveness of Domain Combinations and Topic Features for Word Sense Disambiguation
Sanae Fujita, Kevin Duh, Akinori Fujino, Hirotoshi Taira and Hiroyuki Shindo
show abstracthide abstractWe participated in the SemEval-2010 Japanese Word Sense Disambiguation (WSD) task (Task 16). Our focus was on (1) investigating domain differences, (2) incorporating topic features, (3) predicting new unknown senses. We experimented with Support Vector Machines (SVM) and Maximum Entropy (MEM) classifiers. We achieved an accuracy of 80.1 % in our experiments.
|
123 |
IIITH: Domain Specific Word Sense Disambiguation
Siva Reddy, Abhilash Inumella, Diana McCarthy and Mark Stevenson
show abstracthide abstractWe describe two systems that participated in SemEval-2010 task 17 (All-words Word Sense Disambiguation on a Specific Domain) and were ranked in the third and fourth positions in the formal evaluation. Domain adaptation techniques using the background documents released in the task were used to assign ranking scores to the words and their senses. The test data was disambiguated using the Personalised PageRank algorithm which was applied to a graph constructed from the whole of WordNet in which nodes are initialised with ranking scores of words and their senses. Our systems achieved comparable accuracy of 53.4 and 52.2, which outperforms the most frequent sense baseline (50.5)
|
124 |
UCF-WS: Domain Word Sense Disambiguation using Web Selectors
Hansen A. Schwartz and Fernando Gomez
show abstracthide abstractThis paper studies the application of the Web Selectors word sense disambiguation system on a specific domain. The system was primarily applied without any domain tuning, but the incorporation of domain predominant sense information was explored. Results indicated that the system performs relatively the same with domain predominant sense information as without, scoring well above a random baseline, but still 5 percentage points below results of using the most frequent sense.
|
125 |
TreeMatch: A Fully Unsupervised WSD System Using Dependency Knowledge on a Specific Domain
Andrew Tran, Chris Bowes, David Brown, Ping Chen, Max Choly and Wei Ding
show abstracthide abstractWord sense disambiguation (WSD) is one of the main challenges of applications in Natural Language Processing. TreeMatch is a WSD system originally developed using data from SemEval 2007 Task 7 (Coarse-grained English All-words Task) that has been adapted for use in SemEval 2010 Task 17 (All-words Word Sense Disambiguation on a Specific Domain). The system is based on a fully unsupervised method using dependency knowledge drawn from a domain specific knowledge base that was built for this task. When evaluated on the task, the system precision performs above the First Sense Baseline.
|
126 |
GPLSI-IXA: Using Semantic Classes to Acquire Monosemous Training Examples from Domain Texts
Rubén Izquierdo, Armando Suárez and German Rigau
show abstracthide abstractThis paper summarizes our participation in task #17 of SemEval–2 (All–words WSD on a specific domain) using a supervised class-based Word Sense Disambiguation system. Basically, we use Support Vector Machines (SVM) as learning algorithm and a set of simple features to build three different models. Each model considers a different training corpus: SemCor (SC), examples from monosemous words extracted automatically from background data (BG), and both SC and BG (SCBG). Our system explodes the monosemous words appearing as members of a particular WordNet semantic class to automatically acquire class-based annotated examples from the domain text. We use the class-based examples gathered from the domain corpus to adapt our traditional system trained on SemCor. The evaluation reveal that the best results are achieved training with SemCor and the background examples from monosemous words, obtaining results above the most frequent baseline and the fifth best position in the competition rank.
|
127 |
HIT-CIR: An Unsupervised {WSD} System Based on Domain Most Frequent Sense Estimation
Yuhang Guo, Wanxiang Che, Wei He, Ting Liu and Sheng Li
show abstracthide abstractThis paper presents an unsupervised system for all-word domain specific word sense disambiguation task. This system tags target word with the most frequent sense which is estimated using a thesaurus and the word distribution information in the domain. The thesaurus is automatically constructed from bilingual parallel corpus using paraphrase technique. The recall of this system is 43.5\% on SemEval-2 task 17 English data set.
|
128 |
RACAI: Unsupervised WSD experiments @ SemEval-2, Task #17
Radu Ion and Dan Ştefănescu
show abstracthide abstractThis paper documents the participation of the Research Institute for Artificial Intelligence of the Romanian Academy (RACAI) to the Task 17 – All-words Word Sense Disambiguation on a Specific Domain, of the SemEval-2 competition. We describe three unsupervised WSD systems that make extensive use of the Princeton WordNet (WN) structure and WordNet Domains in order to perform the disambiguation. The best of them has been ranked the 12th by the task organizers out of 29 judged runs.
|
129 |
Kyoto: An Integrated System for Specific Domain WSD
Aitor Soroa, Eneko Agirre, Oier López de Lacalle, Wauter Bosma, Piek Vossen, Monica Monachini, Jessie Lo and Shu-Kai Hsieh
show abstracthide abstractThis document describes the preliminary release of the integrated Kyoto system for specific domain WSD. The system uses concept miners (Tybots) to extract domain-related terms and produces a domain-related thesaurus, followed by knowledge-based WSD based on wordnet graphs (UKB). The resulting system can be applied to any language with a lexical knowledge base, and is based on publicly available software and resources. Our participation in Semeval task #17 focused on producing running systems for all languages in the task, and we attained good results in all except Chinese. Due to the pressure of the time-constraints in the competition, the system is still under development, and we expect results to improve in the near future.
|
130 |
CFILT: Resource Conscious Approaches for All-Words Domain Specific WSD
Anup Kulkarni, Mitesh Khapra, Saurabh Sohoney and Pushpak Bhattacharyya
show abstracthide abstractWe describe two approaches for All-words Word Sense Disambiguation on a Specific Domain}. The first approach is a knowledge based approach which extracts domain-specific largest connected components from the Wordnet graph by exploiting the semantic relations between all candidate synsets appearing in a domain-specific untagged corpus. Given a test word, disambiguation is performed by considering only those candidate synsets that belong to the top-k largest connected components. The second approach is a weakly supervised approach which relies on the "One Sense Per Domain" heuristic and uses a few hand labeled examples for the most frequently appearing words in the target domain. Once the most frequent words have been disambiguated they can provide strong clues for disambiguating other words in the sentence using an iterative disambiguation algorithm. Our weakly supervised system gave the best performance across all systems that participated in the task even when it used as few as 100 hand labeled examples from the target domain.
|
131 |
UMCC-DLSI: Integrative Resource for Disambiguation Task
Yoan Gutiérrez Vázquez, Antonio Fernandez Orquín, Andrés Montoyo Guijarro and Sonia Vázquez Pérez
show abstracthide abstractThis paper describes the UMCC-DLSI system in SemEval-2010 task number 17 (All-words Word Sense Disambiguation on Specific Domain). The main purpose of this work is to evaluate and compare our computational resource of WordNet’s mappings using 3 different methods: Relevant Semantic Tree, Relevant Semantic Tree 2 and an Adaptation of k-clique’s Technique. Our proposal is a non-supervised and knowledge-based system that uses Domains Ontology and SUMO.
|
132 |
HR-WSD: System Description for All-words Word Sense Disambiguation on a Specific Domain at SemEval-2010
Meng-Hsien Shih
show abstracthide abstractThe document describes the knowledge-based Domain-WSD system using heuristic rules (knowledge-base). This HR-WSD system delivered the best performance (55.9%) among all Chinese systems in SemEval-2010 Task 17: All-words WSD on a specific domain.
|
133 |
Twitter Based System: Using Twitter for Disambiguating Sentiment Ambiguous Adjectives
Alexander Pak and Patrick Paroubek
show abstracthide abstractIn this paper, we describe our system which participated in the SemEval 2010 task of disambiguating sentiment ambiguous adjectives for Chinese. Our system uses text messages from Twitter, a popular microblogging platform, for building a dataset of emotional texts. Using the built dataset, the system classifies the meaning of adjectives into positive or negative sentiment polarity according to the given context. Our approach is fully automatic. It does not require any additional hand-built language resources and it is language independent.
|
134 |
YSC-DSAA: An Approach to Disambiguate Sentiment Ambiguous Adjectives Based On SAAOL
Shi-Cai Yang and Mei-Juan Liu
show abstracthide abstractIn this paper, we describe the system we developed for the SemEval-2010 task of Disambiguating Sentiment Ambiguous Adjectives (hereinafter referred to SAA). Our system created a new word library named SAA-Oriented Library consisting of positive words, negative words, negative words related to SAA, positive words related to SAA, and inverse words, etc. Based on the syntactic parsing, we analyzed the relationship between SAA and the keywords and handled other special processes by extracting such words in the relevant sen-tences to disambiguate sentiment ambiguous adjectives. Our micro average accuracy is 0.942, which puts our system in the first place.
|
135 |
OpAL: Applying Opinion Mining Techniques for the Disambiguation of Sentiment Ambiguous Adjectives in SemEval-2 Task 18
Alexandra Balahur and Andrés Montoyo Guijarro
show abstracthide abstractThe task of extracting the opinion expressed in text is challenging due to different reasons. One of them is that the same word (in particular, adjectives) can have different polarities depending on the context. This paper presents the experiments carried out by the OpAL team for the participation in the SemEval 2010 Task 18 – Disambiguation of Sentiment Ambiguous Adjectives. Our approach is based on three different strategies: a) the evaluation of the polarity of the whole context using an opinion mining system; b) the assessment of the polarity of the local context, given by the combinations between the closest nouns and the adjective to be classified; c) rules aiming at refining the local semantics through the spotting of modifiers. The final decision for classification is taken according to the output of the majority of these three approaches. The method used yielded good results, the OpAL system run ranking fifth among 16.
|
136 |
HITSZ_CITYU: Combine Collocation, Context Words and Neighboring Sentence Sentiment in Sentiment Adjectives Disambiguation
Ruifeng Xu, Jun Xu and Chunyu Kit
show abstracthide abstractThis paper presents the HIT_CITYU systems in Semeval-2 Task 18, namely, disambiguat-ing sentiment ambiguous adjectives. The baseline system (HITSZ_CITYU_3) incorporates bi-gram and n-gram collocations of sentiment adjectives, and other context words as features in a one-class Support Vector Machine (SVM) classifier. To enhance the baseline system, collocation set expansion and characteristics learning based on word similarity and semi-supervised learning are investigated, respectively. The final system (HITSZ_CITYU_1/2) combines collocations, context words and neighboring sentence sentiment in a two-class SVM classifier to determine the polarity of sentiment adjectives. The final systems achieved 0.957 and 0.953 (ranked 1st and 2nd) macro accuracy, and 0.936 and 0.933 (ranked 2nd and 3rd) micro accuracy, respectively.
|
137 |
SWAT: Cross-Lingual Lexical Substitution using Local Context Matching, Bilingual Dictionaries and Machine Translation
Richard Wicentowski, Maria Kelly and Rachel Lee
show abstracthide abstractWe present two systems that select the most appropriate Spanish substitutes for a marked word in an English test sentence. These systems were official entries to the SemEval-2010 Cross-Lingual Lexical Substitution task. The first system, Swat-E, finds Spanish substitutions by first finding English substitutions in the English sentence and then translating these substitutions into Spanish using an English-Spanish dictionary. The second system, Swat-S, translates each English sentence into Spanish and then finds the Spanish substitutions in the Spanish sentence. Both systems exceeded the baseline and all other participating systems by a wide margin using one of the two official scoring metrics.
|
138 |
TUD: semantic relatedness for relation classification
György Szarvas and Iryna Gurevych
show abstracthide abstractIn this paper, we describe the system submitted by the team TUD to Task 8 at SemEval 2010. The challenge focused on the identification of semantic relations between pairs of nominals in sentences collected from the web. We applied maximum entropy classification using both lexical and syntactic features to describe the nominals and their context. In addition, we experimented with features describing the semantic relatedness (SR) between the target nominals and a set of clue words characteristic to the relations. Our best submission with SR features achieved 69.23% macro-averaged F-measure, providing 8.73% improvement over our baseline system. Thus, we think SR can serve as a natural way to incorporate external knowledge to relation classification.
|
|