Research in Computational Linguistics (CL) has been growing rapidly in recent years in terms of novel scientific challenges and commercial application opportunities. This is due to the fact that a very large part of theWeb content is textual and written in many languages. A part from linguistic resources (e.g., WordNet), the research trend is moving towards the automatic extraction of semantic information from large corpora to support on-line understanding of textual data. An example of direct outcome is represented by common-sense semantic resources. The main example is ConceptNet, the final result of the Open Mind Common Sense project developed by MIT, which collected unstructured common-sense knowledge by asking people to contribute over the Web. In spite of being promising for its size and broad semantic coverage, few applications appeared in the literature so far, due to a number of issues such as inconsistency and sparseness. In this paper, we present the results of the application of this type of knowledge in two different (supervised and unsupervised) scenarios: the computation of semantic similarity (the keystone of most Computational Linguistics tasks), and the automatic identification of word meanings (Word Sense Induction) in simple syntactic structures. © Springer International Publishing Switzerland 2015.
Common-sense knowledge for natural language understanding: Experiments in unsupervised and supervised settings
DI CARO, Luigi;RUGGERI, ALICE;BOELLA, Guido
2015-01-01
Abstract
Research in Computational Linguistics (CL) has been growing rapidly in recent years in terms of novel scientific challenges and commercial application opportunities. This is due to the fact that a very large part of theWeb content is textual and written in many languages. A part from linguistic resources (e.g., WordNet), the research trend is moving towards the automatic extraction of semantic information from large corpora to support on-line understanding of textual data. An example of direct outcome is represented by common-sense semantic resources. The main example is ConceptNet, the final result of the Open Mind Common Sense project developed by MIT, which collected unstructured common-sense knowledge by asking people to contribute over the Web. In spite of being promising for its size and broad semantic coverage, few applications appeared in the literature so far, due to a number of issues such as inconsistency and sparseness. In this paper, we present the results of the application of this type of knowledge in two different (supervised and unsupervised) scenarios: the computation of semantic similarity (the keystone of most Computational Linguistics tasks), and the automatic identification of word meanings (Word Sense Induction) in simple syntactic structures. © Springer International Publishing Switzerland 2015.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.