URI | http://purl.tuc.gr/dl/dias/6E738048-7889-492D-9368-05CF082D6E7D | - |
Αναγνωριστικό | https://link.springer.com/article/10.1007%2Fs10579-017-9396-5 | - |
Αναγνωριστικό | https://doi.org/10.1007/s10579-017-9396-5 | - |
Γλώσσα | en | - |
Μέγεθος | 24 pages | en |
Τίτλος | Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation | en |
Δημιουργός | Stepanov Evgeny A. | en |
Δημιουργός | Chowdhury Shammur Absar | en |
Δημιουργός | Bayer Ali Orkan | en |
Δημιουργός | Ghosh Arindam | en |
Δημιουργός | Klasinas Ioannis | en |
Δημιουργός | Κλασινας Ιωαννης | el |
Δημιουργός | Calvo Marcos | en |
Δημιουργός | Sanchís Emilio | en |
Δημιουργός | Riccardi Giuseppe | en |
Εκδότης | Springer Verlag | en |
Περίληψη | Modern data-driven spoken language systems (SLS) require manual semantic annotation for training spoken language understanding parsers. Multilingual porting of SLS demands significant manual effort and language resources, as this manual annotation has to be replicated. Crowdsourcing is an accessible and cost-effective alternative to traditional methods of collecting and annotating data. The application of crowdsourcing to simple tasks has been well investigated. However, complex tasks, like cross-language semantic annotation transfer, may generate low judgment agreement and/or poor performance. The most serious issue in cross-language porting is the absence of reference annotations in the target language; thus, crowd quality control and the evaluation of the collected annotations is difficult. In this paper we investigate targeted crowdsourcing for semantic annotation transfer that delegates to crowds a complex task such as segmenting and labeling of concepts taken from a domain ontology; and evaluation using source language annotation. To test the applicability and effectiveness of the crowdsourced annotation transfer we have considered the case of close and distant language pairs: Italian–Spanish and Italian–Greek. The corpora annotated via crowdsourcing are evaluated against source and target language expert annotations. We demonstrate that the two evaluation references (source and target) highly correlate with each other; thus, drastically reduce the need for the target language reference annotations. | en |
Τύπος | Peer-Reviewed Journal Publication | en |
Τύπος | Δημοσίευση σε Περιοδικό με Κριτές | el |
Άδεια Χρήσης | http://creativecommons.org/licenses/by/4.0/ | en |
Ημερομηνία | 2018-04-27 | - |
Ημερομηνία Δημοσίευσης | 2018 | - |
Θεματική Κατηγορία | Cross-language transfer | en |
Θεματική Κατηγορία | Crowdsourcing | en |
Θεματική Κατηγορία | Evaluation | en |
Θεματική Κατηγορία | Semantic annotation | en |
Βιβλιογραφική Αναφορά | E. A. Stepanov, S. A. Chowdhury, A. O. Bayer, A. Ghosh, I. Klasinas, M. Calvo, E. Sanchis and G. Riccardi, "Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation," Lang. Resour. Eval., vol. 52, no. 1, pp. 341-364, Mar. 2018. doi : 10.1007/s10579-017-9396-5 | en |