Το work with title Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation by Stepanov Evgeny A., Chowdhury Shammur Absar, Bayer Ali Orkan, Ghosh Arindam, Klasinas Ioannis, Calvo Marcos, Sanchís Emilio, Riccardi Giuseppe is licensed under Creative Commons Attribution 4.0 International
Bibliographic Citation
E. A. Stepanov, S. A. Chowdhury, A. O. Bayer, A. Ghosh, I. Klasinas, M. Calvo, E. Sanchis and G. Riccardi, "Cross-language transfer of semantic annotation via targeted crowdsourcing: task design and evaluation," Lang. Resour. Eval., vol. 52, no. 1, pp. 341-364, Mar. 2018. doi : 10.1007/s10579-017-9396-5
https://doi.org/10.1007/s10579-017-9396-5
Modern data-driven spoken language systems (SLS) require manual semantic annotation for training spoken language understanding parsers. Multilingual porting of SLS demands significant manual effort and language resources, as this manual annotation has to be replicated. Crowdsourcing is an accessible and cost-effective alternative to traditional methods of collecting and annotating data. The application of crowdsourcing to simple tasks has been well investigated. However, complex tasks, like cross-language semantic annotation transfer, may generate low judgment agreement and/or poor performance. The most serious issue in cross-language porting is the absence of reference annotations in the target language; thus, crowd quality control and the evaluation of the collected annotations is difficult. In this paper we investigate targeted crowdsourcing for semantic annotation transfer that delegates to crowds a complex task such as segmenting and labeling of concepts taken from a domain ontology; and evaluation using source language annotation. To test the applicability and effectiveness of the crowdsourced annotation transfer we have considered the case of close and distant language pairs: Italian–Spanish and Italian–Greek. The corpora annotated via crowdsourcing are evaluated against source and target language expert annotations. We demonstrate that the two evaluation references (source and target) highly correlate with each other; thus, drastically reduce the need for the target language reference annotations.