| Title | Corpus Annotation through Crowdsourcing: Towards Best Practice Guidelines | 
  
  | Authors | Marta Sabou, Kalina Bontcheva, Leon Derczynski and Arno Scharl | 
  
  | Abstract | Crowdsourcing is an emerging collaborative approach that can be used for the acquisition of annotated corpora and a wide range of other linguistic resources. Although the use of this approach is intensifying in all its key genres (paid-for crowdsourcing, games with a purpose, volunteering-based approaches), the community still lacks a set of best-practice guidelines similar to the annotation best practices for traditional, expert-based corpus acquisition. In this paper we focus on the use of crowdsourcing methods for corpus acquisition and propose a set of best practice guidelines based in our own experiences in this area and an overview of related literature. We also introduce GATE Crowd, a plugin of the GATE platform that relies on these guidelines and offers tool support for using crowdsourcing in a more principled and efficient manner. | 
  
  | Topics | Collaborative Resource Construction, Tools, Systems, Applications | 
  
  | Full paper  | Corpus Annotation through Crowdsourcing: Towards Best Practice Guidelines | 
  
  | Bibtex | @InProceedings{SABOU14.497, author =  {Marta Sabou and Kalina Bontcheva and Leon Derczynski and Arno Scharl},
 title =  {Corpus Annotation through Crowdsourcing: Towards Best Practice Guidelines},
 booktitle =  {Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)},
 year =  {2014},
 month =  {may},
 date =  {26-31},
 address =  {Reykjavik, Iceland},
 editor =  {Nicoletta Calzolari (Conference Chair) and Khalid Choukri and Thierry Declerck and Hrafn Loftsson and Bente Maegaard and Joseph Mariani and Asuncion Moreno and Jan Odijk and Stelios Piperidis},
 publisher =  {European Language Resources Association (ELRA)},
 isbn =  {978-2-9517408-8-4},
 language =  {english}
 }
 |