Summary of the paper

Title A Multimodal Corpus of Expert Gaze and Behavior during Phonetic Segmentation Tasks
Authors Arif Khan, Ingmar Steiner, Yusuke Sugano, Andreas Bulling and Ross Macdonald
Abstract Phonetic segmentation is the process of splitting speech into distinct phonetic units. Human experts routinely perform this task manually by analyzing auditory and visual cues using analysis software, which is an extremely time-consuming process. Methods exist for automatic segmentation, but these are not always accurate enough. In order to improve automatic segmentation, we need to model it as close to the manual segmentation as possible. This corpus is an effort to capture the human segmentation behavior by recording experts performing a segmentation task. We believe that this data will enable us to highlight the important aspects of manual segmentation, which can be used in automatic segmentation to improve its accuracy.
Topics Phonetic Databases, Phonology, Speech Resource/Database, Usability, User Satisfaction
Full paper A Multimodal Corpus of Expert Gaze and Behavior during Phonetic Segmentation Tasks
Bibtex @InProceedings{KHAN18.676,
  author = {Arif Khan and Ingmar Steiner and Yusuke Sugano and Andreas Bulling and Ross Macdonald},
  title = "{A Multimodal Corpus of Expert Gaze and Behavior during Phonetic Segmentation Tasks}",
  booktitle = {Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)},
  year = {2018},
  month = {May 7-12, 2018},
  address = {Miyazaki, Japan},
  editor = {Nicoletta Calzolari (Conference chair) and Khalid Choukri and Christopher Cieri and Thierry Declerck and Sara Goggi and Koiti Hasida and Hitoshi Isahara and Bente Maegaard and Joseph Mariani and Hélène Mazo and Asuncion Moreno and Jan Odijk and Stelios Piperidis and Takenobu Tokunaga},
  publisher = {European Language Resources Association (ELRA)},
  isbn = {979-10-95546-00-9},
  language = {english}
  }
Powered by ELDA © 2018 ELDA/ELRA