The MMASCS multi-modal annotated synchronous corpus of audio, video, facial motion and tongue motion data of normal, fast and slow speech

by Dietmar Schabus, Michael Pucher, Phil Hoole
Abstract:
In this paper, we describe and analyze a corpus of speech data that we have recorded in multiple modalities simultaneously: facial motion via optical motion capturing, tongue motion via electro-magnetic articulography, as well as conventional video and high-quality audio. The corpus consists of 320 phonetically diverse sentences uttered by a male Austrian German speaker at normal, fast and slow speaking rate. We analyze the influence of speaking rate on phone durations and on tongue motion. Furthermore, we investigate the correlation between tongue and facial motion. The data corpus is available free of charge for research use, including phonetic annotations and a playback software which visualizes the 3D data, from the website http://cordelia.ftw.at/mmascs
Reference:
Dietmar Schabus, Michael Pucher, Phil Hoole, “The MMASCS multi-modal annotated synchronous corpus of audio, video, facial motion and tongue motion data of normal, fast and slow speech”, In Proceedings of the 9th International Conference on Language Resources and Evaluation (LREC), Reykjavik, Iceland, pp. 3411-3416, 2014.
Bibtex Entry:
@InProceedings{Schabus2014,
  Title                    = {The MMASCS multi-modal annotated synchronous corpus of audio, video, facial motion and tongue motion data of normal, fast and slow speech},
  Author                   = {Dietmar Schabus and Michael Pucher and Phil Hoole},
  Booktitle                = {Proceedings of the 9th International Conference on Language Resources and Evaluation (LREC)},
  Year                     = {2014},

  Address                  = {Reykjavik, Iceland},
  Month                    = may,
  Pages                    = {3411-3416},

  Abstract                 = {In this paper, we describe and analyze a corpus of speech data that we have recorded in multiple modalities simultaneously: facial motion via optical motion capturing, tongue motion via electro-magnetic articulography, as well as conventional video and high-quality audio. The corpus consists of 320 phonetically diverse sentences uttered by a male Austrian German speaker at normal, fast and slow speaking rate. We analyze the influence of speaking rate on phone durations and on tongue motion. Furthermore, we investigate the correlation between tongue and facial motion. The data corpus is available free of charge for research use, including phonetic annotations and a playback software which visualizes the 3D data, from the website http://cordelia.ftw.at/mmascs},
  Url                      = {http://www.lrec-conf.org/proceedings/lrec2014/summaries/192.html}
}