Spoken Query-Answering (SQA) is a core functionality for helpful and interactive synthetic intelligence methods. Just lately, a number of speech-language fashions (SpeechLMs) have been launched with a particular concentrate on enhancing their SQA efficiency. Nonetheless, an absence of managed ablations of pretraining information processing and curation makes it difficult to grasp what elements account for efficiency, regardless of substantial positive aspects from related research in different information modalities. On this work, we tackle this hole by conducting a data-centric exploration for pretraining SpeechLMs. We concentrate on three analysis questions basic to speech-language pretraining information: (1) the way to course of uncooked web-crawled audio content material for speech-text pretraining, (2) the way to assemble artificial pretraining datasets to enhance web-crawled information and (3) the way to interleave (textual content, audio) segments into coaching sequences. We apply the insights from our managed data-centric ablations to pretrain a 3.8B-parameter SpeechLM, referred to as SpeLangy, that outperforms fashions which can be as much as 3x bigger by 10.2% absolute efficiency. We hope our findings spotlight the impression of efficient information curation for speech-language pretraining and information future data-centric exploration in SpeechLMs.
- † College of Cambridge
- ‡ College of Tübingen

