1. INTRODUCTION
Cloud-based ASR systems are easily available to companies building speech-based products. These products cover a wide-range of use cases like speech transcriptions, language understanding, spoken language translation, information extraction, and summarization. Most of these use-cases involve transcribing speech and then performing various downstream language-processing tasks. In these scenarios, there is a break of domain in two places, one for speech-to-text where pre-trained ASR is trained on different domains of data, and another while optimizing NLP downstream tasks with transcriptions from pre-existing ASR trained on another domain. This is a break that also stems from being unable to train in-house competitive ASR on in-domain data alone, which has a lesser chance of out performing pre-trained ASRs on much larger data, even if it is out-of-domain. Towards solving this problem, we propose to carry out ASR error correction via domain adaptation on two pre-existing ASRs: ASPIRE model [1] which is an open-source resource trained on conversational, broadcast, and read speech, and Google Speech API1 which is trained on large quantities of English speech.