Search Results for author: Hervé Bourlard

Found 7 papers, 2 papers with code

Lattice-Free MMI Adaptation Of Self-Supervised Pretrained Acoustic Models

2 code implementations28 Dec 2020 Apoorv Vyas, Srikanth Madikeri, Hervé Bourlard

In this work, we propose lattice-free MMI (LFMMI) for supervised adaptation of self-supervised pretrained acoustic model.

Pkwrap: a PyTorch Package for LF-MMI Training of Acoustic Models

1 code implementation7 Oct 2020 Srikanth Madikeri, Sibo Tong, Juan Zuluaga-Gomez, Apoorv Vyas, Petr Motlicek, Hervé Bourlard

We present a simple wrapper that is useful to train acoustic models in PyTorch using Kaldi's LF-MMI training framework.

Audio and Speech Processing Sound

Neural Network based End-to-End Query by Example Spoken Term Detection

no code implementations19 Nov 2019 Dhananjay Ram, Lesly Miculicich, Hervé Bourlard

Here, we show that the CNN based matching outperforms DTW based matching using bottleneck features as well.

Dynamic Time Warping Template Matching

Multilingual Bottleneck Features for Query by Example Spoken Term Detection

no code implementations30 Jun 2019 Dhananjay Ram, Lesly Miculicich, Hervé Bourlard

State of the art solutions to query by example spoken term detection (QbE-STD) usually rely on bottleneck feature representation of the query and audio document to perform dynamic time warping (DTW) based template matching.

Dynamic Time Warping Template Matching

Information Theoretic Analysis of DNN-HMM Acoustic Modeling

no code implementations29 Aug 2017 Pranay Dighe, Afsaneh Asaei, Hervé Bourlard

We propose an information theoretic framework for quantitative assessment of acoustic modeling for hidden Markov model (HMM) based automatic speech recognition (ASR).

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

On Structured Sparsity of Phonological Posteriors for Linguistic Parsing

no code implementations21 Jan 2016 Milos Cernak, Afsaneh Asaei, Hervé Bourlard

Building on findings from converging linguistic evidence on the gestural model of Articulatory Phonology as well as the neural basis of speech perception, we hypothesize that phonological posteriors convey properties of linguistic classes at multiple time scales, and this information is embedded in their support (index) of active coefficients.

Cannot find the paper you are looking for? You can Submit a new open access paper.