Timezone: »

 
Coffee break + posters 1
Samuel Myer · Wei-Ning Hsu · Jialu Li · Monica Dinculescu · Lea Schönherr · Ehsan Hosseini-Asl · Skyler Seto · Oiwi Parker Jones · Imran Sheikh · Thomas Manzini · Yonatan Belinkov · Nadir Durrani · Alexander Amini · Johanna Hansen · Gabi Shalev · Jamin Shin · Paul Smolensky · Lisa Fan · Zining Zhu · Hamid Eghbal-zadeh · Benjamin Baer · Abelino Jimenez · Joao Felipe Santos · Jan Kremer · Erik McDermott · Andreas Krug · Tzeviya S Fuchs · Shuai Tang · Brandon Carter · David Gifford · Albert Zeyer · André Merboldt · Krishna Pillutla · Katherine Lee · Titouan Parcollet · Orhan Firat · Gautam Bhattacharya · JAHANGIR ALAM · Mirco Ravanelli

Sat Dec 08 07:30 AM -- 08:00 AM (PST) @

Jamin Shin, Andrea Madotto, Pascale Fung, "Interpreting Word Embeddings with Eigenvector Analysis" Mirco Ravanelli, Yoshua Bengio, "Interpretable Convolutional Filters with SincNet" Shuai Tang, Paul Smolensky, Virginia R. de Sa, "Learning Distributed Representations of Symbolic Structure Using Binding and Unbinding Operations" Lisa Fan, Dong Yu, Lu Wang, "Robust Neural Abstractive Summarization Systems and Evaluation against Adversarial Information" Zining Zhu, Jekaterina Novikova, Frank Rudzicz, "Semi-supervised classification by reaching consensus among modalities" Hamid Eghbal-zadeh, Matthias Dorfer, Gerhard Widmer, "Deep Within-Class Covariance Analysis for Robust Deep Audio Representation Learning" Benjamin Baer, Skyler Seto, Martin T. Wells, "Interpreting Word Embeddings with Generalized Low Rank Models" Abelino Jimenez, Benjamin Elizalde, Bhiksha Raj, "Sound event classification using ontology-based neural netowrks" Hai Pham, Paul Pu Liang, Thomas Manzini, Louis-Philippe Morency, Barnabas Poczos, "Found in Translation: Learning Robust Joint Representations by Cyclic Translations Between Modalities" Sri Harsha Dumpala, Imran Sheikh, Rupayan Chakraborty, Sunil Kumar Kopparapu, "Cycle-Consistent GAN Front-end to Improve ASR Robustness to Perturbed Speech" Joao Felipe Santos, Tiago H. Falk, "Investigating the effect of residual and highway connections in speech enhancement models" Jan Kremer, Lasse Borgholt, Lars Maaløe, "On the Inductive Bias of Word-Character-Level Multi-Task Learning for Speech Recognition" Erik McDermott, "A Deep Generative Acoustic Model for Compositional Automatic Speech Recognition" Andreas Krug, René Knaebel, Sebastian Stober, "Neuron Activation Profiles for Interpreting Convolutional Speech Recognition Models" Anthony Bau, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, James Glass, "Identifying and Controlling Important Neurons in Neural Machine Translation" Oiwi Parker Jones, Brendan Shillingford, "Composing RNNs and FSTs for Small Data: Recovering Missing Characters in Old Hawaiian Text" Tzeviya Fuchs, Joseph Keshet, "Robust Spoken Term Detection Automatically Adjusted for a Given Threshold" Shuai Tang, Virginia R. de Sa, "Improving Sentence Representations with Multi-view Frameworks" Brandon Carter, Jonas Mueller, Siddhartha Jain, David Gifford, "Local and global model interpretability via backward selection and clustering" Albert Zeyer, André Merboldt, Ralf Schlüter, Hermann Ney, "A comprehensive analysis on attention models" Barbara Rychalska, Dominika Basaj, Przemysław Biecek, "Are you tough enough? Framework for Robustness Validation of Machine Comprehension Systems" Jialu Li, Mark Hasegawa-Johnson, "A Comparable Phone Set for the TIMIT Dataset Discovered in Clustering of Listen, Attend and Spell" Loren Lugosch, Samuel Myer, Vikrant Singh Tomar, "DONUT: CTC-based Query-by-Example Keyword Spotting" Titouan Parcollet, Mirco Ravanelli, Mohamed Morchid, Georges Linarès, Renato De Mori, "Speech Recognition with Quaternion Neural Networks" Wei-Ning Hsu, Yu Zhang, Ron J. Weiss, Yu-An Chung, Yuxuan Wang, Yonghui Wu, James Glass, "Disentangling Correlated Speaker and Noise for Speech Synthesis via Data Augmentation and Adversarial Factorization" Jan Buys, Yonatan Bisk, Yejin Choi, "Bridging HMMs and RNNs through Architectural Transformations" Katherine Lee, Orhan Firat, Ashish Agarwal, Clara Fannjiang, David Sussillo, "Hallucinations in neural machine translation" Ehsan Hosseini-Asl, Yingbo Zhou, Caiming Xiong, Richard Socher, "Robust Domain Adaptation By Augmented Cyclic Adversarial Learning" Cheng-Zhi Anna Huang, Monica Dinculescu, Ashish Vaswani, Douglas Eck, "Visualizing Music Transformer" Lea Schönherr, Katharina Kohls, Steffen Zeiler, Dorothea Kolossa, Thorsten Holz, "Adversarial Attacks Against Automatic Speech Recognition Systems via Psychoacoustic Hiding" Gautam Bhattacharya, Joao Monteiro, Jahangir Alam, Patrick Kenny, "SpeakerGAN: Recognizing Speakers in New Languages with Generative Adversarial Networks" Jessica Thompson, Marc Schönwiesner, Yoshua Bengio, Daniel Willett, "How transferable are features in convolutional neural network acoustic models across languages?" Ramin M. Hasani, Alexander Amini, Mathias Lechner, Felix Naser, Radu Grosu, Daniela Rus, "Response Characterization for Auditing Cell Dynamics in Long Short-term Memory Networks"

Author Information

Samuel Myer (fluent.ai)

I am a software engineer working at Montreal-based startup fluent.ai. I am involved in research and development of low-footprint speech recognition software. We are working towards speech recognition which works in any language, respects user privacy, and is robust to noise.

Wei-Ning Hsu (Massachusetts Institute of Technology)
Jialu Li (University of Illinois at Urbana-Champaign)
Monica Dinculescu (Google)
Lea Schönherr (Ruhr University Bochum)
Ehsan Hosseini-Asl (Salesforce Research)
Skyler Seto (Cornell University)
Oiwi Parker Jones (University of Oxford)

Human and Animal Neuroscience, Deep Learning, NLP, Prosthetics.

Imran Sheikh (TCS Research and Innovation)
Thomas Manzini (Microsoft AI Development Acceleration Program)
Yonatan Belinkov (MIT)
Nadir Durrani (QCRI)
Alexander Amini (MIT)
Johanna Hansen (McGill University)
Gabi Shalev (Dept. of Computer Science, Bar-Ilan University)
Jamin Shin (HKUST CAiRE)
Paul Smolensky (Microsoft Research/Johns Hopkins)
Lisa Fan (Northeastern University)
Zining Zhu (University of Toronto)
Hamid Eghbal-zadeh (Johannes Kepler University of Linz)
Benjamin Baer (Cornell University)
Abelino Jimenez (Carnegie Mellon University)
Joao Felipe Santos (Institut National de la Recherche Scientifique)
Jan Kremer (Corti)

Jan Kremer is a senior machine learning researcher at Corti in Denmark. There he works on automatic speech recognition in critical conversations. He is particularly interested in multi-task and robust learning. After receiving an MSc in Computer Science from the Technical University of Munich in 2013, he obtained a PhD in Machine Learning from the University of Copenhagen in 2016.

Erik McDermott (Google)
Andreas Krug (University of Potsdam)
Tzeviya S Fuchs (Bar Ilan University)
Shuai Tang (CogSci, UC San Diego)
Brandon Carter (MIT CSAIL / Google)
David Gifford (MIT)
Albert Zeyer (RWTH Aachen University)
André Merboldt (RWTH Aachen University)
Krishna Pillutla (University of Washington)
Katherine Lee (Google)
Titouan Parcollet (University of Avignon)
Orhan Firat (Google)
Gautam Bhattacharya (Mcgill University)
JAHANGIR ALAM (Computer Research Institute of Montreal (CRIM))
Mirco Ravanelli (Montreal Istitute for Learning Algorithms)

I received my master's degree in Telecommunications Engineering (full marks and honours) from the University of Trento, Italy in 2011. I then joined the SHINE research group (led by Prof. Maurizio Omologo) of the Bruno Kessler Foundation (FBK), contributing to some projects on distant-talking speech recognition in noisy and reverberant environments, such as DIRHA and DOMHOS. In 2013 I was visiting researcher at the International Computer Science Institute (University of California, Berkeley) working on deep neural networks for large-vocabulary speech recognition in the context of the IARPA BABEL project (led by Prof. Nelson Morgan). I received my PhD (with cum laude distinction) in Information and Communication Technology from the University of Trento in December 2017. During my PhD I worked on “deep learning for distant speech recognition”, with a particular focus on recurrent and cooperative neural networks (see my PhD thesis here). In the context of my PhD I recently spent 6 months in the MILA lab led by Prof. Yoshua Bengio. I'm currently a post-doc researcher at the University of Montreal, working on deep learning for speech recognition in the MILA Lab.

More from the Same Authors