User profiles for Vikram Ramanarayanan

Vikram Ramanarayanan

Chief Scientist, Modality.AI & Associate Adjunct Professor, UC San Francisco
Verified email at modality.ai
Cited by 1953

Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (TC)

S Narayanan, A Toutios, V Ramanarayanan… - The Journal of the …, 2014 - pubs.aip.org
USC-TIMIT is an extensive database of multimodal speech production data, developed to
complement existing resources available to the speech research community and with the …

Analysis of pausing behavior in spontaneous speech using real-time magnetic resonance imaging of articulation

V Ramanarayanan, E Bresch, D Byrd… - The Journal of the …, 2009 - pubs.aip.org
It is hypothesized that pauses at major syntactic boundaries (ie, grammatical pauses), but
not ungrammatical (eg, word search) pauses, are planned by a high-level cognitive …

Evaluating speech, face, emotion and body movement time-series features for automated multimodal presentation scoring

V Ramanarayanan, CW Leong, L Chen… - Proceedings of the …, 2015 - dl.acm.org
We analyze how fusing features obtained from different multimodal data streams such as
speech, face, body movement and emotion tracks can be applied to the scoring of multimodal …

Analysis of speech production real-time MRI

V Ramanarayanan, S Tilsen, M Proctor, J Töger… - Computer Speech & …, 2018 - Elsevier
Recent advances in real-time magnetic resonance imaging (RT-MRI) have made it possible
to study the anatomy and dynamic motion of the vocal tract during speech production with …

Using bidirectional LSTM recurrent neural networks to learn high-level abstractions of sequential features for automated scoring of non-native spontaneous speech

Z Yu, V Ramanarayanan… - … IEEE Workshop on …, 2015 - ieeexplore.ieee.org
We introduce a new method to grade non-native spoken language tests automatically.
Traditional automated response grading approaches use manually engineered time-aggregated …

[HTML][HTML] The FACTS model of speech motor control: Fusing state estimation and task-based control

B Parrell, V Ramanarayanan… - PLoS computational …, 2019 - journals.plos.org
We present a new computational model of speech motor control: the Feedback-Aware
Control of Tasks in Speech or FACTS model. FACTS employs a hierarchical state feedback …

Simulated speaking environments for language learning: Insights from three cases

…, TFH Smits, K Evanini, V Ramanarayanan - Computer Assisted …, 2019 - Taylor & Francis
Recent CALL technology reviews cover a plethora of technologies available to language
learners to improve a variety of skills, including speaking. However, few technology-enhanced …

Speech as a biomarker: Opportunities, interpretability, and challenges

V Ramanarayanan, AC Lammert, HP Rowe… - Perspectives of the ASHA …, 2022 - ASHA
Purpose: Over the past decade, the signal processing and machine learning literature has
demonstrated notable advancements in automated speech processing with the use of artificial …

An investigation of articulatory setting using real-time magnetic resonance imaging

V Ramanarayanan, L Goldstein, D Byrd… - The Journal of the …, 2013 - pubs.aip.org
This paper presents an automatic procedure to analyze articulatory setting in speech
production using real-time magnetic resonance imaging of the moving human vocal tract. The …

[PDF][PDF] A multimodal real-time MRI articulatory corpus for speech research

…, A Lammert, M Proctor, V Ramanarayanan… - … Annual Conference of …, 2011 - sail.usc.edu
We present MRI-TIMIT: a large-scale database of synchronized audio and real-time magnetic
resonance imaging (rtMRI) data for speech research. The database currently consists of …