User profiles for Vikram Ramanarayanan
Vikram RamanarayananChief Scientist, Modality.AI & Associate Adjunct Professor, UC San Francisco Verified email at modality.ai Cited by 1953 |
Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (TC)
USC-TIMIT is an extensive database of multimodal speech production data, developed to
complement existing resources available to the speech research community and with the …
complement existing resources available to the speech research community and with the …
Analysis of pausing behavior in spontaneous speech using real-time magnetic resonance imaging of articulation
It is hypothesized that pauses at major syntactic boundaries (ie, grammatical pauses), but
not ungrammatical (eg, word search) pauses, are planned by a high-level cognitive …
not ungrammatical (eg, word search) pauses, are planned by a high-level cognitive …
Evaluating speech, face, emotion and body movement time-series features for automated multimodal presentation scoring
We analyze how fusing features obtained from different multimodal data streams such as
speech, face, body movement and emotion tracks can be applied to the scoring of multimodal …
speech, face, body movement and emotion tracks can be applied to the scoring of multimodal …
Analysis of speech production real-time MRI
Recent advances in real-time magnetic resonance imaging (RT-MRI) have made it possible
to study the anatomy and dynamic motion of the vocal tract during speech production with …
to study the anatomy and dynamic motion of the vocal tract during speech production with …
Using bidirectional LSTM recurrent neural networks to learn high-level abstractions of sequential features for automated scoring of non-native spontaneous speech
Z Yu, V Ramanarayanan… - … IEEE Workshop on …, 2015 - ieeexplore.ieee.org
We introduce a new method to grade non-native spoken language tests automatically.
Traditional automated response grading approaches use manually engineered time-aggregated …
Traditional automated response grading approaches use manually engineered time-aggregated …
[HTML][HTML] The FACTS model of speech motor control: Fusing state estimation and task-based control
B Parrell, V Ramanarayanan… - PLoS computational …, 2019 - journals.plos.org
We present a new computational model of speech motor control: the Feedback-Aware
Control of Tasks in Speech or FACTS model. FACTS employs a hierarchical state feedback …
Control of Tasks in Speech or FACTS model. FACTS employs a hierarchical state feedback …
Simulated speaking environments for language learning: Insights from three cases
Recent CALL technology reviews cover a plethora of technologies available to language
learners to improve a variety of skills, including speaking. However, few technology-enhanced …
learners to improve a variety of skills, including speaking. However, few technology-enhanced …
Speech as a biomarker: Opportunities, interpretability, and challenges
Purpose: Over the past decade, the signal processing and machine learning literature has
demonstrated notable advancements in automated speech processing with the use of artificial …
demonstrated notable advancements in automated speech processing with the use of artificial …
An investigation of articulatory setting using real-time magnetic resonance imaging
This paper presents an automatic procedure to analyze articulatory setting in speech
production using real-time magnetic resonance imaging of the moving human vocal tract. The …
production using real-time magnetic resonance imaging of the moving human vocal tract. The …
[PDF][PDF] A multimodal real-time MRI articulatory corpus for speech research
We present MRI-TIMIT: a large-scale database of synchronized audio and real-time magnetic
resonance imaging (rtMRI) data for speech research. The database currently consists of …
resonance imaging (rtMRI) data for speech research. The database currently consists of …