Éva Székely

Orcid: 0000-0003-1175-840X

According to our database1, Éva Székely authored at least 38 papers between 2011 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
Unified speech and gesture synthesis using flow matching.
CoRR, 2023

Matcha-TTS: A fast TTS architecture with conditional flow matching.
CoRR, 2023

On the Use of Self-Supervised Speech Representations in Spontaneous Speech Synthesis.
CoRR, 2023

Diff-TTSG: Denoising probabilistic integrated speech and gesture synthesis.
CoRR, 2023

Automatic Evaluation of Turn-taking Cues in Conversational Speech Synthesis.
CoRR, 2023

Can a gender-ambiguous voice reduce gender stereotypes in human-robot interactions?
Proceedings of the 32nd IEEE International Conference on Robot and Human Interactive Communication, 2023

Hi robot, it's not what you say, it's how you say it.
Proceedings of the 32nd IEEE International Conference on Robot and Human Interactive Communication, 2023

Generation of speech and facial animation with controllable articulatory effort for amusing conversational characters.
Proceedings of the 23rd ACM International Conference on Intelligent Virtual Agents, 2023

A Comparative Study of Self-Supervised Speech Representations in Read and Spontaneous TTS.
Proceedings of the IEEE International Conference on Acoustics, 2023

Prosody-Controllable Spontaneous TTS with Neural HMMS.
Proceedings of the IEEE International Conference on Acoustics, 2023

Why is my Agent so Slow? Deploying Human-Like Conversational Turn-Taking.
Proceedings of the International Conference on Human-Agent Interaction, 2023

Casual chatter or speaking up? Adjusting articulatory effort in generation of speech and animation for conversational characters.
Proceedings of the 17th IEEE International Conference on Automatic Face and Gesture Recognition, 2023

2022
OverFlow: Putting flows on top of neural transducers for better TTS.
CoRR, 2022

Evaluating Sampling-based Filler Insertion with Spontaneous TTS.
Proceedings of the Thirteenth Language Resources and Evaluation Conference, 2022

Where's the uh, hesitation? The interplay between filled pause location, speech rate and fundamental frequency in perception of confidence.
Proceedings of the Interspeech 2022, 2022

Neural HMMS Are All You Need (For High-Quality Attention-Free TTS).
Proceedings of the IEEE International Conference on Acoustics, 2022

2021
Integrated Speech and Gesture Synthesis.
Proceedings of the ICMI '21: International Conference on Multimodal Interaction, 2021

2020
Augmented Prompt Selection for Evaluation of Spontaneous Speech Synthesis.
Proceedings of The 12th Language Resources and Evaluation Conference, 2020

Generating coherent spontaneous speech and gesture from text.
Proceedings of the IVA '20: ACM International Conference on Intelligent Virtual Agents, 2020

Breathing and Speech Planning in Spontaneous Speech Synthesis.
Proceedings of the 2020 IEEE International Conference on Acoustics, 2020

2019
Spontaneous Conversational Speech Synthesis from Found Data.
Proceedings of the Interspeech 2019, 2019

Off the Cuff: Exploring Extemporaneous Speech Delivery with TTS.
Proceedings of the Interspeech 2019, 2019

The Greennn Tree - Lengthening Position Influences Uncertainty Perception.
Proceedings of the Interspeech 2019, 2019

Casting to Corpus: Segmenting and Selecting Spontaneous Dialogue for Tts with a Cnn-lstm Speaker-dependent Breath Detector.
Proceedings of the IEEE International Conference on Acoustics, 2019

Mapping Theoretical and Methodological Perspectives for Understanding Speech Interface Interactions.
Proceedings of the Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, 2019

2017
Synthesising Uncertainty: The Interplay of Vocal Effort and Hesitation Disfluencies.
Proceedings of the Interspeech 2017, 2017

Using crowd-sourcing for the design of listening agents: challenges and opportunities.
Proceedings of the 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents, 2017

They Know as Much as We Do: Knowledge Estimation and Partner Modelling of Artificial Partners.
Proceedings of the 39th Annual Meeting of the Cognitive Science Society, 2017

2015
The effect of soft, modal and loud voice levels on entrainment in noisy conditions.
Proceedings of the INTERSPEECH 2015, 2015

2014
Predicting synthetic voice style from facial expressions. An application for augmented conversations.
Speech Commun., 2014

Facial expression-based affective speech translation.
J. Multimodal User Interfaces, 2014

2013
A system for facial expression-based affective speech translation.
Proceedings of the 18th International Conference on Intelligent User Interfaces, 2013

2012
Synthesizing expressive speech from amateur audiobook recordings.
Proceedings of the 2012 IEEE Spoken Language Technology Workshop (SLT), 2012

WinkTalk: a demonstration of a multimodal speech synthesis platform linking facial expressions to expressive synthetic voices.
Proceedings of the Third Workshop on Speech and Language Processing for Assistive Technologies, 2012

Evaluating expressive speech synthesis from audiobook corpora for conversational phrases.
Proceedings of the Eighth International Conference on Language Resources and Evaluation, 2012

Rapidly Testing the Interaction Model of a Pronunciation Training System via Wizard-of-Oz.
Proceedings of the Eighth International Conference on Language Resources and Evaluation, 2012

Detecting a targeted voice style in an audiobook using voice quality features.
Proceedings of the 2012 IEEE International Conference on Acoustics, 2012

2011
Clustering Expressive Speech Styles in Audiobooks Using Glottal Source Parameters.
Proceedings of the INTERSPEECH 2011, 2011


  Loading...