Name
Affiliation
Papers
GUSTAV EJE HENTER
Sound and Image Processing Laboratory, School of Electrical Engineering, KTH - Royal Institute of Technology, SE-100 44 Stockholm, Sweden
31
Collaborators
Citations 
PageRank 
52
37
11.40
Referers 
Referees 
References 
121
365
191
Search Limit
100365
Title
Citations
PageRank
Year
A Large, Crowdsourced Evaluation of Gesture Generation Systems on Common Data: The GENEA Challenge 202010.352021
Full-Glow - Fully Conditional Glow for More Realistic Image Generation.00.342021
Integrated Speech and Gesture Synthesis00.342021
HEMVIP: Human Evaluation of Multiple Videos in Parallel00.342021
Moving Fast And Slow: Analysis Of Representations And Post-Processing In Speech-Driven Automatic Gesture Generation00.342021
Transflower: probabilistic autoregressive dance generation with multimodal attention00.342021
Style-Controllable Speech-Driven Gesture Synthesis Using Normalising FlowsKeywords30.382020
Gesticulator: A framework for semantically-aware speech-driven gesture generation30.372020
Let's face it: Probabilistic multi-modal interlocutor-aware generation of facial gestures in dyadic settings00.342020
Generating coherent spontaneous speech and gesture from text.00.342020
Casting To Corpus: Segmenting And Selecting Spontaneous Dialogue For Tts With A Cnn-Lstm Speaker-Dependent Breath Detector00.342019
Analyzing Input and Output Representations for Speech-Driven Gesture Generation.00.342019
MoGlow: probabilistic and controllable motion synthesis using normalising flows20.362019
Spontaneous Conversational Speech Synthesis from Found Data00.342019
Off the Cuff - Exploring Extemporaneous Speech Delivery with TTS.00.342019
On the Importance of Representations for Speech-Driven Gesture Generation00.342019
Kernel Density Estimation-Based Markov Models with Hidden State00.342018
Investigating different representations for modeling and controlling multiple emotions in DNN-based speech synthesis.80.552018
Analysing Shortcomings of Statistical Parametric Speech Synthesis.00.342018
Deep Encoder-Decoder Models for Unsupervised Learning of Controllable Speech Synthesis.30.402018
Principles For Learning Controllable Tts From Annotated And Latent Variation40.452017
Misperceptions Of The Emotional Content Of Natural And Vocoded Speech In A Car10.372017
Bayesian Analysis of Phoneme Confusion Matrices10.372016
Minimum Entropy Rate Simplification of Stochastic Processes00.342016
Testing the consistency assumption: Pronunciation variant forced alignment in read and spontaneous speech synthesis.40.402016
Median-Based Generation Of Synthetic Speech Durations Using A Non-Parametric Approach30.432016
Are We Using Enough Listeners? No! An Empirically-Supported Critique Of Interspeech 2014 Tts Evaluations10.382015
Measuring the perceptual effects of modelling assumptions in speech synthesis using stimuli constructed from repeated natural speech.00.342014
Picking up the pieces: Causal states in noisy data, and how to recover them10.362013
Intermediate-State Hmms To Capture Continuously-Changing Signal Features20.482011
Simplified probability models for generative tasks: A rate-distortion approach00.342010