Kalin Stefanov

Orcid: 0000-0002-0861-8660

According to our database1, Kalin Stefanov authored at least 33 papers between 2012 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
GTA-HDR: A Large-Scale Synthetic Dataset for HDR Image Reconstruction.
CoRR, 2024

Human Brain Exhibits Distinct Patterns When Listening to Fake Versus Real Audio: Preliminary Evidence.
CoRR, 2024

HistoHDR-Net: Histogram Equalization for Single LDR to HDR Image Translation.
CoRR, 2024

2023
<i>Glitch in the matrix</i>: A large scale benchmark for content driven audio-visual forgery detection and localization.
Comput. Vis. Image Underst., November, 2023

AV-Deepfake1M: A Large-Scale LLM-Driven Audio-Visual Deepfake Dataset.
CoRR, 2023

S-HR-VQVAE: Sequential Hierarchical Residual Learning Vector Quantized Variational Autoencoder for Video Prediction.
CoRR, 2023

"Glitch in the Matrix!": A Large Scale Benchmark for Content Driven Audio-Visual Forgery Detection and Localization.
CoRR, 2023

MARLIN: Masked Autoencoder for facial video Representation LearnINg.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

ArtHDR-Net: Perceptually Realistic and Accurate HDR Content Creation.
Proceedings of the Asia Pacific Signal and Information Processing Association Annual Summit and Conference, 2023

2022
Visual Representations of Physiological Signals for Fake Video Detection.
CoRR, 2022

Do You Really Mean That? Content Driven Audio-Visual Deepfake Dataset and Multimodal Method for Temporal Forgery Localization.
CoRR, 2022

Graph-based Group Modelling for Backchannel Detection.
Proceedings of the MM '22: The 30th ACM International Conference on Multimedia, Lisboa, Portugal, October 10, 2022

Hierarchical Residual Learning Based Vector Quantized Variational Autoencoder for Image Reconstruction and Generation.
Proceedings of the 33rd British Machine Vision Conference 2022, 2022

2021
Group-Level Focus of Visual Attention for Improved Next Speaker Prediction.
Proceedings of the MM '21: ACM Multimedia Conference, Virtual Event, China, October 20, 2021

Group-Level Focus of Visual Attention for Improved Active Speaker Detection.
Proceedings of the ICMI '21 Companion: Companion Publication of the 2021 International Conference on Multimodal Interaction, Montreal, QC, Canada, October 18, 2021

2020
Self-Supervised Vision-Based Detection of the Active Speaker as Support for Socially Aware Language Acquisition.
IEEE Trans. Cogn. Dev. Syst., 2020

Spatial Bias in Vision-Based Voice Activity Detection.
Proceedings of the 25th International Conference on Pattern Recognition, 2020

Multimodal Automatic Coding of Client Behavior in Motivational Interviewing.
Proceedings of the ICMI '20: International Conference on Multimodal Interaction, 2020

OpenSense: A Platform for Multimodal Data Acquisition and Behavior Perception.
Proceedings of the ICMI '20: International Conference on Multimodal Interaction, 2020

Emotion or expressivity? An automated analysis of nonverbal perception in a social dilemma.
Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition, 2020

2019
Multimodal Learning for Identifying Opportunities for Empathetic Responses.
Proceedings of the International Conference on Multimodal Interaction, 2019

Multimodal Analysis and Estimation of Intimate Self-Disclosure.
Proceedings of the International Conference on Multimodal Interaction, 2019

Towards Digitally-Mediated Sign Language Communication.
Proceedings of the 7th International Conference on Human-Agent Interaction, 2019

2018
Recognition and Generation of Communicative Signals: Modeling of Hand Gestures, Speech Activity and Eye-Gaze in Human-Machine Interaction.
PhD thesis, 2018

Webcam-based Eye Gaze Tracking under Natural Head Movement.
CoRR, 2018

2017
Self-Supervised Vision-Based Detection of the Active Speaker as a Prerequisite for Socially-Aware Language Acquisition.
CoRR, 2017

2016
A Multi-party Multi-modal Dataset for Focus of Visual Attention in Human-human and Human-robot Interaction.
Proceedings of the Tenth International Conference on Language Resources and Evaluation LREC 2016, 2016

Look who's talking: visual identification of the active speaker in multi-party human-robot interaction.
Proceedings of the 2nd Workshop on Advancements in Social Signal Processing for Multimodal Interaction, 2016

2015
Public Speaking Training with a Multimodal Interactive Virtual Audience Framework.
Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, November 09, 2015

2014
The Tutorbot Corpus ― A Corpus for Studying Tutoring Behaviour in Multiparty Face-to-Face Spoken Dialogue.
Proceedings of the Ninth International Conference on Language Resources and Evaluation, 2014

Human-robot collaborative tutoring using multiparty multimodal spoken dialogue.
Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, 2014

2013
Tutoring Robots - Multiparty Multimodal Social Dialogue with an Embodied Tutor.
Proceedings of the Innovative and Creative Developments in Multimodal Interaction Systems, 2013

2012
Multimodal multiparty social interaction with the furhat head.
Proceedings of the International Conference on Multimodal Interaction, 2012


  Loading...