Picture for Divesh Lala

Divesh Lala

Multilingual and Continuous Backchannel Prediction: A Cross-lingual Study

Add code
Dec 16, 2025
Figure 1 for Multilingual and Continuous Backchannel Prediction: A Cross-lingual Study
Figure 2 for Multilingual and Continuous Backchannel Prediction: A Cross-lingual Study
Figure 3 for Multilingual and Continuous Backchannel Prediction: A Cross-lingual Study
Figure 4 for Multilingual and Continuous Backchannel Prediction: A Cross-lingual Study
Viaarxiv icon

Triadic Multi-party Voice Activity Projection for Turn-taking in Spoken Dialogue Systems

Add code
Jul 10, 2025
Viaarxiv icon

Does the Appearance of Autonomous Conversational Robots Affect User Spoken Behaviors in Real-World Conference Interactions?

Add code
Mar 17, 2025
Viaarxiv icon

An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue

Add code
Jan 28, 2025
Figure 1 for An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue
Figure 2 for An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue
Figure 3 for An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue
Figure 4 for An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue
Viaarxiv icon

Why Do We Laugh? Annotation and Taxonomy Generation for Laughable Contexts in Spontaneous Text Conversation

Add code
Jan 28, 2025
Viaarxiv icon

Human-Like Embodied AI Interviewer: Employing Android ERICA in Real International Conference

Add code
Dec 13, 2024
Figure 1 for Human-Like Embodied AI Interviewer: Employing Android ERICA in Real International Conference
Figure 2 for Human-Like Embodied AI Interviewer: Employing Android ERICA in Real International Conference
Figure 3 for Human-Like Embodied AI Interviewer: Employing Android ERICA in Real International Conference
Figure 4 for Human-Like Embodied AI Interviewer: Employing Android ERICA in Real International Conference
Viaarxiv icon

Yeah, Un, Oh: Continuous and Real-time Backchannel Prediction with Fine-tuning of Voice Activity Projection

Add code
Oct 21, 2024
Figure 1 for Yeah, Un, Oh: Continuous and Real-time Backchannel Prediction with Fine-tuning of Voice Activity Projection
Figure 2 for Yeah, Un, Oh: Continuous and Real-time Backchannel Prediction with Fine-tuning of Voice Activity Projection
Figure 3 for Yeah, Un, Oh: Continuous and Real-time Backchannel Prediction with Fine-tuning of Voice Activity Projection
Figure 4 for Yeah, Un, Oh: Continuous and Real-time Backchannel Prediction with Fine-tuning of Voice Activity Projection
Viaarxiv icon

Analysis and Detection of Differences in Spoken User Behaviors between Autonomous and Wizard-of-Oz Systems

Add code
Oct 04, 2024
Figure 1 for Analysis and Detection of Differences in Spoken User Behaviors between Autonomous and Wizard-of-Oz Systems
Figure 2 for Analysis and Detection of Differences in Spoken User Behaviors between Autonomous and Wizard-of-Oz Systems
Figure 3 for Analysis and Detection of Differences in Spoken User Behaviors between Autonomous and Wizard-of-Oz Systems
Figure 4 for Analysis and Detection of Differences in Spoken User Behaviors between Autonomous and Wizard-of-Oz Systems
Viaarxiv icon

Evaluation of a semi-autonomous attentive listening system with takeover prompting

Add code
Feb 21, 2024
Viaarxiv icon

Acknowledgment of Emotional States: Generating Validating Responses for Empathetic Dialogue

Add code
Feb 20, 2024
Figure 1 for Acknowledgment of Emotional States: Generating Validating Responses for Empathetic Dialogue
Figure 2 for Acknowledgment of Emotional States: Generating Validating Responses for Empathetic Dialogue
Figure 3 for Acknowledgment of Emotional States: Generating Validating Responses for Empathetic Dialogue
Figure 4 for Acknowledgment of Emotional States: Generating Validating Responses for Empathetic Dialogue
Viaarxiv icon