RUS  ENG
Full version
JOURNALS // Doklady Rossijskoj Akademii Nauk. Mathematika, Informatika, Processy Upravlenia // Archive

Dokl. RAN. Math. Inf. Proc. Upr., 2025 Volume 527, Pages 332–345 (Mi danma691)

SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES

JDCEmb: joint distillation and contrastive learning for embeddings in goal-oriented dialogue systems

A. I. Burykinaa, D. R. Lednevab, D. P. Kuznetsovb

a St. Petersburg National Research University of Information Technologies, Mechanics and Optics
b Moscow Institute of Physics and Technology (National Research University), Dolgoprudny, Moscow Region

Abstract: We present JDCEmb – a new framework for training universal vector representations in goal-oriented dialogue tasks. Text encoders play a crucial role in such systems, and their quality determines the effectiveness of dialogue systems. Modern approaches to training dialogue encoders often rely on contrastive methods, which improve the distinguishability of representations but are sensitive to the selection of positive and negative pairs. This can lead to loss of important semantic information. Knowledge distillation-based methods, on the other hand, transfer more context but struggle to distinguish similar utterances and perform poorly with subtle semantic differences. JDCEmb combines the strengths of both approaches using a teacher-student architecture, where the student model is trained contrastively and aligned with the teacher model's vector representations simultaneously. This combination allows maintaining semantic richness while enhancing the distinctiveness of vector representations-crucial for dialogue systems. Experimental results on key dialogue tasks demonstrate the effectiveness of the approach: JDCEmb consistently reaches or surpasses state-of-the-art levels, outperforming strong current baseline models.

Keywords: goal-oriented dialogues, vector representations, knowledge distillation, contrastive learning, dialogue systems.

UDC: 004.8

Received: 21.08.2025
Accepted: 29.09.2025

DOI: 10.7868/S268695432507029X



Bibliographic databases:


© Steklov Math. Inst. of RAS, 2026