Hyper-CL: Conditioning Sentence Representations with Hypernetworks
arxiv(2024)
摘要
While the introduction of contrastive learning frameworks in sentence
representation learning has significantly contributed to advancements in the
field, it still remains unclear whether state-of-the-art sentence embeddings
can capture the fine-grained semantics of sentences, particularly when
conditioned on specific perspectives. In this paper, we introduce Hyper-CL, an
efficient methodology that integrates hypernetworks with contrastive learning
to compute conditioned sentence representations. In our proposed approach, the
hypernetwork is responsible for transforming pre-computed condition embeddings
into corresponding projection layers. This enables the same sentence embeddings
to be projected differently according to various conditions. Evaluation on two
representative conditioning benchmarks, namely conditional semantic text
similarity and knowledge graph completion, demonstrates that Hyper-CL is
effective in flexibly conditioning sentence representations, showcasing its
computational efficiency at the same time. We also provide a comprehensive
analysis of the inner workings of our approach, leading to a better
interpretation of its mechanisms.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要