Tag: Contrastive Learning
All the articles with the tag "Contrastive Learning".
-
SoftCoT++: Test-Time Scaling with Soft Chain-of-Thought Reasoning
SoftCoT++ 通过在连续潜在空间中引入多样化初始令牌和对比学习实现测试时扩展,显著提升了大型语言模型在多个推理任务上的性能,并与传统离散空间扩展方法展现出协同效应。
-
Diffusion vs. Autoregressive Language Models: A Text Embedding Perspective
本文提出基于扩散语言模型的文本嵌入方法DIFFEMBED,利用其双向注意力机制在长文档检索和推理密集型任务上显著优于自回归LLM嵌入模型,同时在传统嵌入任务上表现相当。
-
Language Models are Universal Embedders
本文基于多语言解码器模型(如BLOOM)提出通用嵌入器构建方法,通过对比学习和参数高效微调实现跨语言、跨任务的高质量嵌入,实验表明其在多语言和多任务场景中具有显著潜力和泛化能力。
-
Breaking the Modality Barrier: Universal Embedding Learning with Multimodal LLMs
本文提出UniME框架,通过文本判别知识蒸馏和硬负例增强指令微调,利用多模态大语言模型学习通用的多模态嵌入,提高了下游任务的判别性和组合能力。
-
Style Feature Extraction Using Contrastive Conditioned Variational Autoencoders with Mutual Information Constraints
This paper proposes a novel method combining contrastive learning with conditional variational autoencoders and mutual information constraints to extract style features from unlabeled data, demonstrating effectiveness on simple datasets like MNIST while facing challenges with natural image datasets due to augmentation limitations and qualitative evaluation.