Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation Apple Machine Learning Research
Aligning large language models (LLMs) with human expectations without human-annotated preference data is an important problem. In this paper, we propose a method to evaluate the response preference by using the output probabilities of response pairs under contrastive prompt pairs, which could achieve better performance… Read More »Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation Apple Machine Learning Research