What Can CLIP Learn From Task-specific Experts? Apple Machine Learning Research
[[{“value”:”This paper has been accepted to the UniReps Workshop in NeurIPS 2023. Contrastive language image pretraining has become the standard approach for training vision language models. Despite the utility of CLIP visual features as global representations for images, they have limitations when it comes to… Read More »What Can CLIP Learn From Task-specific Experts? Apple Machine Learning Research