Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation Apple Machine Learning Research
This paper was accepted at the workshop I Can’t Believe It’s Not Better! (ICBINB) at NeurIPS 2023. Recent advances in image tokenizers, such as VQ-VAE, have enabled text-to-image generation using auto-regressive methods, similar to language modeling. However, these methods have yet to leverage pre-trained language… Read More »Pre-trained Language Models Do Not Help Auto-regressive Text-to-Image Generation Apple Machine Learning Research