
LLM visual priors
Can Reading Countless Books Help Large Models “See” the Visual World? Meta Reveals the Origins of LLM Visual Priors
How LLMs Learn to “See” Without Visual Input A Large Language Model (LLM) trained exclusively on text — without any visual data — can still develop prior abilities transferable to vision tasks. This surprising discovery comes from a recent paper by Meta Superintelligence Labs and the University of Oxford. --- Study Overview