Comment by jprete

17 hours ago

This probably isn't backed by an LLM but instead some kind of geometric shape model.

How do you explain a horse 2 legs become 4 legs when rotated assuming they only drew 2 legs on the side view

  • The second L in LLM stands for "language". Nothing of what you're describing has to do with language modeling.

    They could be using transformers, sure. But plenty of transformers-based models are not LLMs.

    • They are probably looking for LGMs - Large Generative Models which encapsulate vision & multi-modal models.

  • The model need only recognize from the shape that it is a horse, and would know to extrapolate from there. It would presumably have some text encoding as residual from training, but it doesn't need to be fed text from the text encoder side to know that. Think of the CLIP encoder used in stable diffusion.