I think there’s a vast difference between “I say I can take in images as input for prompts with limitations “ and “I’m using the wrong tool for a completely absurd use case” like your microscope analogy implies.
LLM is the wrong tool for image analysis, even if the providers say that it is possible. Possibility doesn’t mean effectiveness or even usefulness. Like a microscope and onions.
Now read that FAQ. I see just a bunch of limitations descriptions, not a “I can read and correctly understand 100 percent of the images”
I think there’s a vast difference between “I say I can take in images as input for prompts with limitations “ and “I’m using the wrong tool for a completely absurd use case” like your microscope analogy implies.
LLM is the wrong tool for image analysis, even if the providers say that it is possible. Possibility doesn’t mean effectiveness or even usefulness. Like a microscope and onions.