Hacker News new | past | comments | ask | show | jobs | submit login

Isn't GPT-4o multimodal? Shouldn't I be able to just feed in an image of the rendered HTML, instead of doing work to strip tags out?



it is theoretically possible, but the results and bandwidth would be worse. sending an image that large would take a lot longer than sending text


This. Images passed to LLMs are typically downsampled to something like 512x512 because that’s perfectly good for feature extraction. Getting text would mean very large images so the text is still readable.


Images are much less reliable than text, unfortunately.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: