A new study entitled ‘Image-to-Text Logic Jailbreak: Your Imagination Can Help You Do Anything’ has found visual language models, like GPT-4o, can be tricked into producing harmful text outputs but feeding them a flowchart image depicting a harmful activity alongside a text prompt asking for details about the process.
The researchers of the study found that GPT-4o, probably the most popular visual language model, is…
Read the full article here