Vision language models (VLMs) can be manipulated through adversarial image perturbations, similar to traditional image classifiers. Using techniques like Projected Gradient Descent (PGD), attackers can craft pixel-level modifications or adversarial patches that cause VLMs to generate unexpected outputs. The article demonstrates
•9m read time• From developer.nvidia.com
Table of contents
Vision language modelsEvading image classifiersBuilding adversarial images for VLMsThe difference with VLMsExtending the attackLearn moreSort: