top of page
CertificateBackground.png

Certificate of Completion

ErdosHorizontal.png

THIS ACKNOWLEDGES THAT

HAS COMPLETED THE FALL 2024 DATA SCIENCE BOOT CAMP

Gayatri Davuluri

clear.png

Roman Holowinsky, PhD

December 11, 2024

DIRECTOR

DATE

TEAM

Evaluating Security and Robustness of Vision Language Models

Gayatri Davuluri

clear.png

This project evaluates the safety, robustness, and reliability of vision-language models (VLMs) like GPT-4o and GPT-4o-mini in Out-of-Distribution (OOD) and challenging scenarios. Using the VLLM Safety Benchmark, it explores their performance on datasets such as OODCV-VQA, Counterfactual VQA, and Sketchy-VQA, highlighting their limitations in handling counterfactual reasoning, abstract sketches, and ambiguous inputs. The findings aim to identify safety concerns, quantify model capabilities, and contribute to the development of secure, reliable VLMs for real-world applications involving nuanced and complex visual contexts.

Screen Shot 2022-06-03 at 11.31.35 AM.png
github URL
bottom of page