Certificate of Completion
THIS ACKNOWLEDGES THAT
HAS COMPLETED THE FALL 2024 DATA SCIENCE BOOT CAMP
Gayatri Davuluri
Roman Holowinsky, PhD
December 11, 2024
DIRECTOR
DATE
TEAM
Evaluating Security and Robustness of Vision Language Models
Gayatri Davuluri
This project evaluates the safety, robustness, and reliability of vision-language models (VLMs) like GPT-4o and GPT-4o-mini in Out-of-Distribution (OOD) and challenging scenarios. Using the VLLM Safety Benchmark, it explores their performance on datasets such as OODCV-VQA, Counterfactual VQA, and Sketchy-VQA, highlighting their limitations in handling counterfactual reasoning, abstract sketches, and ambiguous inputs. The findings aim to identify safety concerns, quantify model capabilities, and contribute to the development of secure, reliable VLMs for real-world applications involving nuanced and complex visual contexts.