Evaluating Adversarial Vulnerabilities in Modern Large Language Models