Adversarial Attacks on Large Language Models and Defense Mechanisms
In the past few years, surprisingly, Large Language Models (LLMs) have advanced very rapidly, from search engines and chatbots to assistants for enterprises. Although these models are very good at understanding and even writing text, new security risks are posed by their complexity. First, adversarial inputs are created by attackers that can disturb the normal behavior of a model. Unlike normal vulnerabilities and weaknesses in … Read more