Publications Repository - Gdańsk University of Technology

Page settings

polski
Publications Repository
Gdańsk University of Technology

Treść strony

An Adversarial Machine Learning Approach on Securing Large Language Model with Vigil, an Open-Source Initiative

Several security concerns and efforts to breach system security and prompt safety concerns have been brought to light as a result of the expanding use of LLMs. These vulnerabilities are evident and LLM models have been showing many signs of hallucination, repetitive content generation, and biases, which makes them vulnerable to malicious prompts that raise substantial concerns in regard to the dependability and efficiency of such models. It is vital to have a complete grasp of the complex behaviours of malicious attackers in order to build effective strategies for protecting modern artificial intelligence (AI) systems through the development of effective tactics. The purpose of this study is to look into some of these aspects and propose a method for preventing devastating possibilities and protecting LLMs from potential threats that attackers may pose. Vigil is an open-source LLM prompt security scanner, that is accessible as a Python library and REST API, specifically to solve these problems by employing a sophisticated adversarial machine-learning algorithm. The entire objective of this study is to make use of Vigil as a security scanner. and asses its efficiency. In this case study, we shed some light on Vigil, which effectively recognises and helps LLM prompts by identifying two varieties of threats: malicious and benign.

Authors

Additional information

DOI
Digital Object Identifier link open in new tab 10.1016/j.procs.2024.09.486
Category
Aktywność konferencyjna
Type
publikacja w wydawnictwie zbiorowym recenzowanym (także w materiałach konferencyjnych)
Language
angielski
Publication year
2024

Source: MOSTWiedzy.pl - publication "An Adversarial Machine Learning Approach on Securing Large Language Model with Vigil, an Open-Source Initiative" link open in new tab

Portal MOST Wiedzy link open in new tab