Quantifying the Impact of Model Quantization on the Transferability of Adversarial Attacks in Large Language Models

DSpace/Manakin Repository

Show simple item record

dc.contributor.advisor Pelillo, Marcello it_IT
dc.contributor.author Novello, Martina <2000> it_IT
dc.date.accessioned 2024-09-30 it_IT
dc.date.accessioned 2024-11-13T12:08:25Z
dc.date.issued 2024-10-17 it_IT
dc.identifier.uri http://hdl.handle.net/10579/27720
dc.description.abstract In recent years, Large Language Models (LLMs) have gained significant popularity due to their remarkable abilities in understanding, processing, and generating human language. The rapid advancement of these models has contributed to their growing adoption in a variety of industrial environments. However, the deployment of these models comes with significant challenges, particularly in terms of security and computational efficiency. In particular, LLMs are vulnerable to small perturbations in the input data. Even minor changes, such as slight modifications to the input text or injection of suffixes that seem random, can lead the model to change its decision or generate incorrect, biased, or harmful content. This thesis examines the relationship between model quantization, the process of decreasing the precision of neural network weights, and the transferability of adversarial attacks in LLMs. The primary goals are to assess the effectiveness of adversarial attacks and to determine whether attacks crafted on quantized models can successfully transfer to their non-quantized counterparts, exposing potential security risks. By conducting experiments across a range of models and attack scenarios, the research demonstrates that attacks targeting low-precision models can effectively compromise models of higher precision. This finding highlights a critical security gap that could be exploited by malicious actors, emphasizing the need for more secure quantization strategies. it_IT
dc.language.iso en it_IT
dc.publisher Università Ca' Foscari Venezia it_IT
dc.rights © Martina Novello, 2024 it_IT
dc.title Quantifying the Impact of Model Quantization on the Transferability of Adversarial Attacks in Large Language Models it_IT
dc.title.alternative Quantifying the Impact of Model Quantization on the Transferability of Adversarial Attacks in Large Language Models it_IT
dc.type Master's Degree Thesis it_IT
dc.degree.name Computer science and information technology it_IT
dc.degree.level Laurea magistrale it_IT
dc.degree.grantor Dipartimento di Scienze Ambientali, Informatica e Statistica it_IT
dc.description.academicyear sessione_autunnale_23-24_appello_14-10-24 it_IT
dc.rights.accessrights closedAccess it_IT
dc.thesis.matricno 880893 it_IT
dc.subject.miur INF/01 INFORMATICA it_IT
dc.description.note it_IT
dc.degree.discipline it_IT
dc.contributor.co-advisor it_IT
dc.date.embargoend 10000-01-01
dc.provenance.upload Martina Novello (880893@stud.unive.it), 2024-09-30 it_IT
dc.provenance.plagiarycheck None it_IT


Files in this item

This item appears in the following Collection(s)

Show simple item record