Data Poisoning Attacks during Training/Fine-tuning
Was this section helpful?
Backdoor Attacks on Large Language Models, Tessa Bauman, Bruno Gašperov, Stjepan Begušić, Zvonko Kostanjčar, 2023arXiv preprint arXiv:2307.13501DOI: 10.48550/arXiv.2307.13501 - Discusses various backdoor attack strategies for LLMs during both pre-training and fine-tuning stages, including methods to create stealthy triggers.