Data Poisoning Attacks during Training/Fine-tuning
New · Open Source
Kerb - LLM Development Toolkit
Python toolkit for building production-ready LLM applications. Modular utilities for prompts, RAG, agents, structured outputs, and multi-provider support.
Was this section helpful?
Backdoor Attacks on Large Language Models, Tessa Bauman, Bruno Gašperov, Stjepan Begušić, Zvonko Kostanjčar, 2023arXiv preprint arXiv:2307.13501DOI: 10.48550/arXiv.2307.13501 - Discusses various backdoor attack strategies for LLMs during both pre-training and fine-tuning stages, including methods to create stealthy triggers.