Page history
15 March 2025
14 March 2025
no edit summary
−41
no edit summary
no edit summary
+85
no edit summary
+1,183
Created page with "'''Backdooring Large Language Models (LLMs)''' refers to the process of intentionally embedding hidden, malicious behaviors—known as Backdoors—into LLMs during their training or fine-tuning phases. These Backdoors enable the model to behave normally under typical conditions but trigger undesirable outputs, such as malicious code or deceptive responses, when specific conditions or inputs are met. This phenomenon raises significant concerns about the se..."
+6,845