As end users progressively depend on Large Language Models (LLMs) to perform their day by day jobs, their worries regarding the likely leakage of personal knowledge by these versions have surged.Adversarial Attacks: Attackers are acquiring tactics to manipulate AI models by poisoned training knowledge, adversarial illustrations, and various strateg