“…Traditional security threats have prompted significant exploration into areas such as membership inference attacks (Shi et al, 2023b), backdoor attacks (Shi et al, 2023a;Xu et al, 2023), andothers (Wan et al, 2023;Shi et al, 2024). (Wang et al, 2023a;Huang et al, 2023c;Bi et al, 2023). A multitude of studies have extensively examined the trustworthiness of LLMs including the alignment (Wang et al, 2023b;Liu et al, 2023a), truthfulness (e.g., misinformation (Huang and Sun, 2023;Chen and Shu, 2023b,a) and hallucination (Xu et al, 2024;Tonmoy et al, 2024;Huang et al, 2023a;), accountability (He et al, 2024;, and fairness (Wang et al, 2023a;Huang et al, 2023c;Bi et al, 2023).…”