git-disl
Pinned Loading
Repositories
-           awesome_LLM-harmful-fine-tuning-papers PublicA survey on harmful fine-tuning attack for large language model git-disl/awesome_LLM-harmful-fine-tuning-papers’s past year of commit activity 
-           Fed-alphaCDP PublicThis repo is for paper Securing Distributed SGD against Gradient Leakage Threats submitted to IEEE TPDS. git-disl/Fed-alphaCDP’s past year of commit activity 
-           GradCloak PublicCode for GradCloak: Gradient Obfuscation for Privacy-Preserving Distributed Learning as a Service git-disl/GradCloak’s past year of commit activity 
-           AFOG Public Forked from zacharyyahn/AFOGUnofficial re-implementation of "Adversarial Attention Perturbations for Large Object Detection Transformers" git-disl/AFOG’s past year of commit activity 
-           Antidote PublicThis is the unofficial re-implementation of "Antidote: Post-fine-tuning Safety Alignment for Large Language Models against Harmful Fine-tuning Attack" (ICML2025) git-disl/Antidote’s past year of commit activity 
-           Fusion-Shot Publicgit-disl/Fusion-Shot’s past year of commit activity 
-           GTLLMZoo PublicGTLLMZoo: A comprehensive framework that aggregates LLM benchmark data from multiple sources with an interactive UI for efficient model comparison, filtering, and evaluation across performance, safety, and efficiency metrics. git-disl/GTLLMZoo’s past year of commit activity 
-           Booster PublicThis is the official code for the paper "Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturbation" (ICLR2025 Oral). git-disl/Booster’s past year of commit activity 
-           Safety-Tax PublicThis is the official code for the paper "Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable". git-disl/Safety-Tax’s past year of commit activity 
Top languages
Loading…