The official repository for Improved Techniques for Optimization-Based Jailbreaking on Large Language Models.
Please feel free to contact [email protected] if you have any question.
python attack_llm_core_best_update_our_target.py --behaviors_config=behaviors_ours_config.json
python generate_our_config.py
python run_multiple_attack_our_target.py --behaviors_config=behaviours_gcss_config_init_v2_continued.json --output_path=gcss --model_path="/home/LLM/Llama-2-7b-chat-hf"
Kindly include a reference to this paper in your publications if it helps your research:
@article{jia2024improved,
title={Improved Techniques for Optimization-Based Jailbreaking on Large Language Models},
author={Xiaojun Jia and Tianyu Pang and Chao Du and Yihao Huang and Jindong Gu and Yang Liu and Xiaochun Cao and Min Lin},
year={2024},
eprint={2405.21018}
}