The rise and potential of large language model based agents: A survey Z Xi, W Chen, X Guo, W He, Y Ding, B Hong, M Zhang, J Wang, S Jin, ... arXiv preprint arXiv:2309.07864, 2023 | 486 | 2023 |
Textflint: Unified multilingual robustness evaluation toolkit for natural language processing X Wang, Q Liu, T Gui, Q Zhang, Y Zou, X Zhou, J Ye, Y Zhang, R Zheng, ... Proceedings of the 59th Annual Meeting of the Association for Computational …, 2021 | 121* | 2021 |
Shadow alignment: The ease of subverting safely-aligned language models X Yang*, X Wang*, Q Zhang, L Petzold, WY Wang, X Zhao, D Lin arXiv preprint arXiv:2310.02949, 2023 | 78 | 2023 |
Instructuie: Multi-task instruction tuning for unified information extraction X Wang, W Zhou, C Zu, H Xia, T Chen, Y Zhang, R Zheng, J Ye, Q Zhang, ... arXiv preprint arXiv:2304.08085, 2023 | 71* | 2023 |
LoRAMoE: Revolutionizing mixture of experts for maintaining world knowledge in language model alignment S Dou, E Zhou, Y Liu, S Gao, J Zhao, W Shen, Y Zhou, Z Xi, X Wang, ... ACL 2024, 2023 | 49* | 2023 |
Secrets of rlhf in large language models part ii: Reward modeling B Wang, R Zheng, L Chen, Y Liu, S Dou, C Huang, W Shen, S Jin, E Zhou, ... arXiv preprint arXiv:2401.06080, 2024 | 44* | 2024 |
Orthogonal Subspace Learning for Language Model Continual Learning X Wang, T Chen, Q Ge, H Xia, R Bao, R Zheng, Q Zhang, T Gui, X Huang EMNLP 2023 findings, 2023 | 33 | 2023 |
MINER: Improving out-of-vocabulary named entity recognition from an information theoretic perspective X Wang, S Dou, L Xiong, Y Zou, Q Zhang, T Gui, L Qiao, Z Cheng, ... Proceedings of the 60th Annual Meeting of the Association for Computational …, 2022 | 30 | 2022 |
EasyJailbreak: A Unified Framework for Jailbreaking Large Language Models W Zhou*, X Wang*, L Xiong, H Xia, Y Gu, M Chai, F Zhu, C Huang, S Dou, ... arXiv preprint arXiv:2403.12171, 2024 | 24* | 2024 |
TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models X Wang, Y Zhang, T Chen, S Gao, S Jin, X Yang, Z Xi, R Zheng, Y Zou, ... arXiv preprint arXiv:2310.06762, 2023 | 17* | 2023 |
CodeChameleon: Personalized Encryption Framework for Jailbreaking Large Language Models H Lv*, X Wang*, Y Zhang, C Huang, S Dou, J Ye, T Gui, Q Zhang, ... arXiv preprint arXiv:2402.16717, 2024 | 15 | 2024 |
StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback S Dou, Y Liu, H Jia, L Xiong, E Zhou, J Shan, C Huang, X Wang, W Shen, ... ACL 2024, 2024 | 13* | 2024 |
Improving generalization of alignment with human preferences through group invariant learning R Zheng, W Shen, Y Hua, W Lai, S Dou, Y Zhou, Z Xi, X Wang, H Huang, ... ICLR 2024, 2023 | 9* | 2023 |
Farewell to Aimless Large-scale Pretraining: Influential Subset Selection for Language Model X Wang, W Zhou, Q Zhang, J Zhou, S Gao, J Wang, M Zhang, X Gao, ... ACL 2023 findings, 2023 | 8 | 2023 |
Navigating the OverKill in Large Language Models C Shi*, X Wang*, Q Ge, S Gao, X Yang, T Gui, Q Zhang, X Huang, X Zhao, ... ACL 2024, 2024 | 7 | 2024 |
Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning Z Xi, W Chen, B Hong, S Jin, R Zheng, W He, Y Ding, S Liu, X Guo, ... ICML 2024, 2024 | 5 | 2024 |
Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning X Wang, T Chen, X Yang, Q Zhang, X Zhao, D Lin arXiv preprint arXiv:2404.10552, 2024 | 3 | 2024 |
Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback S Gao, Q Ge, W Shen, S Dou, J Ye, X Wang, R Zheng, Y Zou, Z Chen, ... ICML 2024, 2024 | 3 | 2024 |
DSRM: Boost Textual Adversarial Training with Distribution Shift Risk Minimization S Gao, S Dou, Y Liu, X Wang, Q Zhang, Z Wei, J Ma, Y Shan ACL 2023, 2023 | 3 | 2023 |
A Confidence-based Partial Label Learning Model for Crowd-Annotated Named Entity Recognition L Xiong, J Zhou, Q Zhu, X Wang, Y Wu, Q Zhang, T Gui, X Huang, J Ma, ... ACL 2023 findings, 2023 | 3 | 2023 |