BMCook: A task-agnostic compression toolkit for big models Z Zhang, B Gong, Y Chen, X Han, G Zeng, W Zhao, Y Chen, Z Liu, M Sun Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022 | 5 | 2022 |
SHUOWEN-JIEZI: Linguistically informed tokenizers for Chinese language model pretraining C Si, Z Zhang, Y Chen, F Qi, X Wang, Z Liu, M Sun arXiv preprint arXiv:2106.00400, 2021 | 5 | 2021 |
Sub-character tokenization for Chinese pretrained language models C Si, Z Zhang, Y Chen, F Qi, X Wang, Z Liu, Y Wang, Q Liu, M Sun Transactions of the Association for Computational Linguistics 11, 469-487, 2023 | 3 | 2023 |
CFDBench: A Comprehensive Benchmark for Machine Learning Methods in Fluid Dynamics Y Luo, Y Chen, Z Zhang arXiv preprint arXiv:2310.05963, 2023 | 2 | 2023 |
READIN: A Chinese multi-task benchmark with realistic and diverse input noises C Si, Z Zhang, Y Chen, X Wang, Z Liu, M Sun arXiv preprint arXiv:2302.07324, 2023 | 1 | 2023 |
Robust and Scalable Model Editing for Large Language Models Y Chen, Z Zhang, X Han, C Xiao, Z Liu, C Chen, K Li, T Yang, M Sun arXiv preprint arXiv:2403.17431, 2024 | | 2024 |
Bench: Extending Long Context Evaluation Beyond 100K Tokens X Zhang, Y Chen, S Hu, Z Xu, J Chen, MK Hao, X Han, ZL Thai, S Wang, ... arXiv preprint arXiv:2402.13718, 2024 | | 2024 |