关注
Zhongzhu Zhou
Zhongzhu Zhou
其他姓名Zhongzhu Charlie Zhou
Ph.D. Candidate at the University of Sydney
在 sydney.edu.au 的电子邮件经过验证 - 首页
标题
引用次数
引用次数
年份
Deepspeed-chat: Easy, fast and affordable rlhf training of chatgpt-like models at all scales
Z Yao, RY Aminabadi, O Ruwase, S Rajbhandari, X Wu, AA Awan, ...
arXiv preprint arXiv:2308.01320, 2023
532023
Flash-llm: Enabling cost-effective and highly-efficient large generative model inference with unstructured sparsity
H Xia, Z Zheng, Y Li, D Zhuang, Z Zhou, X Qiu, Y Li, W Lin, SL Song
arXiv preprint arXiv:2309.10285, 2023
392023
Renaissance: A survey into ai text-to-image generation in the era of large model
F Bie, Y Yang, Z Zhou, A Ghanem, M Zhang, Z Yao, X Wu, C Holmes, ...
arXiv preprint arXiv:2309.00810, 2023
192023
Fp6-llm: Efficiently serving large language models through fp6-centric algorithm-system co-design
H Xia, Z Zheng, X Wu, S Chen, Z Yao, S Youn, A Bakhtiari, M Wyatt, ...
arXiv preprint arXiv:2401.14112, 2024
112024
DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies
SL Song, B Kruft, M Zhang, C Li, S Chen, C Zhang, M Tanaka, X Wu, ...
arXiv preprint arXiv:2310.04610, 2023
62023
Binary neural network for automated visual surface defect detection
W Liu, J Zhang, Z Su, Z Zhou, L Liu
Sensors 21 (20), 6868, 2021
62021
CorDA: Context-Oriented Decomposition Adaptation of Large Language Models
Y Yang, X Li, Z Zhou, SL Song, J Wu, L Nie, B Ghanem
arXiv preprint arXiv:2406.05223, 2024
32024
{Quant-LLM}: Accelerating the Serving of Large Language Models via {FP6-Centric}{Algorithm-System}{Co-Design} on Modern {GPUs}
H Xia, Z Zheng, X Wu, S Chen, Z Yao, S Youn, A Bakhtiari, M Wyatt, ...
2024 USENIX Annual Technical Conference (USENIX ATC 24), 699-713, 2024
32024
JSidentify: A hybrid framework for detecting plagiarism among JavaScript code in online mini games
Q Xia, Z Zhou, Z Li, B Xu, W Zou, Z Chen, H Ma, G Liang, H Lu, S Guo, ...
Proceedings of the ACM/IEEE 42nd International Conference on Software …, 2020
32020
Flash-LLM: Enabling Cost-E ective and Highly-E icient Large Generative Model Inference with Unstructured Sparsity
H Xia, Z Zheng, Y Li, D Zhuang, Z Zhou, X Qiu, Y Li, W Lin, SL Song
系统目前无法执行此操作,请稍后再试。
文章 1–10