| # QServe benchmarks | |
| This huggingface repository contains configurations and tokenizer files for all models benchmarked in our [QServe](https://github.com/mit-han-lab/qserve) project: | |
| - Llama-3-8B | |
| - Llama-2-7B | |
| - Llama-2-13B | |
| - Llama-2-70B | |
| - Llama-30B | |
| - Mistral-7B | |
| - Yi-34B | |
| - Qwen1.5-72B | |
| Please clone this repository if you wish to run our QServe benchmark code without cloning full models. | |
| Please consider citing our paper if it is helpful: | |
| ``` | |
| @article{lin2024qserve, | |
| title={QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving}, | |
| author={Lin*, Yujun and Tang*, Haotian and Yang*, Shang and Zhang, Zhekai and Xiao, Guangxuan and Gan, Chuang and Han, Song}, | |
| year={2024} | |
| } | |
| ``` |