Skip to content

Commit bf57d0b

Browse files
authored
[Readme] Add accepted papers (#465)
1 parent 98c9d32 commit bf57d0b

File tree

1 file changed

+16
-0
lines changed

1 file changed

+16
-0
lines changed

README.md

Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@ xFasterTransformer is an exceptionally optimized solution for large language mod
4242
- [MLServer](#mlserver)
4343
- [Benchmark](#benchmark)
4444
- [Support](#support)
45+
- [Accepted Papers](#accepted-papers)
4546
- [Q\&A](#qa)
4647

4748
## Models overview
@@ -388,6 +389,21 @@ Benchmark scripts are provided to get the model inference performance quickly.
388389
- xFasterTransformer email: [email protected]
389390
- xFasterTransformer [wechat](https://github.com/intel/xFasterTransformer/wiki)
390391

392+
## Accepted Papers
393+
- ICLR'2024 on practical ML for limited/low resource settings: [Distributed Inference Performance Optimization for LLMs on CPUs](https://arxiv.org/abs/2407.00029)
394+
- ICML'2024 on Foundation Models in the Wild: Inference Performance Optimization for Large Language Models on CPUs
395+
- IEEE ICSESS 2024: All-in-one Approach for Large Language Models Inference
396+
397+
If xFT is useful for your research, please cite:
398+
```latex
399+
@article{he2024distributed,
400+
title={Distributed Inference Performance Optimization for LLMs on CPUs},
401+
author={He, Pujiang and Zhou, Shan and Li, Changqing and Huang, Wenhuan and Yu, Weifei and Wang, Duyi and Meng, Chen and Gui, Sheng},
402+
journal={arXiv preprint arXiv:2407.00029},
403+
year={2024}
404+
}
405+
```
406+
391407
## Q&A
392408

393409
- ***Q***: Can xFasterTransformer run on a Intel® Core™ CPU?

0 commit comments

Comments
 (0)