zhulinJulia24
|
fb6a0df652
|
[ci] fix test env for vllm and add vllm baselines (#1481)
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
* update
---------
Co-authored-by: zhulin1 <zhulin1@pjlab.org.cn>
|
2024-09-04 19:24:09 +08:00 |
|
Lyu Han
|
1013dce60c
|
adapt to lmdeploy v0.4.0 (#1073)
* adapt to lmdeploy v0.4.0
* compatible
|
2024-04-28 19:57:40 +08:00 |
|
Fengzhe Zhou
|
b39f501563
|
[Sync] update taco (#1030)
|
2024-04-09 17:50:23 +08:00 |
|
RunningLeon
|
c54a5d3b0f
|
Support get_ppl for TurbomindModel (#878)
* update ppl for turbomindmodel
* update api_server
* rename config and set thread_safe for pytorch engine if possible
|
2024-03-06 11:44:19 +08:00 |
|
RunningLeon
|
32ba0b074e
|
Support lmdeploy pytorch engine (#875)
* add lmdeploy pytorch model
* fix
* speed up encoding and decoding
* fix
* change tokenizer
|
2024-02-22 03:46:07 -03:00 |
|