Skip to content

Commit 08272d3

Browse files
committed
升级vllm==0.8.5.post1 lmdeopy==0.8.0 sglang==0.4.6.post2 flashtts==0.1.5 infinity-emb[all]==0.0.76 版本
1 parent 57ec9da commit 08272d3

File tree

4 files changed

+116
-70
lines changed

4 files changed

+116
-70
lines changed

gpt_server/model_worker/spark_tts.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -68,14 +68,14 @@ def __init__(
6868
conv_template,
6969
model_type="tts",
7070
)
71-
71+
backend = os.environ["backend"]
7272
self.engine = AutoEngine(
7373
model_path=model_path,
7474
max_length=32768,
7575
llm_device="auto",
7676
tokenizer_device="auto",
7777
detokenizer_device="auto",
78-
backend="vllm",
78+
backend=backend,
7979
wav2vec_attn_implementation="sdpa", # 使用flash attn加速wav2vec
8080
llm_gpu_memory_utilization=0.6,
8181
seed=0,

pyproject.toml

+6-6
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[project]
22
name = "gpt_server"
3-
version = "0.4.3"
3+
version = "0.4.4"
44
description = "gpt_server是一个用于生产级部署LLMs或Embedding的开源框架。"
55
readme = "README.md"
66
license = { text = "Apache 2.0" }
@@ -11,23 +11,23 @@ dependencies = [
1111
"fastapi==0.115.0",
1212
"ffmpy",
1313
"fschat==0.2.36",
14-
"infinity-emb[all]==0.0.73",
15-
"lmdeploy==0.7.3",
14+
"infinity-emb[all]==0.0.76",
15+
"lmdeploy==0.8.0",
1616
"loguru>=0.7.2",
1717
"openai==1.55.3",
1818
"setuptools==75.2.0",
1919
"streamlit==1.39.0",
2020
"torch==2.6.0",
2121
"torchvision==0.20.1",
22-
"vllm==0.8.5",
22+
"vllm==0.8.5.post1",
2323
"qwen_vl_utils",
2424
"evalscope[perf]==0.10.1",
2525
"modelscope==1.20.1",
2626
"edge-tts>=7.0.0",
2727
"funasr>=1.2.6",
28-
"sglang[all]>=0.4.6.post1",
28+
"sglang[all]>=0.4.6.post2",
2929
"flashinfer-python",
30-
"flashtts>=0.1.0",
30+
"flashtts>=0.1.5",
3131
]
3232

3333
[tool.uv]

requirements.txt

+39-33
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,7 @@ aiohttp==3.11.18
2727
# sglang
2828
# vllm
2929
aiosignal==1.3.2
30-
# via
31-
# aiohttp
32-
# ray
30+
# via aiohttp
3331
airportsdata==20250224
3432
# via outlines
3533
aliyun-python-sdk-core==2.16.0
@@ -42,7 +40,7 @@ altair==5.5.0
4240
# via streamlit
4341
annotated-types==0.7.0
4442
# via pydantic
45-
anthropic==0.50.0
43+
anthropic==0.51.0
4644
# via sglang
4745
antlr4-python3-runtime==4.9.3
4846
# via
@@ -81,6 +79,8 @@ blake3==1.0.4
8179
# via vllm
8280
blinker==1.9.0
8381
# via streamlit
82+
blobfile==3.0.0
83+
# via sglang
8484
cachetools==5.5.2
8585
# via
8686
# evalscope
@@ -96,7 +96,7 @@ cffi==1.17.1
9696
# via
9797
# cryptography
9898
# soundfile
99-
charset-normalizer==3.4.1
99+
charset-normalizer==3.4.2
100100
# via requests
101101
click==8.1.8
102102
# via
@@ -125,13 +125,13 @@ contourpy==1.3.2
125125
# via matplotlib
126126
crcmod==1.7
127127
# via oss2
128-
cryptography==44.0.2
128+
cryptography==44.0.3
129129
# via aliyun-python-sdk-core
130130
ctranslate2==4.6.0
131131
# via infinity-emb
132-
cuda-bindings==12.8.0
132+
cuda-bindings==12.9.0
133133
# via cuda-python
134-
cuda-python==12.8.0
134+
cuda-python==12.9.0
135135
# via sglang
136136
cupy-cuda12x==13.4.1
137137
# via ray
@@ -175,7 +175,7 @@ distro==1.9.0
175175
# posthog
176176
dnspython==2.7.0
177177
# via email-validator
178-
edge-tts==7.0.1
178+
edge-tts==7.0.2
179179
# via gpt-server (pyproject.toml)
180180
editdistance==0.8.1
181181
# via
@@ -222,6 +222,7 @@ ffmpy==0.5.0
222222
# via gpt-server (pyproject.toml)
223223
filelock==3.18.0
224224
# via
225+
# blobfile
225226
# datasets
226227
# huggingface-hub
227228
# ray
@@ -230,11 +231,11 @@ filelock==3.18.0
230231
# vllm
231232
fire==0.7.0
232233
# via lmdeploy
233-
flashinfer-python==0.2.3+cu124torch2.5
234+
flashinfer-python==0.2.5+cu124torch2.5
234235
# via
235236
# gpt-server (pyproject.toml)
236237
# sglang
237-
flashtts==0.1.4
238+
flashtts==0.1.5
238239
# via gpt-server (pyproject.toml)
239240
flatbuffers==25.2.10
240241
# via onnxruntime
@@ -246,7 +247,6 @@ frozenlist==1.6.0
246247
# via
247248
# aiohttp
248249
# aiosignal
249-
# ray
250250
fschat==0.2.36
251251
# via gpt-server (pyproject.toml)
252252
fsspec==2024.6.1
@@ -259,7 +259,7 @@ funasr==1.2.6
259259
# via gpt-server (pyproject.toml)
260260
future==1.0.0
261261
# via pyloudnorm
262-
gguf==0.16.2
262+
gguf==0.16.3
263263
# via vllm
264264
gitdb==4.0.12
265265
# via gitpython
@@ -279,7 +279,7 @@ hf-transfer==0.1.9
279279
# via
280280
# infinity-emb
281281
# sglang
282-
hf-xet==1.0.5
282+
hf-xet==1.1.0
283283
# via huggingface-hub
284284
httpcore==1.0.9
285285
# via httpx
@@ -293,7 +293,7 @@ httpx==0.27.2
293293
# fschat
294294
# litellm
295295
# openai
296-
huggingface-hub==0.30.2
296+
huggingface-hub==0.31.1
297297
# via
298298
# accelerate
299299
# datasets
@@ -326,7 +326,7 @@ importlib-metadata==8.0.0
326326
# vllm
327327
importlib-resources==6.5.2
328328
# via wetextprocessing
329-
infinity-emb==0.0.73
329+
infinity-emb==0.0.76
330330
# via gpt-server (pyproject.toml)
331331
interegular==0.3.3
332332
# via
@@ -360,7 +360,7 @@ jiter==0.9.0
360360
# openai
361361
jmespath==0.10.0
362362
# via aliyun-python-sdk-core
363-
joblib==1.4.2
363+
joblib==1.5.0
364364
# via
365365
# librosa
366366
# nltk
@@ -386,7 +386,7 @@ lark==1.2.2
386386
# via
387387
# outlines
388388
# vllm
389-
latex2mathml==3.77.0
389+
latex2mathml==3.78.0
390390
# via markdown2
391391
lazy-loader==0.4
392392
# via librosa
@@ -406,12 +406,14 @@ llvmlite==0.44.0
406406
# pynndescent
407407
lm-format-enforcer==0.10.11
408408
# via vllm
409-
lmdeploy==0.7.3
409+
lmdeploy==0.8.0
410410
# via gpt-server (pyproject.toml)
411411
loguru==0.7.3
412412
# via gpt-server (pyproject.toml)
413413
lxml==5.4.0
414-
# via sacrebleu
414+
# via
415+
# blobfile
416+
# sacrebleu
415417
markdown-it-py==3.0.0
416418
# via rich
417419
markdown2==2.5.3
@@ -454,7 +456,7 @@ multiprocess==0.70.16
454456
# via
455457
# datasets
456458
# evaluate
457-
narwhals==1.37.1
459+
narwhals==1.38.0
458460
# via
459461
# altair
460462
# plotly
@@ -560,7 +562,7 @@ nvidia-cusparse-cu12==12.3.1.170
560562
# torch
561563
nvidia-cusparselt-cu12==0.6.2
562564
# via torch
563-
nvidia-ml-py==12.570.86
565+
nvidia-ml-py==12.575.51
564566
# via pynvml
565567
nvidia-nccl-cu12==2.21.5
566568
# via
@@ -624,11 +626,11 @@ opentelemetry-sdk==1.26.0
624626
# vllm
625627
opentelemetry-semantic-conventions==0.47b0
626628
# via opentelemetry-sdk
627-
opentelemetry-semantic-conventions-ai==0.4.3
629+
opentelemetry-semantic-conventions-ai==0.4.7
628630
# via vllm
629631
optimum==1.24.0
630632
# via infinity-emb
631-
orjson==3.10.17
633+
orjson==3.10.18
632634
# via
633635
# infinity-emb
634636
# sglang
@@ -678,6 +680,7 @@ parso==0.8.4
678680
# via jedi
679681
partial-json-parser==0.2.1.1.post5
680682
# via
683+
# lmdeploy
681684
# sglang
682685
# vllm
683686
peft==0.14.0
@@ -700,7 +703,7 @@ pillow==10.4.0
700703
# streamlit
701704
# torchvision
702705
# vllm
703-
platformdirs==4.3.7
706+
platformdirs==4.3.8
704707
# via
705708
# pooch
706709
# yapf
@@ -766,7 +769,9 @@ pycparser==2.22
766769
# via cffi
767770
pycryptodome==3.22.0
768771
# via oss2
769-
pydantic==2.11.3
772+
pycryptodomex==3.22.0
773+
# via blobfile
774+
pydantic==2.11.4
770775
# via
771776
# anthropic
772777
# compressed-tensors
@@ -782,7 +787,7 @@ pydantic==2.11.3
782787
# sglang
783788
# vllm
784789
# xgrammar
785-
pydantic-core==2.33.1
790+
pydantic-core==2.33.2
786791
# via pydantic
787792
pydeck==0.9.1
788793
# via streamlit
@@ -851,7 +856,7 @@ pyzmq==26.4.0
851856
# vllm
852857
qwen-vl-utils==0.0.11
853858
# via gpt-server (pyproject.toml)
854-
ray==2.43.0
859+
ray==2.46.0
855860
# via
856861
# lmdeploy
857862
# vllm
@@ -901,7 +906,7 @@ rich==13.9.4
901906
# rich-toolkit
902907
# streamlit
903908
# typer
904-
rich-toolkit==0.14.3
909+
rich-toolkit==0.14.5
905910
# via fastapi-cli
906911
rouge-chinese==1.0.3
907912
# via evalscope
@@ -963,9 +968,9 @@ setuptools==75.2.0
963968
# torch
964969
# triton
965970
# vllm
966-
sgl-kernel==0.1.0
971+
sgl-kernel==0.1.1
967972
# via sglang
968-
sglang==0.4.6.post1
973+
sglang==0.4.6.post2
969974
# via gpt-server (pyproject.toml)
970975
shellingham==1.5.4
971976
# via typer
@@ -1038,7 +1043,7 @@ tenacity==9.1.2
10381043
# via streamlit
10391044
tensorboardx==2.6.2.2
10401045
# via funasr
1041-
termcolor==3.0.1
1046+
termcolor==3.1.0
10421047
# via
10431048
# fire
10441049
# mmengine-lite
@@ -1195,6 +1200,7 @@ unicorn==2.1.3
11951200
# via evalscope
11961201
urllib3==2.4.0
11971202
# via
1203+
# blobfile
11981204
# modelscope
11991205
# requests
12001206
uvicorn==0.32.1
@@ -1211,7 +1217,7 @@ uvloop==0.21.0
12111217
# via
12121218
# sglang
12131219
# uvicorn
1214-
vllm==0.8.5
1220+
vllm==0.8.5.post1
12151221
# via gpt-server (pyproject.toml)
12161222
watchdog==5.0.3
12171223
# via streamlit

0 commit comments

Comments
 (0)