使用vllm部署
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 

30 lines
639 B

Flask==3.0.0
gevent==23.9.1
greenlet==3.0.3
gunicorn==21.2.0
numpy==1.26.3
nvidia-cublas-cu12==12.1.3.1
nvidia-cuda-cupti-cu12==12.1.105
nvidia-cuda-nvrtc-cu12==12.1.105
nvidia-cuda-runtime-cu12==12.1.105
nvidia-cudnn-cu12==8.9.2.26
nvidia-cufft-cu12==11.0.2.54
nvidia-curand-cu12==10.3.2.106
nvidia-cusolver-cu12==11.4.5.107
nvidia-cusparse-cu12==12.1.0.106
nvidia-nccl-cu12==2.18.1
nvidia-nvjitlink-cu12==12.3.101
nvidia-nvtx-cu12==12.1.105
pandas==2.1.4
redis==5.0.1
requests==2.31.0
safetensors==0.4.1
tokenizers==0.15.0
torch==2.1.2
tqdm==4.66.1
transformers==4.36.2
urllib3==2.1.0
uvicorn==0.25.0
vllm==0.2.0
xformers==0.0.23.post1