vllm文档
This commit is contained in:
parent
d8f24e7b96
commit
1c8f9338bb
|
@ -29,7 +29,7 @@
|
||||||
|
|
||||||
$(document).ready(function() {
|
$(document).ready(function() {
|
||||||
var host = window.location.hostname
|
var host = window.location.hostname
|
||||||
var ws = new WebSocket("ws://"+host+":8000/humanecho");
|
var ws = new WebSocket("ws://"+host+":8000/humanchat");
|
||||||
//document.getElementsByTagName("video")[0].setAttribute("src", aa["video"]);
|
//document.getElementsByTagName("video")[0].setAttribute("src", aa["video"]);
|
||||||
ws.onopen = function() {
|
ws.onopen = function() {
|
||||||
console.log('Connected');
|
console.log('Connected');
|
||||||
|
|
|
@ -0,0 +1,8 @@
|
||||||
|
1、推理加速
|
||||||
|
conda create -n vllm python=3.10
|
||||||
|
conda install pytorch==2.1.0 torchvision==0.16.0 torchaudio==2.1.0 pytorch-cuda=12.1 -c pytorch -c nvidia
|
||||||
|
|
||||||
|
|
||||||
|
python -m vllm.entrypoints.openai.api_server --tensor-parallel-size=1 --trust-remote-code --max-model-len 1024 --model THUDM/chatglm3-6b
|
||||||
|
|
||||||
|
python -m vllm.entrypoints.openai.api_server --host 127.0.0.1 --port 8101 --tensor-parallel-size=1 --trust-remote-code --max-model-len 1024 --model THUDM/chatglm3-6b
|
Loading…
Reference in New Issue