diff --git a/vllm/vllm-deployment.yaml b/vllm/vllm-deployment.yaml index 4c8e733..dd9d994 100644 --- a/vllm/vllm-deployment.yaml +++ b/vllm/vllm-deployment.yaml @@ -28,7 +28,9 @@ spec: value: /.cache - name: shm-size value: 1g - command: ["watch", "/bin/ls"] + command: ["/bin/bash", "-c"] + args: + - while true; do sleep 2600; done #command: ["python3", "-m", "vllm.entrypoints.openai.api_server"] #args: ["--model=meta-llama/Llama-2-7b-hf", # "--gpu-memory-utilization=0.95",