From cea6dd868f51c4a168794386eccdd282381c5898 Mon Sep 17 00:00:00 2001 From: Tyler Perkins Date: Sat, 30 Mar 2024 22:29:42 -0400 Subject: [PATCH] Update container to sleep forever --- vllm/vllm-deployment.yaml | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/vllm/vllm-deployment.yaml b/vllm/vllm-deployment.yaml index 4c8e733..dd9d994 100644 --- a/vllm/vllm-deployment.yaml +++ b/vllm/vllm-deployment.yaml @@ -28,7 +28,9 @@ spec: value: /.cache - name: shm-size value: 1g - command: ["watch", "/bin/ls"] + command: ["/bin/bash", "-c"] + args: + - while true; do sleep 2600; done #command: ["python3", "-m", "vllm.entrypoints.openai.api_server"] #args: ["--model=meta-llama/Llama-2-7b-hf", # "--gpu-memory-utilization=0.95",