| FROM runpod/vllm:0.7.3-cuda12 | |
| COPY DeepSeek-V3-AWQ /workspace/model | |
| ENV HF_MODEL=local:///workspace/model \ | |
| VLLM_DTYPE=float16 VLLM_MAX_MODEL_LEN=8192 VLLM_NUM_GPUS=8 \ | |
| VLLM_USE_V1=0 VLLM_WORKER_MULTIPROC_METHOD=spawn \ | |
| VLLM_MARLIN_USE_ATOMIC_ADD=1 # fixes fp16 overflow noted in model card | |
| CMD ["python", "-m", "vllm.entrypoints.openai.api_server", | |
| "--host", "0.0.0.0", "--port", "8000", | |
| "--tensor-parallel-size", "8", | |
| "--gpu-memory-utilization", "0.95", | |
| "--trust-remote-code", | |
| "--served-model-name", "deepseek-chat"] | |