Update Docker command arguments for llama-server and vllm with volume mounts

This commit is contained in:
2025-09-25 22:51:51 +02:00
parent 282344af23
commit 031d6c7017

View File

@@ -137,9 +137,11 @@ func LoadConfig(configPath string) (AppConfig, error) {
Command: "llama-server", Command: "llama-server",
Args: []string{}, Args: []string{},
Docker: &DockerSettings{ Docker: &DockerSettings{
Enabled: false, Enabled: false,
Image: "ghcr.io/ggml-org/llama.cpp:server", Image: "ghcr.io/ggml-org/llama.cpp:server",
Args: []string{"run", "--rm", "--network", "host", "--gpus", "all"}, Args: []string{
"run", "--rm", "--network", "host", "--gpus", "all",
"-v", filepath.Join(getDefaultDataDirectory(), "llama.cpp") + ":/root/.cache/llama.cpp"},
Environment: map[string]string{}, Environment: map[string]string{},
}, },
}, },
@@ -147,9 +149,12 @@ func LoadConfig(configPath string) (AppConfig, error) {
Command: "vllm", Command: "vllm",
Args: []string{"serve"}, Args: []string{"serve"},
Docker: &DockerSettings{ Docker: &DockerSettings{
Enabled: false, Enabled: false,
Image: "vllm/vllm-openai:latest", Image: "vllm/vllm-openai:latest",
Args: []string{"run", "--rm", "--network", "host", "--gpus", "all", "--shm-size", "1g"}, Args: []string{
"run", "--rm", "--network", "host", "--gpus", "all", "--shm-size", "1g",
"-v", filepath.Join(getDefaultDataDirectory(), "huggingface") + ":/root/.cache/huggingface",
},
Environment: map[string]string{}, Environment: map[string]string{},
}, },
}, },