Update Docker command arguments for llama-server and vllm with volume mounts

This commit is contained in:
2025-09-25 22:51:51 +02:00
parent 282344af23
commit 031d6c7017

View File

@@ -137,9 +137,11 @@ func LoadConfig(configPath string) (AppConfig, error) {
Command: "llama-server",
Args: []string{},
Docker: &DockerSettings{
Enabled: false,
Image: "ghcr.io/ggml-org/llama.cpp:server",
Args: []string{"run", "--rm", "--network", "host", "--gpus", "all"},
Enabled: false,
Image: "ghcr.io/ggml-org/llama.cpp:server",
Args: []string{
"run", "--rm", "--network", "host", "--gpus", "all",
"-v", filepath.Join(getDefaultDataDirectory(), "llama.cpp") + ":/root/.cache/llama.cpp"},
Environment: map[string]string{},
},
},
@@ -147,9 +149,12 @@ func LoadConfig(configPath string) (AppConfig, error) {
Command: "vllm",
Args: []string{"serve"},
Docker: &DockerSettings{
Enabled: false,
Image: "vllm/vllm-openai:latest",
Args: []string{"run", "--rm", "--network", "host", "--gpus", "all", "--shm-size", "1g"},
Enabled: false,
Image: "vllm/vllm-openai:latest",
Args: []string{
"run", "--rm", "--network", "host", "--gpus", "all", "--shm-size", "1g",
"-v", filepath.Join(getDefaultDataDirectory(), "huggingface") + ":/root/.cache/huggingface",
},
Environment: map[string]string{},
},
},