Files
llamactl/pkg/backends/llamacpp/parser.go

36 lines
1010 B
Go

package llamacpp
import (
"llamactl/pkg/backends"
)
// ParseLlamaCommand parses a llama-server command string into LlamaServerOptions
// Supports multiple formats:
// 1. Full command: "llama-server --model file.gguf"
// 2. Full path: "/usr/local/bin/llama-server --model file.gguf"
// 3. Args only: "--model file.gguf --gpu-layers 32"
// 4. Multiline commands with backslashes
func ParseLlamaCommand(command string) (*LlamaServerOptions, error) {
config := backends.CommandParserConfig{
ExecutableNames: []string{"llama-server"},
MultiValuedFlags: map[string]struct{}{
"override_tensor": {},
"override_kv": {},
"lora": {},
"lora_scaled": {},
"control_vector": {},
"control_vector_scaled": {},
"dry_sequence_breaker": {},
"logit_bias": {},
},
}
var llamaOptions LlamaServerOptions
if err := backends.ParseCommand(command, config, &llamaOptions); err != nil {
return nil, err
}
return &llamaOptions, nil
}