mirror of
https://github.com/lordmathis/llamactl.git
synced 2025-11-06 00:54:23 +00:00
Add LRU eviction configuration for instances
This commit is contained in:
@@ -58,6 +58,9 @@ type InstancesConfig struct {
|
|||||||
// Maximum number of instances that can be running at the same time
|
// Maximum number of instances that can be running at the same time
|
||||||
MaxRunningInstances int `yaml:"max_running_instances,omitempty"`
|
MaxRunningInstances int `yaml:"max_running_instances,omitempty"`
|
||||||
|
|
||||||
|
// Enable LRU eviction for instance logs
|
||||||
|
EnableLRUEviction bool `yaml:"enable_lru_eviction"`
|
||||||
|
|
||||||
// Path to llama-server executable
|
// Path to llama-server executable
|
||||||
LlamaExecutable string `yaml:"llama_executable"`
|
LlamaExecutable string `yaml:"llama_executable"`
|
||||||
|
|
||||||
@@ -117,6 +120,7 @@ func LoadConfig(configPath string) (AppConfig, error) {
|
|||||||
AutoCreateDirs: true,
|
AutoCreateDirs: true,
|
||||||
MaxInstances: -1, // -1 means unlimited
|
MaxInstances: -1, // -1 means unlimited
|
||||||
MaxRunningInstances: -1, // -1 means unlimited
|
MaxRunningInstances: -1, // -1 means unlimited
|
||||||
|
EnableLRUEviction: true,
|
||||||
LlamaExecutable: "llama-server",
|
LlamaExecutable: "llama-server",
|
||||||
DefaultAutoRestart: true,
|
DefaultAutoRestart: true,
|
||||||
DefaultMaxRestarts: 3,
|
DefaultMaxRestarts: 3,
|
||||||
@@ -220,6 +224,11 @@ func loadEnvVars(cfg *AppConfig) {
|
|||||||
cfg.Instances.MaxRunningInstances = m
|
cfg.Instances.MaxRunningInstances = m
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if enableLRUEviction := os.Getenv("LLAMACTL_ENABLE_LRU_EVICTION"); enableLRUEviction != "" {
|
||||||
|
if b, err := strconv.ParseBool(enableLRUEviction); err == nil {
|
||||||
|
cfg.Instances.EnableLRUEviction = b
|
||||||
|
}
|
||||||
|
}
|
||||||
if llamaExec := os.Getenv("LLAMACTL_LLAMA_EXECUTABLE"); llamaExec != "" {
|
if llamaExec := os.Getenv("LLAMACTL_LLAMA_EXECUTABLE"); llamaExec != "" {
|
||||||
cfg.Instances.LlamaExecutable = llamaExec
|
cfg.Instances.LlamaExecutable = llamaExec
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user