mirror of
https://github.com/lordmathis/llamactl.git
synced 2025-11-06 00:54:23 +00:00
Add MaxRunningInstances to InstancesConfig and implement IsRunning method
This commit is contained in:
@@ -11,6 +11,12 @@ import (
|
||||
"time"
|
||||
)
|
||||
|
||||
func (i *Process) IsRunning() bool {
|
||||
i.mu.RLock()
|
||||
defer i.mu.RUnlock()
|
||||
return i.Running
|
||||
}
|
||||
|
||||
// Start starts the llama server instance and returns an error if it fails.
|
||||
func (i *Process) Start() error {
|
||||
i.mu.Lock()
|
||||
|
||||
Reference in New Issue
Block a user