4.1 KiB
Troubleshooting
Issues specific to Llamactl deployment and operation.
Configuration Issues
Invalid Configuration
Problem: Invalid configuration preventing startup
Solutions:
-
Use minimal configuration:
server: host: "0.0.0.0" port: 8080 instances: port_range: [8000, 9000] -
Check data directory permissions:
# Ensure data directory is writable (default: ~/.local/share/llamactl) mkdir -p ~/.local/share/llamactl/{instances,logs}
Instance Management Issues
Model Loading Failures
Problem: Instance fails to start with model loading errors
Common Solutions:
- llama-server not found: Ensure
llama-serverbinary is in PATH - Wrong model format: Ensure model is in GGUF format
- Insufficient memory: Use smaller model or reduce context size
- Path issues: Use absolute paths to model files
Memory Issues
Problem: Out of memory errors or system becomes unresponsive
Solutions:
-
Reduce context size:
{ "n_ctx": 1024 } -
Use quantized models:
- Try Q4_K_M instead of higher precision models
- Use smaller model variants (7B instead of 13B)
GPU Configuration
Problem: GPU not being used effectively
Solutions:
- Configure GPU layers:
{ "n_gpu_layers": 35 }
Advanced Instance Issues
Problem: Complex model loading, performance, or compatibility issues
Since llamactl uses llama-server under the hood, many instance-related issues are actually llama.cpp issues. For advanced troubleshooting check llama.cpp resources:
- llama.cpp Documentation: https://github.com/ggml-org/llama.cpp
Testing directly with llama-server:
# Test your model and parameters directly with llama-server
llama-server --model /path/to/model.gguf --port 8081 --n-gpu-layers 35
This helps determine if the issue is with llamactl or with the underlying llama.cpp/llama-server.
API and Network Issues
CORS Errors
Problem: Web UI shows CORS errors in browser console
Solutions:
- Configure allowed origins:
server: allowed_origins: - "http://localhost:3000" - "https://yourdomain.com"
Authentication Issues
Problem: API requests failing with authentication errors
Solutions:
-
Disable authentication temporarily:
auth: require_management_auth: false require_inference_auth: false -
Configure API keys:
auth: management_keys: - "your-management-key" inference_keys: - "your-inference-key" -
Use correct Authorization header:
curl -H "Authorization: Bearer your-api-key" \ http://localhost:8080/api/v1/instances
Remote Node Issues
Node Configuration
Problem: Remote instances not appearing or cannot be managed
Solutions:
-
Verify node configuration:
local_node: "main" # Must match a key in nodes map nodes: main: address: "" # Empty for local node worker1: address: "http://worker1.internal:8080" api_key: "secure-key" # Must match worker1's management key -
Check node name consistency:
local_nodeon each node must match what other nodes call it- Node names are case-sensitive
-
Test remote node connectivity:
curl -H "Authorization: Bearer remote-node-key" \ http://remote-node:8080/api/v1/instances
Debugging and Logs
Viewing Instance Logs
# Get instance logs via API
curl http://localhost:8080/api/v1/instances/{name}/logs
# Or check log files directly
tail -f ~/.local/share/llamactl/logs/{instance-name}.log
Enable Debug Logging
export LLAMACTL_LOG_LEVEL=debug
llamactl
Getting Help
When reporting issues, include:
-
System information:
llamactl --version -
Configuration file (remove sensitive keys)
-
Relevant log output
-
Steps to reproduce the issue