Update API endpoints in managing instances and quick start documentation

This commit is contained in:
2025-10-26 16:35:58 +01:00
parent 59c954811d
commit eac4f834c0
2 changed files with 16 additions and 16 deletions

View File

@@ -69,7 +69,7 @@ Each instance is displayed as a card showing:
```bash ```bash
# Create llama.cpp instance with local model file # Create llama.cpp instance with local model file
curl -X POST http://localhost:8080/api/instances/my-llama-instance \ curl -X POST http://localhost:8080/api/v1/instances/my-llama-instance \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "llama_cpp", "backend_type": "llama_cpp",
@@ -83,7 +83,7 @@ curl -X POST http://localhost:8080/api/instances/my-llama-instance \
}' }'
# Create MLX instance (macOS only) # Create MLX instance (macOS only)
curl -X POST http://localhost:8080/api/instances/my-mlx-instance \ curl -X POST http://localhost:8080/api/v1/instances/my-mlx-instance \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "mlx_lm", "backend_type": "mlx_lm",
@@ -99,7 +99,7 @@ curl -X POST http://localhost:8080/api/instances/my-mlx-instance \
}' }'
# Create vLLM instance # Create vLLM instance
curl -X POST http://localhost:8080/api/instances/my-vllm-instance \ curl -X POST http://localhost:8080/api/v1/instances/my-vllm-instance \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "vllm", "backend_type": "vllm",
@@ -119,7 +119,7 @@ curl -X POST http://localhost:8080/api/instances/my-vllm-instance \
}' }'
# Create llama.cpp instance with HuggingFace model # Create llama.cpp instance with HuggingFace model
curl -X POST http://localhost:8080/api/instances/gemma-3-27b \ curl -X POST http://localhost:8080/api/v1/instances/gemma-3-27b \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "llama_cpp", "backend_type": "llama_cpp",
@@ -132,7 +132,7 @@ curl -X POST http://localhost:8080/api/instances/gemma-3-27b \
}' }'
# Create instance on specific remote node # Create instance on specific remote node
curl -X POST http://localhost:8080/api/instances/remote-llama \ curl -X POST http://localhost:8080/api/v1/instances/remote-llama \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "llama_cpp", "backend_type": "llama_cpp",
@@ -144,7 +144,7 @@ curl -X POST http://localhost:8080/api/instances/remote-llama \
}' }'
# Create instance on multiple nodes for high availability # Create instance on multiple nodes for high availability
curl -X POST http://localhost:8080/api/instances/multi-node-llama \ curl -X POST http://localhost:8080/api/v1/instances/multi-node-llama \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "llama_cpp", "backend_type": "llama_cpp",
@@ -166,7 +166,7 @@ curl -X POST http://localhost:8080/api/instances/multi-node-llama \
**Via API** **Via API**
```bash ```bash
curl -X POST http://localhost:8080/api/instances/{name}/start curl -X POST http://localhost:8080/api/v1/instances/{name}/start
``` ```
## Stop Instance ## Stop Instance
@@ -177,7 +177,7 @@ curl -X POST http://localhost:8080/api/instances/{name}/start
**Via API** **Via API**
```bash ```bash
curl -X POST http://localhost:8080/api/instances/{name}/stop curl -X POST http://localhost:8080/api/v1/instances/{name}/stop
``` ```
## Edit Instance ## Edit Instance
@@ -192,7 +192,7 @@ curl -X POST http://localhost:8080/api/instances/{name}/stop
Modify instance settings: Modify instance settings:
```bash ```bash
curl -X PUT http://localhost:8080/api/instances/{name} \ curl -X PUT http://localhost:8080/api/v1/instances/{name} \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_options": { "backend_options": {
@@ -218,7 +218,7 @@ Check instance status in real-time:
```bash ```bash
# Get instance details # Get instance details
curl http://localhost:8080/api/instances/{name}/logs curl http://localhost:8080/api/v1/instances/{name}/logs
``` ```
## Delete Instance ## Delete Instance
@@ -230,7 +230,7 @@ curl http://localhost:8080/api/instances/{name}/logs
**Via API** **Via API**
```bash ```bash
curl -X DELETE http://localhost:8080/api/instances/{name} curl -X DELETE http://localhost:8080/api/v1/instances/{name}
``` ```
## Instance Proxy ## Instance Proxy
@@ -239,7 +239,7 @@ Llamactl proxies all requests to the underlying backend instances (llama-server,
```bash ```bash
# Get instance details # Get instance details
curl http://localhost:8080/api/instances/{name}/proxy/ curl http://localhost:8080/api/v1/instances/{name}/proxy/
``` ```
All backends provide OpenAI-compatible endpoints. Check the respective documentation: All backends provide OpenAI-compatible endpoints. Check the respective documentation:
@@ -258,6 +258,6 @@ All backends provide OpenAI-compatible endpoints. Check the respective documenta
Check the health status of your instances: Check the health status of your instances:
```bash ```bash
curl http://localhost:8080/api/instances/{name}/proxy/health curl http://localhost:8080/api/v1/instances/{name}/proxy/health
``` ```

View File

@@ -132,10 +132,10 @@ You can also manage instances via the REST API:
```bash ```bash
# List all instances # List all instances
curl http://localhost:8080/api/instances curl http://localhost:8080/api/v1/instances
# Create a new llama.cpp instance # Create a new llama.cpp instance
curl -X POST http://localhost:8080/api/instances/my-model \ curl -X POST http://localhost:8080/api/v1/instances/my-model \
-H "Content-Type: application/json" \ -H "Content-Type: application/json" \
-d '{ -d '{
"backend_type": "llama_cpp", "backend_type": "llama_cpp",
@@ -145,7 +145,7 @@ curl -X POST http://localhost:8080/api/instances/my-model \
}' }'
# Start an instance # Start an instance
curl -X POST http://localhost:8080/api/instances/my-model/start curl -X POST http://localhost:8080/api/v1/instances/my-model/start
``` ```
## OpenAI Compatible API ## OpenAI Compatible API