Capitalize godoc tags

This commit is contained in:
2025-10-26 16:49:27 +01:00
parent 6c522a2199
commit 249ff2a7aa
8 changed files with 125 additions and 125 deletions

View File

@@ -28,7 +28,7 @@ const docTemplate = `{
], ],
"description": "Returns a list of available devices for the llama server", "description": "Returns a list of available devices for the llama server",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "List available devices for llama server", "summary": "List available devices for llama server",
"responses": { "responses": {
@@ -56,7 +56,7 @@ const docTemplate = `{
], ],
"description": "Returns the help text for the llama server command", "description": "Returns the help text for the llama server command",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Get help for llama server", "summary": "Get help for llama server",
"responses": { "responses": {
@@ -90,7 +90,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse llama-server command", "summary": "Parse llama-server command",
"parameters": [ "parameters": [
@@ -141,7 +141,7 @@ const docTemplate = `{
], ],
"description": "Returns the version of the llama server command", "description": "Returns the version of the llama server command",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Get version of llama server", "summary": "Get version of llama server",
"responses": { "responses": {
@@ -175,7 +175,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse mlx_lm.server command", "summary": "Parse mlx_lm.server command",
"parameters": [ "parameters": [
@@ -223,7 +223,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse vllm serve command", "summary": "Parse vllm serve command",
"parameters": [ "parameters": [
@@ -265,7 +265,7 @@ const docTemplate = `{
], ],
"description": "Returns a list of all instances managed by the server", "description": "Returns a list of all instances managed by the server",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "List all instances", "summary": "List all instances",
"responses": { "responses": {
@@ -296,7 +296,7 @@ const docTemplate = `{
], ],
"description": "Returns the details of a specific instance by name", "description": "Returns the details of a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Get details of a specific instance", "summary": "Get details of a specific instance",
"parameters": [ "parameters": [
@@ -340,7 +340,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Update an instance's configuration", "summary": "Update an instance's configuration",
"parameters": [ "parameters": [
@@ -393,7 +393,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Create and start a new instance", "summary": "Create and start a new instance",
"parameters": [ "parameters": [
@@ -443,7 +443,7 @@ const docTemplate = `{
], ],
"description": "Stops and removes a specific instance by name", "description": "Stops and removes a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Delete an instance", "summary": "Delete an instance",
"parameters": [ "parameters": [
@@ -483,7 +483,7 @@ const docTemplate = `{
], ],
"description": "Returns the logs from a specific instance by name with optional line limit", "description": "Returns the logs from a specific instance by name with optional line limit",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Get logs from a specific instance", "summary": "Get logs from a specific instance",
"parameters": [ "parameters": [
@@ -532,7 +532,7 @@ const docTemplate = `{
], ],
"description": "Forwards HTTP requests to the llama-server instance running on a specific port", "description": "Forwards HTTP requests to the llama-server instance running on a specific port",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Proxy requests to a specific instance, does not autostart instance if stopped", "summary": "Proxy requests to a specific instance, does not autostart instance if stopped",
"parameters": [ "parameters": [
@@ -576,7 +576,7 @@ const docTemplate = `{
], ],
"description": "Forwards HTTP requests to the llama-server instance running on a specific port", "description": "Forwards HTTP requests to the llama-server instance running on a specific port",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Proxy requests to a specific instance, does not autostart instance if stopped", "summary": "Proxy requests to a specific instance, does not autostart instance if stopped",
"parameters": [ "parameters": [
@@ -622,7 +622,7 @@ const docTemplate = `{
], ],
"description": "Restarts a specific instance by name", "description": "Restarts a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Restart a running instance", "summary": "Restart a running instance",
"parameters": [ "parameters": [
@@ -665,7 +665,7 @@ const docTemplate = `{
], ],
"description": "Starts a specific instance by name", "description": "Starts a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Start a stopped instance", "summary": "Start a stopped instance",
"parameters": [ "parameters": [
@@ -708,7 +708,7 @@ const docTemplate = `{
], ],
"description": "Stops a specific instance by name", "description": "Stops a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Stop a running instance", "summary": "Stop a running instance",
"parameters": [ "parameters": [
@@ -751,7 +751,7 @@ const docTemplate = `{
], ],
"description": "Returns a map of all nodes configured in the server (node name -\u003e node config)", "description": "Returns a map of all nodes configured in the server (node name -\u003e node config)",
"tags": [ "tags": [
"nodes" "Nodes"
], ],
"summary": "List all configured nodes", "summary": "List all configured nodes",
"responses": { "responses": {
@@ -782,7 +782,7 @@ const docTemplate = `{
], ],
"description": "Returns the details of a specific node by name", "description": "Returns the details of a specific node by name",
"tags": [ "tags": [
"nodes" "Nodes"
], ],
"summary": "Get details of a specific node", "summary": "Get details of a specific node",
"parameters": [ "parameters": [
@@ -831,7 +831,7 @@ const docTemplate = `{
], ],
"description": "Returns the version of the llamactl command", "description": "Returns the version of the llamactl command",
"tags": [ "tags": [
"version" "System"
], ],
"summary": "Get llamactl version", "summary": "Get llamactl version",
"responses": { "responses": {
@@ -862,7 +862,7 @@ const docTemplate = `{
"text/html" "text/html"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp UI for the instance", "summary": "Proxy requests to llama.cpp UI for the instance",
"parameters": [ "parameters": [
@@ -908,7 +908,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -955,7 +955,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1002,7 +1002,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1049,7 +1049,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1096,7 +1096,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1143,7 +1143,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1190,7 +1190,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1235,7 +1235,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1282,7 +1282,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1329,7 +1329,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1376,7 +1376,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1423,7 +1423,7 @@ const docTemplate = `{
"application/json" "application/json"
], ],
"tags": [ "tags": [
"openai" "OpenAI"
], ],
"summary": "OpenAI-compatible proxy endpoint", "summary": "OpenAI-compatible proxy endpoint",
"responses": { "responses": {
@@ -1454,7 +1454,7 @@ const docTemplate = `{
], ],
"description": "Returns a list of instances in a format compatible with OpenAI API", "description": "Returns a list of instances in a format compatible with OpenAI API",
"tags": [ "tags": [
"openai" "OpenAI"
], ],
"summary": "List instances in OpenAI-compatible format", "summary": "List instances in OpenAI-compatible format",
"responses": { "responses": {

View File

@@ -21,7 +21,7 @@
], ],
"description": "Returns a list of available devices for the llama server", "description": "Returns a list of available devices for the llama server",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "List available devices for llama server", "summary": "List available devices for llama server",
"responses": { "responses": {
@@ -49,7 +49,7 @@
], ],
"description": "Returns the help text for the llama server command", "description": "Returns the help text for the llama server command",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Get help for llama server", "summary": "Get help for llama server",
"responses": { "responses": {
@@ -83,7 +83,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse llama-server command", "summary": "Parse llama-server command",
"parameters": [ "parameters": [
@@ -134,7 +134,7 @@
], ],
"description": "Returns the version of the llama server command", "description": "Returns the version of the llama server command",
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Get version of llama server", "summary": "Get version of llama server",
"responses": { "responses": {
@@ -168,7 +168,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse mlx_lm.server command", "summary": "Parse mlx_lm.server command",
"parameters": [ "parameters": [
@@ -216,7 +216,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Backends"
], ],
"summary": "Parse vllm serve command", "summary": "Parse vllm serve command",
"parameters": [ "parameters": [
@@ -258,7 +258,7 @@
], ],
"description": "Returns a list of all instances managed by the server", "description": "Returns a list of all instances managed by the server",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "List all instances", "summary": "List all instances",
"responses": { "responses": {
@@ -289,7 +289,7 @@
], ],
"description": "Returns the details of a specific instance by name", "description": "Returns the details of a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Get details of a specific instance", "summary": "Get details of a specific instance",
"parameters": [ "parameters": [
@@ -333,7 +333,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Update an instance's configuration", "summary": "Update an instance's configuration",
"parameters": [ "parameters": [
@@ -386,7 +386,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Create and start a new instance", "summary": "Create and start a new instance",
"parameters": [ "parameters": [
@@ -436,7 +436,7 @@
], ],
"description": "Stops and removes a specific instance by name", "description": "Stops and removes a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Delete an instance", "summary": "Delete an instance",
"parameters": [ "parameters": [
@@ -476,7 +476,7 @@
], ],
"description": "Returns the logs from a specific instance by name with optional line limit", "description": "Returns the logs from a specific instance by name with optional line limit",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Get logs from a specific instance", "summary": "Get logs from a specific instance",
"parameters": [ "parameters": [
@@ -525,7 +525,7 @@
], ],
"description": "Forwards HTTP requests to the llama-server instance running on a specific port", "description": "Forwards HTTP requests to the llama-server instance running on a specific port",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Proxy requests to a specific instance, does not autostart instance if stopped", "summary": "Proxy requests to a specific instance, does not autostart instance if stopped",
"parameters": [ "parameters": [
@@ -569,7 +569,7 @@
], ],
"description": "Forwards HTTP requests to the llama-server instance running on a specific port", "description": "Forwards HTTP requests to the llama-server instance running on a specific port",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Proxy requests to a specific instance, does not autostart instance if stopped", "summary": "Proxy requests to a specific instance, does not autostart instance if stopped",
"parameters": [ "parameters": [
@@ -615,7 +615,7 @@
], ],
"description": "Restarts a specific instance by name", "description": "Restarts a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Restart a running instance", "summary": "Restart a running instance",
"parameters": [ "parameters": [
@@ -658,7 +658,7 @@
], ],
"description": "Starts a specific instance by name", "description": "Starts a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Start a stopped instance", "summary": "Start a stopped instance",
"parameters": [ "parameters": [
@@ -701,7 +701,7 @@
], ],
"description": "Stops a specific instance by name", "description": "Stops a specific instance by name",
"tags": [ "tags": [
"instances" "Instances"
], ],
"summary": "Stop a running instance", "summary": "Stop a running instance",
"parameters": [ "parameters": [
@@ -744,7 +744,7 @@
], ],
"description": "Returns a map of all nodes configured in the server (node name -\u003e node config)", "description": "Returns a map of all nodes configured in the server (node name -\u003e node config)",
"tags": [ "tags": [
"nodes" "Nodes"
], ],
"summary": "List all configured nodes", "summary": "List all configured nodes",
"responses": { "responses": {
@@ -775,7 +775,7 @@
], ],
"description": "Returns the details of a specific node by name", "description": "Returns the details of a specific node by name",
"tags": [ "tags": [
"nodes" "Nodes"
], ],
"summary": "Get details of a specific node", "summary": "Get details of a specific node",
"parameters": [ "parameters": [
@@ -824,7 +824,7 @@
], ],
"description": "Returns the version of the llamactl command", "description": "Returns the version of the llamactl command",
"tags": [ "tags": [
"version" "System"
], ],
"summary": "Get llamactl version", "summary": "Get llamactl version",
"responses": { "responses": {
@@ -855,7 +855,7 @@
"text/html" "text/html"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp UI for the instance", "summary": "Proxy requests to llama.cpp UI for the instance",
"parameters": [ "parameters": [
@@ -901,7 +901,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -948,7 +948,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -995,7 +995,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1042,7 +1042,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1089,7 +1089,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1136,7 +1136,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1183,7 +1183,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1228,7 +1228,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1275,7 +1275,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1322,7 +1322,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1369,7 +1369,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"backends" "Llama.cpp"
], ],
"summary": "Proxy requests to llama.cpp server instance", "summary": "Proxy requests to llama.cpp server instance",
"parameters": [ "parameters": [
@@ -1416,7 +1416,7 @@
"application/json" "application/json"
], ],
"tags": [ "tags": [
"openai" "OpenAI"
], ],
"summary": "OpenAI-compatible proxy endpoint", "summary": "OpenAI-compatible proxy endpoint",
"responses": { "responses": {
@@ -1447,7 +1447,7 @@
], ],
"description": "Returns a list of instances in a format compatible with OpenAI API", "description": "Returns a list of instances in a format compatible with OpenAI API",
"tags": [ "tags": [
"openai" "OpenAI"
], ],
"summary": "List instances in OpenAI-compatible format", "summary": "List instances in OpenAI-compatible format",
"responses": { "responses": {

View File

@@ -85,7 +85,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: List available devices for llama server summary: List available devices for llama server
tags: tags:
- backends - Backends
/api/v1/backends/llama-cpp/help: /api/v1/backends/llama-cpp/help:
get: get:
description: Returns the help text for the llama server command description: Returns the help text for the llama server command
@@ -102,7 +102,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get help for llama server summary: Get help for llama server
tags: tags:
- backends - Backends
/api/v1/backends/llama-cpp/parse-command: /api/v1/backends/llama-cpp/parse-command:
post: post:
consumes: consumes:
@@ -138,7 +138,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Parse llama-server command summary: Parse llama-server command
tags: tags:
- backends - Backends
/api/v1/backends/llama-cpp/version: /api/v1/backends/llama-cpp/version:
get: get:
description: Returns the version of the llama server command description: Returns the version of the llama server command
@@ -155,7 +155,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get version of llama server summary: Get version of llama server
tags: tags:
- backends - Backends
/api/v1/backends/mlx/parse-command: /api/v1/backends/mlx/parse-command:
post: post:
consumes: consumes:
@@ -185,7 +185,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Parse mlx_lm.server command summary: Parse mlx_lm.server command
tags: tags:
- backends - Backends
/api/v1/backends/vllm/parse-command: /api/v1/backends/vllm/parse-command:
post: post:
consumes: consumes:
@@ -215,7 +215,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Parse vllm serve command summary: Parse vllm serve command
tags: tags:
- backends - Backends
/api/v1/instances: /api/v1/instances:
get: get:
description: Returns a list of all instances managed by the server description: Returns a list of all instances managed by the server
@@ -234,7 +234,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: List all instances summary: List all instances
tags: tags:
- instances - Instances
/api/v1/instances/{name}: /api/v1/instances/{name}:
delete: delete:
description: Stops and removes a specific instance by name description: Stops and removes a specific instance by name
@@ -259,7 +259,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Delete an instance summary: Delete an instance
tags: tags:
- instances - Instances
get: get:
description: Returns the details of a specific instance by name description: Returns the details of a specific instance by name
parameters: parameters:
@@ -285,7 +285,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get details of a specific instance summary: Get details of a specific instance
tags: tags:
- instances - Instances
post: post:
consumes: consumes:
- application/json - application/json
@@ -319,7 +319,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Create and start a new instance summary: Create and start a new instance
tags: tags:
- instances - Instances
put: put:
consumes: consumes:
- application/json - application/json
@@ -353,7 +353,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Update an instance's configuration summary: Update an instance's configuration
tags: tags:
- instances - Instances
/api/v1/instances/{name}/logs: /api/v1/instances/{name}/logs:
get: get:
description: Returns the logs from a specific instance by name with optional description: Returns the logs from a specific instance by name with optional
@@ -385,7 +385,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get logs from a specific instance summary: Get logs from a specific instance
tags: tags:
- instances - Instances
/api/v1/instances/{name}/proxy: /api/v1/instances/{name}/proxy:
get: get:
description: Forwards HTTP requests to the llama-server instance running on description: Forwards HTTP requests to the llama-server instance running on
@@ -416,7 +416,7 @@ paths:
summary: Proxy requests to a specific instance, does not autostart instance summary: Proxy requests to a specific instance, does not autostart instance
if stopped if stopped
tags: tags:
- instances - Instances
post: post:
description: Forwards HTTP requests to the llama-server instance running on description: Forwards HTTP requests to the llama-server instance running on
a specific port a specific port
@@ -446,7 +446,7 @@ paths:
summary: Proxy requests to a specific instance, does not autostart instance summary: Proxy requests to a specific instance, does not autostart instance
if stopped if stopped
tags: tags:
- instances - Instances
/api/v1/instances/{name}/restart: /api/v1/instances/{name}/restart:
post: post:
description: Restarts a specific instance by name description: Restarts a specific instance by name
@@ -473,7 +473,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Restart a running instance summary: Restart a running instance
tags: tags:
- instances - Instances
/api/v1/instances/{name}/start: /api/v1/instances/{name}/start:
post: post:
description: Starts a specific instance by name description: Starts a specific instance by name
@@ -500,7 +500,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Start a stopped instance summary: Start a stopped instance
tags: tags:
- instances - Instances
/api/v1/instances/{name}/stop: /api/v1/instances/{name}/stop:
post: post:
description: Stops a specific instance by name description: Stops a specific instance by name
@@ -527,7 +527,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Stop a running instance summary: Stop a running instance
tags: tags:
- instances - Instances
/api/v1/nodes: /api/v1/nodes:
get: get:
description: Returns a map of all nodes configured in the server (node name description: Returns a map of all nodes configured in the server (node name
@@ -547,7 +547,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: List all configured nodes summary: List all configured nodes
tags: tags:
- nodes - Nodes
/api/v1/nodes/{name}: /api/v1/nodes/{name}:
get: get:
description: Returns the details of a specific node by name description: Returns the details of a specific node by name
@@ -578,7 +578,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get details of a specific node summary: Get details of a specific node
tags: tags:
- nodes - Nodes
/api/v1/version: /api/v1/version:
get: get:
description: Returns the version of the llamactl command description: Returns the version of the llamactl command
@@ -595,7 +595,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Get llamactl version summary: Get llamactl version
tags: tags:
- version - System
/llama-cpp/{name}/: /llama-cpp/{name}/:
get: get:
description: Proxies requests to the llama.cpp UI for the specified instance description: Proxies requests to the llama.cpp UI for the specified instance
@@ -624,7 +624,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp UI for the instance summary: Proxy requests to llama.cpp UI for the instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/apply-template: /llama-cpp/{name}/apply-template:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -655,7 +655,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/completion: /llama-cpp/{name}/completion:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -686,7 +686,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/detokenize: /llama-cpp/{name}/detokenize:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -717,7 +717,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/embeddings: /llama-cpp/{name}/embeddings:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -748,7 +748,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/infill: /llama-cpp/{name}/infill:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -779,7 +779,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/metrics: /llama-cpp/{name}/metrics:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -810,7 +810,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/props: /llama-cpp/{name}/props:
get: get:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -841,7 +841,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
it on-demand if configured it on-demand if configured
@@ -871,7 +871,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/reranking: /llama-cpp/{name}/reranking:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -902,7 +902,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/slots: /llama-cpp/{name}/slots:
get: get:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -933,7 +933,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/llama-cpp/{name}/tokenize: /llama-cpp/{name}/tokenize:
post: post:
description: Proxies requests to the specified llama.cpp server instance, starting description: Proxies requests to the specified llama.cpp server instance, starting
@@ -964,7 +964,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: Proxy requests to llama.cpp server instance summary: Proxy requests to llama.cpp server instance
tags: tags:
- backends - Llama.cpp
/v1/: /v1/:
post: post:
consumes: consumes:
@@ -987,7 +987,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: OpenAI-compatible proxy endpoint summary: OpenAI-compatible proxy endpoint
tags: tags:
- openai - OpenAI
/v1/models: /v1/models:
get: get:
description: Returns a list of instances in a format compatible with OpenAI description: Returns a list of instances in a format compatible with OpenAI
@@ -1005,7 +1005,7 @@ paths:
- ApiKeyAuth: [] - ApiKeyAuth: []
summary: List instances in OpenAI-compatible format summary: List instances in OpenAI-compatible format
tags: tags:
- openai - OpenAI
securityDefinitions: securityDefinitions:
ApiKeyAuth: ApiKeyAuth:
in: header in: header

View File

@@ -44,7 +44,7 @@ func (h *Handler) stripLlamaCppPrefix(r *http.Request, instName string) {
// LlamaCppUIProxy godoc // LlamaCppUIProxy godoc
// @Summary Proxy requests to llama.cpp UI for the instance // @Summary Proxy requests to llama.cpp UI for the instance
// @Description Proxies requests to the llama.cpp UI for the specified instance // @Description Proxies requests to the llama.cpp UI for the specified instance
// @Tags backends // @Tags Llama.cpp
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produce html // @Produce html
// @Param name query string true "Instance Name" // @Param name query string true "Instance Name"
@@ -83,7 +83,7 @@ func (h *Handler) LlamaCppUIProxy() http.HandlerFunc {
// LlamaCppProxy godoc // LlamaCppProxy godoc
// @Summary Proxy requests to llama.cpp server instance // @Summary Proxy requests to llama.cpp server instance
// @Description Proxies requests to the specified llama.cpp server instance, starting it on-demand if configured // @Description Proxies requests to the specified llama.cpp server instance, starting it on-demand if configured
// @Tags backends // @Tags Llama.cpp
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produce json // @Produce json
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
@@ -160,7 +160,7 @@ func parseHelper(w http.ResponseWriter, r *http.Request, backend interface {
// ParseLlamaCommand godoc // ParseLlamaCommand godoc
// @Summary Parse llama-server command // @Summary Parse llama-server command
// @Description Parses a llama-server command string into instance options // @Description Parses a llama-server command string into instance options
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produce json // @Produce json
@@ -190,7 +190,7 @@ func (h *Handler) ParseLlamaCommand() http.HandlerFunc {
// ParseMlxCommand godoc // ParseMlxCommand godoc
// @Summary Parse mlx_lm.server command // @Summary Parse mlx_lm.server command
// @Description Parses MLX-LM server command string into instance options // @Description Parses MLX-LM server command string into instance options
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produce json // @Produce json
@@ -219,7 +219,7 @@ func (h *Handler) ParseMlxCommand() http.HandlerFunc {
// ParseVllmCommand godoc // ParseVllmCommand godoc
// @Summary Parse vllm serve command // @Summary Parse vllm serve command
// @Description Parses a vLLM serve command string into instance options // @Description Parses a vLLM serve command string into instance options
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produce json // @Produce json
@@ -261,7 +261,7 @@ func (h *Handler) executeLlamaServerCommand(flag, errorMsg string) http.HandlerF
// LlamaServerHelpHandler godoc // LlamaServerHelpHandler godoc
// @Summary Get help for llama server // @Summary Get help for llama server
// @Description Returns the help text for the llama server command // @Description Returns the help text for the llama server command
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces text/plain // @Produces text/plain
// @Success 200 {string} string "Help text" // @Success 200 {string} string "Help text"
@@ -274,7 +274,7 @@ func (h *Handler) LlamaServerHelpHandler() http.HandlerFunc {
// LlamaServerVersionHandler godoc // LlamaServerVersionHandler godoc
// @Summary Get version of llama server // @Summary Get version of llama server
// @Description Returns the version of the llama server command // @Description Returns the version of the llama server command
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces text/plain // @Produces text/plain
// @Success 200 {string} string "Version information" // @Success 200 {string} string "Version information"
@@ -287,7 +287,7 @@ func (h *Handler) LlamaServerVersionHandler() http.HandlerFunc {
// LlamaServerListDevicesHandler godoc // LlamaServerListDevicesHandler godoc
// @Summary List available devices for llama server // @Summary List available devices for llama server
// @Description Returns a list of available devices for the llama server // @Description Returns a list of available devices for the llama server
// @Tags backends // @Tags Backends
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces text/plain // @Produces text/plain
// @Success 200 {string} string "List of devices" // @Success 200 {string} string "List of devices"

View File

@@ -16,7 +16,7 @@ import (
// ListInstances godoc // ListInstances godoc
// @Summary List all instances // @Summary List all instances
// @Description Returns a list of all instances managed by the server // @Description Returns a list of all instances managed by the server
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Success 200 {array} instance.Instance "List of instances" // @Success 200 {array} instance.Instance "List of instances"
@@ -37,7 +37,7 @@ func (h *Handler) ListInstances() http.HandlerFunc {
// CreateInstance godoc // CreateInstance godoc
// @Summary Create and start a new instance // @Summary Create and start a new instance
// @Description Creates a new instance with the provided configuration options // @Description Creates a new instance with the provided configuration options
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produces json // @Produces json
@@ -75,7 +75,7 @@ func (h *Handler) CreateInstance() http.HandlerFunc {
// GetInstance godoc // GetInstance godoc
// @Summary Get details of a specific instance // @Summary Get details of a specific instance
// @Description Returns the details of a specific instance by name // @Description Returns the details of a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
@@ -105,7 +105,7 @@ func (h *Handler) GetInstance() http.HandlerFunc {
// UpdateInstance godoc // UpdateInstance godoc
// @Summary Update an instance's configuration // @Summary Update an instance's configuration
// @Description Updates the configuration of a specific instance by name // @Description Updates the configuration of a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produces json // @Produces json
@@ -143,7 +143,7 @@ func (h *Handler) UpdateInstance() http.HandlerFunc {
// StartInstance godoc // StartInstance godoc
// @Summary Start a stopped instance // @Summary Start a stopped instance
// @Description Starts a specific instance by name // @Description Starts a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
@@ -179,7 +179,7 @@ func (h *Handler) StartInstance() http.HandlerFunc {
// StopInstance godoc // StopInstance godoc
// @Summary Stop a running instance // @Summary Stop a running instance
// @Description Stops a specific instance by name // @Description Stops a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
@@ -209,7 +209,7 @@ func (h *Handler) StopInstance() http.HandlerFunc {
// RestartInstance godoc // RestartInstance godoc
// @Summary Restart a running instance // @Summary Restart a running instance
// @Description Restarts a specific instance by name // @Description Restarts a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
@@ -239,7 +239,7 @@ func (h *Handler) RestartInstance() http.HandlerFunc {
// DeleteInstance godoc // DeleteInstance godoc
// @Summary Delete an instance // @Summary Delete an instance
// @Description Stops and removes a specific instance by name // @Description Stops and removes a specific instance by name
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
// @Success 204 "No Content" // @Success 204 "No Content"
@@ -267,7 +267,7 @@ func (h *Handler) DeleteInstance() http.HandlerFunc {
// GetInstanceLogs godoc // GetInstanceLogs godoc
// @Summary Get logs from a specific instance // @Summary Get logs from a specific instance
// @Description Returns the logs from a specific instance by name with optional line limit // @Description Returns the logs from a specific instance by name with optional line limit
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
// @Param lines query string false "Number of lines to retrieve (default: all lines)" // @Param lines query string false "Number of lines to retrieve (default: all lines)"
@@ -310,7 +310,7 @@ func (h *Handler) GetInstanceLogs() http.HandlerFunc {
// InstanceProxy godoc // InstanceProxy godoc
// @Summary Proxy requests to a specific instance, does not autostart instance if stopped // @Summary Proxy requests to a specific instance, does not autostart instance if stopped
// @Description Forwards HTTP requests to the llama-server instance running on a specific port // @Description Forwards HTTP requests to the llama-server instance running on a specific port
// @Tags instances // @Tags Instances
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Param name path string true "Instance Name" // @Param name path string true "Instance Name"
// @Success 200 "Request successfully proxied to instance" // @Success 200 "Request successfully proxied to instance"

View File

@@ -14,7 +14,7 @@ type NodeResponse struct {
// ListNodes godoc // ListNodes godoc
// @Summary List all configured nodes // @Summary List all configured nodes
// @Description Returns a map of all nodes configured in the server (node name -> node config) // @Description Returns a map of all nodes configured in the server (node name -> node config)
// @Tags nodes // @Tags Nodes
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Success 200 {object} map[string]NodeResponse "Map of nodes" // @Success 200 {object} map[string]NodeResponse "Map of nodes"
@@ -37,7 +37,7 @@ func (h *Handler) ListNodes() http.HandlerFunc {
// GetNode godoc // GetNode godoc
// @Summary Get details of a specific node // @Summary Get details of a specific node
// @Description Returns the details of a specific node by name // @Description Returns the details of a specific node by name
// @Tags nodes // @Tags Nodes
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Param name path string true "Node Name" // @Param name path string true "Node Name"

View File

@@ -25,7 +25,7 @@ type OpenAIInstance struct {
// OpenAIListInstances godoc // OpenAIListInstances godoc
// @Summary List instances in OpenAI-compatible format // @Summary List instances in OpenAI-compatible format
// @Description Returns a list of instances in a format compatible with OpenAI API // @Description Returns a list of instances in a format compatible with OpenAI API
// @Tags openai // @Tags OpenAI
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces json // @Produces json
// @Success 200 {object} OpenAIListInstancesResponse "List of OpenAI-compatible instances" // @Success 200 {object} OpenAIListInstancesResponse "List of OpenAI-compatible instances"
@@ -61,7 +61,7 @@ func (h *Handler) OpenAIListInstances() http.HandlerFunc {
// OpenAIProxy godoc // OpenAIProxy godoc
// @Summary OpenAI-compatible proxy endpoint // @Summary OpenAI-compatible proxy endpoint
// @Description Handles all POST requests to /v1/*, routing to the appropriate instance based on the request body. Requires API key authentication via the `Authorization` header. // @Description Handles all POST requests to /v1/*, routing to the appropriate instance based on the request body. Requires API key authentication via the `Authorization` header.
// @Tags openai // @Tags OpenAI
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Accept json // @Accept json
// @Produces json // @Produces json

View File

@@ -8,7 +8,7 @@ import (
// VersionHandler godoc // VersionHandler godoc
// @Summary Get llamactl version // @Summary Get llamactl version
// @Description Returns the version of the llamactl command // @Description Returns the version of the llamactl command
// @Tags version // @Tags System
// @Security ApiKeyAuth // @Security ApiKeyAuth
// @Produces text/plain // @Produces text/plain
// @Success 200 {string} string "Version information" // @Success 200 {string} string "Version information"