From 4f94f63de30da1e9ebaa5ea9d2e2656641b8dc7a Mon Sep 17 00:00:00 2001 From: LordMathis Date: Sun, 26 Oct 2025 17:19:53 +0100 Subject: [PATCH] Minor docs improvements --- docs/configuration.md | 2 +- docs/quick-start.md | 8 +++++--- docs/troubleshooting.md | 2 +- 3 files changed, 7 insertions(+), 5 deletions(-) diff --git a/docs/configuration.md b/docs/configuration.md index 6f9ee98..c271f29 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -80,7 +80,7 @@ nodes: # Node configuration for multi-node deployment ### Configuration File Locations -Configuration files are searched in the following locations (in order of precedence): +Configuration files are searched in the following locations (in order of precedence, first found is used): **Linux:** - `./llamactl.yaml` or `./config.yaml` (current directory) diff --git a/docs/quick-start.md b/docs/quick-start.md index b4867b5..f65f9b2 100644 --- a/docs/quick-start.md +++ b/docs/quick-start.md @@ -2,6 +2,8 @@ This guide will help you get Llamactl up and running in just a few minutes. +**Before you begin:** Ensure you have at least one backend installed (llama.cpp, MLX, or vLLM). See the [Installation Guide](installation.md#prerequisites) for backend setup. + ## Core Concepts Before you start, let's clarify a few key terms: @@ -53,7 +55,7 @@ llamactl Llamactl server listening on 0.0.0.0:8080 ``` -Copy the **Management API Key** from the terminal - you'll need it to access the web UI. +Copy the **Management** and **Inference** API Keys from the terminal - you'll need them to access the web UI and make inference requests. By default, Llamactl will start on `http://localhost:8080`. @@ -143,7 +145,7 @@ Here are basic example configurations for each backend: } ``` -**Multi-node deployment example:** +**Remote node deployment example:** ```json { "name": "distributed-model", @@ -152,7 +154,7 @@ Here are basic example configurations for each backend: "model": "/path/to/model.gguf", "gpu_layers": 32 }, - "nodes": ["worker1", "worker2"] + "nodes": ["worker1"] } ``` diff --git a/docs/troubleshooting.md b/docs/troubleshooting.md index e7d9d80..1f33c0a 100644 --- a/docs/troubleshooting.md +++ b/docs/troubleshooting.md @@ -85,7 +85,7 @@ llama-server --model /path/to/model.gguf --port 8081 mlx_lm.server --model mlx-community/Mistral-7B-Instruct-v0.3-4bit --port 8081 # vLLM -python -m vllm.entrypoints.openai.api_server --model microsoft/DialoGPT-medium --port 8081 +vllm serve microsoft/DialoGPT-medium --port 8081 ``` ## API and Network Issues