From 8154b8d0ab6a332904084303535a4271ed98d1ce Mon Sep 17 00:00:00 2001 From: LordMathis Date: Wed, 6 Aug 2025 18:49:36 +0200 Subject: [PATCH] Fix temp in tests --- pkg/backends/llamacpp/llama_test.go | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/pkg/backends/llamacpp/llama_test.go b/pkg/backends/llamacpp/llama_test.go index 239703b..9c1162e 100644 --- a/pkg/backends/llamacpp/llama_test.go +++ b/pkg/backends/llamacpp/llama_test.go @@ -109,13 +109,13 @@ func TestBuildCommandArgs_NumericFields(t *testing.T) { args := options.BuildCommandArgs() expectedPairs := map[string]string{ - "--port": "8080", - "--threads": "4", - "--ctx-size": "2048", - "--gpu-layers": "16", - "--temperature": "0.7", - "--top-k": "40", - "--top-p": "0.9", + "--port": "8080", + "--threads": "4", + "--ctx-size": "2048", + "--gpu-layers": "16", + "--temp": "0.7", + "--top-k": "40", + "--top-p": "0.9", } for flag, expectedValue := range expectedPairs { @@ -231,7 +231,7 @@ func TestUnmarshalJSON_StandardFields(t *testing.T) { "verbose": true, "ctx_size": 4096, "gpu_layers": 32, - "temperature": 0.7 + "temp": 0.7 }` var options llamacpp.LlamaServerOptions