|
|
4df02a6519
|
Initial vLLM backend support
|
2025-09-19 18:05:12 +02:00 |
|
|
|
6a580667ed
|
Remove LlamaExecutable checks from default and file loading tests
|
2025-09-18 20:30:26 +02:00 |
|
|
|
5121f0e302
|
Remove PythonPath references from MlxServerOptions and related configurations
|
2025-09-17 21:59:55 +02:00 |
|
|
|
cc5d8acd92
|
Refactor instance and manager tests to use BackendConfig for LlamaExecutable and MLXLMExecutable
|
2025-09-16 21:45:50 +02:00 |
|
|
|
468688cdbc
|
Pass backend options to instances
|
2025-09-16 21:37:48 +02:00 |
|
|
|
988c4aca40
|
Add MLX backend config options
|
2025-09-16 21:14:19 +02:00 |
|
|
|
a9f1c1a619
|
Add LRU eviction configuration for instances
|
2025-08-30 22:26:02 +02:00 |
|
|
|
615c2ac54e
|
Add MaxRunningInstances to InstancesConfig and implement IsRunning method
|
2025-08-27 18:42:34 +02:00 |
|
|
|
ddb54763f6
|
Add OnDemandStartTimeout configuration and update OpenAIProxy to use it
|
2025-08-20 14:25:43 +02:00 |
|
|
|
7b4adfa0cd
|
Add DefaultOnDemandStart configuration and update instance options
|
2025-08-20 13:50:43 +02:00 |
|
|
|
ccffbca6b2
|
Add timeout check interval and update instance configuration
|
2025-08-17 19:26:21 +02:00 |
|
|
|
e2b64620b5
|
Expose version endpoint
|
2025-08-07 19:10:06 +02:00 |
|
|
|
2abe9c282e
|
Rename config and instance struct to avoid awkward naming
|
2025-08-04 19:30:50 +02:00 |
|
|
|
6a7a9a2d09
|
Split large package into subpackages
|
2025-08-04 19:23:56 +02:00 |
|