Skip to content

Commit 7795dae

Browse files
author
Paweł Kędzia
committed
Merge branch 'features/lib'
2 parents e622709 + 8b49e5d commit 7795dae

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

56 files changed

+2473
-341
lines changed

.version

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
0.2.0

CHANGELOG.md

Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,11 @@
11
## Changelog
22

3-
| Version | Changelog |
4-
|---------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
5-
| 0.0.1 | Initialization, License, setup, interface for each endpoint and sample `ping` EP. Autoloader of builtin endpoints and for the future implementations. |
6-
| 0.0.2 | Add base models for api call (module `llm_proxy_rest.data_models` with `error.py` handling. Decorators to check required params and to measure the response time. |
7-
| 0.0.3 | Proper `AutoLoading` for each found endpoint. Implementation of `ApiTypesDispatcher`, `ApiModelConfig`, `ModelHandler`. Ollama endpoints: `/`, `tags`. Added endpoint to full proxy with params. Streaming in case when external api provides stream. |
8-
| 0.0.4 | All llama-service endpoints are refactored to `llm-proxy-api`. Refactoring base `ep_run` method. Proper handling system message, prompt name, model etc. |
9-
| 0.1.0 | Repository name changed from `llm-proxy-api` to `llm-router`. Added class `HttpRequestExecutor` to handle http requests from `EndpointWithHttpRequestI`. Handled routing between any models: `openai -> ollama` and `ollama -> openai` |
10-
| 0.1.1 | Prometheus metrics logging. Workers/Threads/Workers class is able to set by environments. Streaming fixes. Multi-providers for single model with default-balanced strategy. |
3+
| Version | Changelog |
4+
|---------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
5+
| 0.0.1 | Initialization, License, setup, interface for each endpoint and sample `ping` EP. Autoloader of builtin endpoints and for the future implementations. |
6+
| 0.0.2 | Add base models for api call (module `llm_proxy_rest.data_models` with `error.py` handling. Decorators to check required params and to measure the response time. |
7+
| 0.0.3 | Proper `AutoLoading` for each found endpoint. Implementation of `ApiTypesDispatcher`, `ApiModelConfig`, `ModelHandler`. Ollama endpoints: `/`, `tags`. Added endpoint to full proxy with params. Streaming in case when external api provides stream. |
8+
| 0.0.4 | All llama-service endpoints are refactored to `llm-proxy-api`. Refactoring base `ep_run` method. Proper handling system message, prompt name, model etc. |
9+
| 0.1.0 | Repository name changed from `llm-proxy-api` to `llm-router`. Added class `HttpRequestExecutor` to handle http requests from `EndpointWithHttpRequestI`. Handled routing between any models: `openai -> ollama` and `ollama -> openai` |
10+
| 0.1.1 | Prometheus metrics logging. Workers/Threads/Workers class is able to set by environments. Streaming fixes. Multi-providers for single model with default-balanced strategy. |
11+
| 0.2.0 | Add balancing strategies: `balanced`, `weighted`, `dynamic_weighted` and `first_available` which works for streaming and non streaming requests. Included Prometheus metrics logging via `/metrics` endpoint. First stage of `llm_router_lib` library, to simply usage of `llm-router-api`. |

0 commit comments

Comments
 (0)