Skip to content

Commit d3582ca

Browse files
author
Paweł Kędzia
committed
Update CHANGELOG 0.2.2: fix routing with vLLM
1 parent e6b413b commit d3582ca

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

CHANGELOG.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,4 +10,4 @@
1010
| 0.1.1 | Prometheus metrics logging. Workers/Threads/Workers class is able to set by environments. Streaming fixes. Multi-providers for single model with default-balanced strategy. |
1111
| 0.2.0 | Add balancing strategies: `balanced`, `weighted`, `dynamic_weighted` and `first_available` which works for streaming and non streaming requests. Included Prometheus metrics logging via `/metrics` endpoint. First stage of `llm_router_lib` library, to simply usage of `llm-router-api`. |
1212
| 0.2.1 | Fix stream: OpenAI->Ollama, Ollama->OpenAI. Add Redis caching of availability of model providers (when using `first_available` strategy). Add `llm_router_web` module with simple flask-based frontend to manage llm-router config files. |
13-
| 0.2.2 | Update dockerfile and requirements. |
13+
| 0.2.2 | Update dockerfile and requirements. Fix routing with vLLM. |

0 commit comments

Comments
 (0)