Minimalist Python framework for AI agents logic-only coding with streaming, tool calls, and multi-LLM provider support.
pip install open-taranis --upgradeimport open_taranis as T
client = T.clients.openrouter("api_key")
messages = [
T.create_user_prompt("Tell me about yourself")
]
stream = T.clients.openrouter_request(
client=client,
messages=messages,
model="qwen/qwen3-4b:free",
)
print("assistant : ",end="")
for token, tool, tool_bool in T.handle_streaming(stream) :
if token :
print(token, end="")- Base of the docs (coding some things before the real docs)
- v0.0.1: start
- v0.0.x: Add and confirm other API providers (in the cloud, not locally)
- v0.1.x: Functionality verifications in examples
- > v0.2.0: Add features for logic-only coding approach
- v0.6.x: Add llama.cpp as backend in addition to APIs
- v0.7.x: Add reverse proxy + server to create a dedicated full relay/backend (like OpenRouter), framework usable as server and client
- v0.8.x: Add PyTorch as backend with
transformersto deploy a remote server - v0.9.x: Total reduction of dependencies for built-in functions (unless counter-optimizations)
- v1.0.0: First complete version in Python without dependencies
- v0.0.4 : Add xai and groq provider
- v0.0.6 : Add huggingface provider and args for clients.veniceai_request
- v0.1.0 : Start the docs, add update-checker and preparing for the continuation of the project...