generated from SAP/repository-template
-
Notifications
You must be signed in to change notification settings - Fork 30
Open
Labels
Description
Describe the Problem
Hi there,
i am struggling to get LLMS running that arent from OpenAI with your langchain module. Per default you seem to filter out LLMS which are in a different "excecutableId" than "azure-openai"
This makes using different Models seemingly impossible (for now)
Propose a Solution
I would suggest that you allow to pass a excecutableId when initializing a langchain chat client
const chatClient = new AzureOpenAiChatClient({
modelName: 'meta--llama3.1-70b-instruct',
excecutableID:'aicore-opensource'
});this should allow the usage of other models quite easily without having to rewrite much.
The chat-completion api ( for example when using llama or mixtral) seems identical to payloads working with OpenAI models.
POST {baseurl}/v2/inference/deployments/{deploymentID}/chat/completions
Body for GPT 4o:
{
"messages": [
{
"role": "user",
"content": "test"
}
],
"model": "gpt-4o",
"max_tokens": 100,
"temperature": 0.0,
"frequency_penalty": 0,
"presence_penalty": 0,
"stop": "null"
}Body For Llama 3.1 Instruct
{
"messages": [
{
"role": "user",
"content": "test"
}
],
"model": "meta--llama3.1-70b-instruct",
"max_tokens": 100,
"temperature": 0.0,
"frequency_penalty": 0,
"presence_penalty": 0,
"stop": "null"
}Describe Alternatives
No response
Affected Development Phase
Development
Impact
Inconvenience
Timeline
No response
Additional Context
No response