Skip to content

Commit e45656f

Browse files
authored
Removed unused properties from OpenAI inference requests (#25008)
1 parent 0762e82 commit e45656f

File tree

6 files changed

+0
-30
lines changed

6 files changed

+0
-30
lines changed

specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-03-01-preview/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -203,11 +203,6 @@
203203
"type": "string",
204204
"nullable": true
205205
},
206-
"cache_level": {
207-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
208-
"type": "integer",
209-
"nullable": true
210-
},
211206
"presence_penalty": {
212207
"description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.",
213208
"type": "number",

specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-06-01-preview/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -203,11 +203,6 @@
203203
"type": "string",
204204
"nullable": true
205205
},
206-
"cache_level": {
207-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
208-
"type": "integer",
209-
"nullable": true
210-
},
211206
"presence_penalty": {
212207
"description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.",
213208
"type": "number",

specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-03-15-preview/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -173,11 +173,6 @@
173173
"type": "string",
174174
"nullable": true
175175
},
176-
"cache_level": {
177-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
178-
"type": "integer",
179-
"nullable": true
180-
},
181176
"presence_penalty": {
182177
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
183178
"type": "number",

specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-06-01-preview/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -165,11 +165,6 @@
165165
"type": "string",
166166
"nullable": true
167167
},
168-
"cache_level": {
169-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
170-
"type": "integer",
171-
"nullable": true
172-
},
173168
"presence_penalty": {
174169
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
175170
"type": "number",

specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2022-12-01/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -173,11 +173,6 @@
173173
"type": "string",
174174
"nullable": true
175175
},
176-
"cache_level": {
177-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
178-
"type": "integer",
179-
"nullable": true
180-
},
181176
"presence_penalty": {
182177
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
183178
"type": "number",

specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2023-05-15/inference.json

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -173,11 +173,6 @@
173173
"type": "string",
174174
"nullable": true
175175
},
176-
"cache_level": {
177-
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
178-
"type": "integer",
179-
"nullable": true
180-
},
181176
"presence_penalty": {
182177
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
183178
"type": "number",

0 commit comments

Comments
 (0)