From e45656f49582887382dec9d3c31bc661796a4fce Mon Sep 17 00:00:00 2001 From: Ralf Beckers Date: Tue, 29 Aug 2023 18:10:29 +0200 Subject: [PATCH] Removed unused properties from OpenAI inference requests (#25008) --- .../inference/preview/2022-03-01-preview/inference.json | 5 ----- .../inference/preview/2022-06-01-preview/inference.json | 5 ----- .../inference/preview/2023-03-15-preview/inference.json | 5 ----- .../inference/preview/2023-06-01-preview/inference.json | 5 ----- .../AzureOpenAI/inference/stable/2022-12-01/inference.json | 5 ----- .../AzureOpenAI/inference/stable/2023-05-15/inference.json | 5 ----- 6 files changed, 30 deletions(-) diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-03-01-preview/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-03-01-preview/inference.json index ee6bbead016a..6b600ec2e724 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-03-01-preview/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-03-01-preview/inference.json @@ -203,11 +203,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.", "type": "number", diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-06-01-preview/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-06-01-preview/inference.json index b27e08ec3682..fb3724a60edd 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-06-01-preview/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2022-06-01-preview/inference.json @@ -203,11 +203,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.", "type": "number", diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-03-15-preview/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-03-15-preview/inference.json index 263c88396230..a958d8d19462 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-03-15-preview/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-03-15-preview/inference.json @@ -173,11 +173,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.", "type": "number", diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-06-01-preview/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-06-01-preview/inference.json index 680d495a3306..a422b785611e 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-06-01-preview/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2023-06-01-preview/inference.json @@ -165,11 +165,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.", "type": "number", diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2022-12-01/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2022-12-01/inference.json index 490205391f67..296ac46fb5dd 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2022-12-01/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2022-12-01/inference.json @@ -173,11 +173,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.", "type": "number", diff --git a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2023-05-15/inference.json b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2023-05-15/inference.json index 1f4b77da87f7..92096f52e60a 100644 --- a/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2023-05-15/inference.json +++ b/specification/cognitiveservices/data-plane/AzureOpenAI/inference/stable/2023-05-15/inference.json @@ -173,11 +173,6 @@ "type": "string", "nullable": true }, - "cache_level": { - "description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache", - "type": "integer", - "nullable": true - }, "presence_penalty": { "description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.", "type": "number",