Skip to content

Commit

Permalink
Removed unused properties from OpenAI inference requests (#25008)
Browse files Browse the repository at this point in the history
  • Loading branch information
bexxx authored Aug 29, 2023
1 parent 0762e82 commit e45656f
Show file tree
Hide file tree
Showing 6 changed files with 0 additions and 30 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -203,11 +203,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.",
"type": "number",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -203,11 +203,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "How much to penalize new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim. Has minimum of -2 and maximum of 2.",
"type": "number",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -173,11 +173,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
"type": "number",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -165,11 +165,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
"type": "number",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -173,11 +173,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
"type": "number",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -173,11 +173,6 @@
"type": "string",
"nullable": true
},
"cache_level": {
"description": "can be used to disable any server-side caching, 0=no cache, 1=prompt prefix enabled, 2=full cache",
"type": "integer",
"nullable": true
},
"presence_penalty": {
"description": "Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics.",
"type": "number",
Expand Down

0 comments on commit e45656f

Please sign in to comment.