From 31312c5850d5e15aeb3f96474a3bbcb2bfb12578 Mon Sep 17 00:00:00 2001 From: Travis Wilson <35748617+trrwilson@users.noreply.github.com> Date: Fri, 3 Feb 2023 11:23:51 -0800 Subject: [PATCH] AOAI: fix CompletionsUsage model (#22439) --- .../OpenAI.Inference/models/completions.create.cadl | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/specification/cognitiveservices/OpenAI.Inference/models/completions.create.cadl b/specification/cognitiveservices/OpenAI.Inference/models/completions.create.cadl index 8093d171b5b4..fd602657dff8 100644 --- a/specification/cognitiveservices/OpenAI.Inference/models/completions.create.cadl +++ b/specification/cognitiveservices/OpenAI.Inference/models/completions.create.cadl @@ -162,10 +162,13 @@ model CompletionsLogProbs { text_offset?: int32[]; } -@doc("Measurment of the amount of tokens used in this request and response") +@doc(""" +Representation of the token counts processed for a completions request. +Counts consider all tokens across prompts, choices, choice alternates, best_of generations, and other consumers. +""") model CompletionsUsage { @doc("Number of tokens received in the completion") - completion_token: int32, + completion_tokens: int32, @doc("Number of tokens sent in the original request") prompt_tokens: int32, @doc("Total number of tokens transacted in this request/response")