You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
When running the evaluation a dataset using evaluate() using the similarity evaluator I have come across some scenarios where the result is not a number. How To Reproduce the bug Model config {azure_deployment= "gpt4-turbo-preview", api_version="2024-02-01"} jsonl file {"Question":"How can you get the version of the Kubernetes cluster?","Answer":"{\"code\": \"kubectl version\" }","output":"{code: kubectl version --output=json}"} Evaluate Config
Checking the actual logged value in _similarity.py suggests the actual returned value is the string 'The'.
I notice that this issue usually occurs when the answer does not match what the LLM response based on the question would be. For example, {Question: What is the capital of France?, Answer: Washington DC, }
The text was updated successfully, but these errors were encountered:
I have added to similarity.prompty the following text: "You will respond with a single digit number between 1 and 5. You will include no other text or information", and this seems to fix the issue.
@bhonris , thank you for reporting the issue and sharing a workaround. It is a known issue that some preview OpenAI models will cause NaN results. Please also try with stable version models.
Hi, we're sending this friendly reminder because we haven't heard back from you in 30 days. We need more information about this issue to help address it. Please be sure to give us your input. If we don't hear back from you within 7 days of this comment, the issue will be automatically closed. Thank you!
Describe the bug
When running the evaluation a dataset using
evaluate()
using the similarity evaluator I have come across some scenarios where the result is not a number.How To Reproduce the bug
Model config
{azure_deployment= "gpt4-turbo-preview", api_version="2024-02-01"}
jsonl file
{"Question":"How can you get the version of the Kubernetes cluster?","Answer":"{\"code\": \"kubectl version\" }","output":"{code: kubectl version --output=json}"}
Evaluate Config
Expected behavior
Value returned is number
Running Information(please complete the following information):
pf -v
:python --version
: 3.10.11Additional context
_similarity.py
suggests the actual returned value is the string 'The'.{Question: What is the capital of France?, Answer: Washington DC, }
The text was updated successfully, but these errors were encountered: