We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I have three possible scores: 0, 0.1, and 1 for a Python assertion, and two basic assertions.
providers: - openai:chat:gpt-4-0613 - openai:chat:gpt-4-turbo-2024-04-09 - anthropic:messages:claude-3-sonnet-20240229 defaultTest: assert: - description: was answered type: not-icontains value: cannot answer - description: has sentences type: javascript value: output.length > 20 - description: check value type: python value: file://assert.py
At the top of my promptfoo view, I see bins around 0.6 and 0.7, which isn't quite making sense to me:
promptfoo view
The request is, can we add a little description such that this figure is easy to understand.
0.6
0.7
The text was updated successfully, but these errors were encountered:
I now understand that I have three assertions:
I realized the histogram plots mean score: 0.7 = (1 + 1 + 0.1) / 3
That being said, I still think perhaps promptfoo can add a little info bubble or hover-over/tooltip that explains this.
promptfoo
Feel free to close this out if uninterested
Sorry, something went wrong.
No branches or pull requests
I have three possible scores: 0, 0.1, and 1 for a Python assertion, and two basic assertions.
At the top of my
promptfoo view
, I see bins around 0.6 and 0.7, which isn't quite making sense to me:The request is, can we add a little description such that this figure is easy to understand.
0.6
and0.7
? Is that like a sum of multiple assertions' scores?The text was updated successfully, but these errors were encountered: