Rares Vernica
09/29/2025, 7:09 PM--filter-failing but it seems broken, see [#5755](https://github.com/promptfoo/promptfoo/issues/5755). Is there a workaround?Jérémie
10/01/2025, 8:04 PMpromptfoo eval command ?
I see there is a way to update eval name from the webpage but I'm wondering how I can let my testers easily access eval results by adopting a naming convention.
Thanks for your insights.
Kind regards,
JérémieFirestorm
10/02/2025, 8:57 PMdulax
10/06/2025, 7:45 PMbeforeEach and just hits the create session endpoint, extracts the session_id from the response and passes it through the context.
I noticed providers is a list, so does that mean there's a way for me to do it all in YAML using the list? I couldn't find an example.Umut
10/07/2025, 11:55 AMGia Duc
10/09/2025, 8:31 AM[matchers.js:121] Provider vertex:text-embedding-005 is not a valid embedding provider for 'similarity check', falling back to default
This is the config in the defaultTest:
provider:
embedding:
id: vertex:text-embedding-005
AFAIK, the text-embedding-005 is available for the similarity task type: https://cloud.google.com/vertex-ai/generative-ai/docs/embeddings/task-types#assess_text_similarity
Also, the syntax is valid according to the Promptfoo document for Vertex: https://www.promptfoo.dev/docs/providers/vertex/#embedding-models
The test works fine currently but because it is using the default provider or something else? How can I use that text embedding model for my similarity assertion? Please help me take a look. Thank youAttila Horvath
10/13/2025, 10:19 AMoyebahadur
10/13/2025, 1:42 PMsinghe.
10/15/2025, 2:22 AMYang
10/15/2025, 7:36 PMapilchand
10/16/2025, 6:20 AMUmut
10/16/2025, 11:46 AMUmut
10/16/2025, 1:09 PMdulax
10/16/2025, 6:21 PMjjczopek
10/20/2025, 12:12 PMdescription - which is a ground-truth, human made description.
My agent is producing a field in output called generated_description which is of course the description agent did.
I would like to run the similar metric on the two, ideally using the default test. How can I configure the test assertion so that it would reference description from the variables of the test case instead of hard-coding it?
something like this:
defaultTest:
assert:
- type: similar
provider: azure:embeddings:text-embedding-3-small
value: {{input_state.description}}
transformm: 'output.generated_description'storied
10/20/2025, 8:26 PMpelopo
10/20/2025, 9:50 PMpromptfoo eval And immediately cancelled it. get the prompt in the database but there are no options to manually annotate them or rank them.
Basically I need the evaluation metrics even if the problem didn't run -
#evals
https://cdn.discordapp.com/attachments/1429949929895493723/1429949931329683586/image.png?ex=68f80026&is=68f6aea6&hm=baba382cc34013385292eccf32ab9408222c2a47ff0715eba7e58f97ba05d31d&
https://cdn.discordapp.com/attachments/1429949929895493723/1429949932080599221/image.png?ex=68f80026&is=68f6aea6&hm=ed959dc3f0dc443702e9aac5887b6df4b4eaccfbd2947aee4ff61e69fa5edff6&gonkm
10/21/2025, 4:57 AMdulax
10/21/2025, 9:50 PMrejectUnauthorized: false I keep getting the above error. Is there anything more I should be doing?
- id: https
config:
url: https://<myhost>:8080/chat
method: POST
tls:
certPath: 'client.cert'
keyPath: 'client.key'
caPAth: 'ca.cert'
rejectUnauthorized: false
Error
Request to https://myhost:8080/v1/run failed, retrying: TypeError: fetch failed (Cause: Error: self-signed certificate in certificate chain)pelopo
10/22/2025, 11:04 AMTeti
10/22/2025, 1:30 PMpelopo
10/23/2025, 1:48 PMpelopo
10/23/2025, 2:15 PMĐức Duy
10/23/2025, 3:28 PMcrizm
10/23/2025, 11:57 PMhaveles
10/24/2025, 2:39 PMtanktg
10/28/2025, 12:10 PMAlex1990
10/28/2025, 3:45 PMdef call_api(prompt, options=None, context=None):
.................. some logic......
data = response.json()
contexts = [source.get('content', '')
for source in data.get('sources', [])]
return {
"output": data.get('content', ''),
"context": context_text
}
and part of YAML for this metric
assert:
- type: context-relevance
contextTransform: context
value: ''
But when I tried to catch this context field from the RAG response, I got an error below
Whatever I did, I tried to use a string or array, just context or output.context, every time I had an error
Error: Failed to transform context using expression 'context': Invariant failed: contextTransform must return a string or array of strings. Got object. Check your transform expression: context
Error: Failed to transform context using expression 'context': Invariant failed: contextTransform must return a string or array of strings. Got object. Check your transform expression: context
at resolveContext (/Users/aleksandrmeskov/.npm/_npx/81bbc6515d992ace/node_modules/promptfoo/dist/src/assertions/contextUtils.js:60:19)
at async handleContextRelevance (/Users/aleksandrmeskov/.npm/_npx/81bbc6515d992ace/node_modules/promptfoo/dist/src/assertions/contextRelevance.js:23:21)
at async runAssertion (/Users/aleksandrmeskov/.npm/_npx/81bbc6515d992ace/node_modules/promptfoo/dist/src/assertions/index.js:353:24)
at async /Users/aleksandrmeskov/.npm/_npx/81bbc6515d992ace/node_modules/promptfoo/dist/src/assertions/index.js:400:24
In documentation, it looks pretty simple, but look like it doesn't work correctly
https://www.promptfoo.dev/docs/configuration/expected-outputs/model-graded/context-relevance/
Any suggestions, how I can handle that?
https://cdn.discordapp.com/attachments/1432757147405651968/1432757147648786584/image.png?ex=69023693&is=6900e513&hm=e3561b5fac664cff41e9131fc0c4327ce0fa1634c74a9240f06dff1d91c6ffb1&Elias_M2M
10/29/2025, 9:38 AMb00l_
10/29/2025, 2:36 PM