Evaluations
All evaluations
All datasets
Filter
inputs
output
model_latency
weave_evaluation_scorer
match
Trace
Feedback
Status
model
self
mean
mean
<bound method TypedPredictor.forward of TypedPredictor(PairwiseSignature(input -> output
instructions='Given the fields `input`, produce the fields `output`.'
input = Field(annotation=Input required=True json_schema_extra={'__dspy_field_type': 'input', 'prefix': 'Input:', 'desc': '${input}'})
output = Field(annotation=Output required=True json_schema_extra={'__dspy_field_type': 'output', 'prefix': 'Output:', 'desc': '${output}'})
))>
N/A
N/A
<bound method PairwiseEvaluationModule.forward of prog.predictor = Predict(PairwiseSignature(input -> output
instructions='Given the fields `input`, produce the fields `output`.'
input = Field(annotation=Input required=True json_schema_extra={'__dspy_field_type': 'input', 'prefix': 'Input:', 'desc': '${input}'})
output = Field(annotation=Output required=True json_schema_extra={'__dspy_field_type': 'output', 'prefix': 'Output:', 'desc': '${output}'})
))>
N/A
N/A
<bound method TypedPredictor.forward of TypedPredictor(PairwiseSignature(input -> output
instructions='Given the fields `input`, produce the fields `output`.'
input = Field(annotation=Input required=True json_schema_extra={'__dspy_field_type': 'input', 'prefix': 'Input:', 'desc': '${input}'})
output = Field(annotation=Output required=True json_schema_extra={'__dspy_field_type': 'output', 'prefix': 'Output:', 'desc': '${output}'})
))>
N/A
N/A
TypedPredictor(PairwiseSignature(input -> output
instructions='Given the fields `input`, produce the fields `output`.'
input = Field(annotation=Input required=True json_schema_extra={'__dspy_field_type': 'input', 'prefix': 'Input:', 'desc': '${input}'})
output = Field(annotation=Output required=True json_schema_extra={'__dspy_field_type': 'output', 'prefix': 'Output:', 'desc': '${output}'})
))
N/A
N/A
1-9 of 9
Per page:
50
Charts
3
Score summary
2
General
Cost
$0.00
↗+ $0.00
Tokens
33.65K
↗+ 675
Latency
4.21s
↗+ 4.21s
weave_evaluation_scorer
match.mean
0.78
↗+ 0.08
model_latency
mean
0.97
↘- 3.42