Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Changes
mfpbench
to now support specifying theMetric
to use for a benchmark, which will be returned byresult.error
orresult.score
. I.e.To help with this, you can see the metrics defined by a benchmark as so:
You can also specify the
value_metric=
in thequery()
callWhen a
Metric
is bounded above and below, the.score
and.error
will always be bounded in(0, 1)
as we can normalize the objective.To view the optimum of all metrics, you can use
benchmark.metric_optimums: dict[str, Metric.Value]
For tabular benchmarks, there is also
.table_optimums
which acts much the same but does not use theoretical optimums, instead using the best value seen in the table provided.Breaking
Usage of
config.dict()
orresult.dict()
should now beconfig.as_dict()
orresult.as_dict()
. You can also now obtain the id in when callingconfig.as_dict(with_id=True)
which defaults toFalse
.