How-To / FAQ
What if I do not have an LLM service to run LLM-as-judge metrics?
# BEFORE : default text metrics including those requiring target (ground_truth) and LLM-as-judge
text_metrics = dbnl.eval.metrics.text_metrics(
prediction="prediction", target="ground_truth", eval_llm_client=oai_client
)
# AFTER : remove the eval_llm_client to exclude LLM-as-judge metrics
text_metrics = dbnl.eval.metrics.text_metrics(
prediction="prediction", target="ground_truth"
)
aug_eval_df = evaluate(eval_df, text_metrics)What if I do not have ground-truth available?
# BEFORE : default text metrics, including those requiring target (ground_truth) and LLM-as-judge
text_metrics = dbnl.eval.metrics.text_metrics(
prediction="prediction", target="ground_truth", eval_llm_client=oai_client
)
# AFTER : remove the target to remove metrics that depend on that value being specified
text_metrics = dbnl.eval.metrics.text_metrics(
prediction="prediction", eval_llm_client=oai_client
)
aug_eval_df = evaluate(eval_df, text_metrics)How do I create a custom LLM-as-judge metric?
Was this helpful?

