eval
evaluate()
get_column_schemas_from_dataframe_and_metrics()
get_column_schemas_from_metrics()
eval.metrics
Metric
RougeScoreType
answer_quality_llm_accuracy()
answer_quality_llm_answer_correctness()
answer_quality_llm_answer_similarity()
answer_quality_llm_coherence()
answer_quality_llm_commital()
answer_quality_llm_completeness()
answer_quality_llm_contextual_relevance()
answer_quality_llm_faithfulness()
answer_quality_llm_grammar_accuracy()
answer_quality_llm_metrics()
answer_quality_llm_originality()
answer_quality_llm_relevance()
answer_viability_llm_metrics()
answer_viability_llm_reading_complexity()
answer_viability_llm_sentiment_assessment()
answer_viability_llm_text_fluency()
answer_viability_llm_text_toxicity()
automated_readability_index()
bleu()
character_count()
context_hit()
count_metrics()
flesch_kincaid_grade()
ground_truth_non_llm_answer_metrics()
ground_truth_non_llm_retrieval_metrics()
inner_product_retrieval()
inner_product_target_prediction()
levenshtein()
mrr()
non_llm_non_ground_truth_metrics()
quality_llm_text_similarity()
question_and_answer_metrics()
question_and_answer_metrics_extended()
rouge1()
rouge2()
rougeL()
rougeLsum()
rouge_metrics()
sentence_count()
summarization_metrics()
token_count()
text_metrics()
word_count()
Last updated 5 months ago
Was this helpful?