From 32109f7a509b1b400471bb14ebb8726d1fcc3312 Mon Sep 17 00:00:00 2001 From: Cleo Schneider Date: Thu, 2 May 2024 11:55:46 +0000 Subject: [PATCH] Incorporating feedback --- docs/evaluation.md | 15 +++++++-------- 1 file changed, 7 insertions(+), 8 deletions(-) diff --git a/docs/evaluation.md b/docs/evaluation.md index a1c63b1c8..e326ab876 100644 --- a/docs/evaluation.md +++ b/docs/evaluation.md @@ -53,23 +53,22 @@ genkit eval:flow bobQA --input testQuestions.json --output eval-result.json Note: Below you can see an example of how an LLM can help you generate the test cases. -## Supported Evaluator Plugins +## Supported evaluators -### Genkit Eval +### Genkit evaluators -We have created a small number of native evaluators to help developers get started that are inspired by [RAGAS](https://docs.ragas.io/en/latest/index.html) metrics including: +Genkit includes a small number of native evaluators, inspired by RAGAS, to help you get started: - Faithfulness - Answer Relevancy - Maliciousness -### VertexAI Rapid Evaluators +### Evaluator plugins -We support a handful of VertexAI Rapid Evaluators via the [VertexAI Plugin](/docs/plugins/vertex-ai#evaluation). +Genkit supports additional evaluators through plugins: -### Langchain Evaluators - -Firebase Genkit supports [Langchain Criteria Evaluation](https://python.langchain.com/docs/guides/productionization/evaluation/string/criteria_eval_chain/) via the Langchain Plugin. +- VertexAI Rapid Evaluators via the [VertexAI Plugin](plugins/vertex-ai#evaluation). +- [LangChain Criteria Evaluation](https://python.langchain.com/docs/guides/productionization/evaluation/string/criteria_eval_chain/) via the [LangChain plugin](plugins/langchain.md). ## Advanced use