13. Repeat evaluation
Repeat evaluation
You can add iterations to the experiment.
This is useful when you can repeat the evaluation multiple times:
For larger evaluation sets
For chains that can generate variable responses
Assessments that can generate variable scores (e.g.
llm-as-judge)
Reference
# installation
# !pip install -qU langsmith langchain-teddynotew# Configuration file for managing API KEY as environment variable
from dotenv import load_dotenv
# Load API KEY information
load_dotenv() True # Set up LangSmith tracking. https://smith.langchain.com
# !pip install -qU langchain-teddynote
from langchain_teddynote import logging
# Enter a project name.
logging.langsmith("CH16-Evaluations")Define functions for RAG performance testing
Repeat evaluation for RAG using GPT model

Repeat evaluation for RAG using Ollama model

Last updated