13. Repeat evaluation

Repeat evaluation

You can add iterations to the experiment.

This is useful when you can repeat the evaluation multiple times:

  • For larger evaluation sets

  • For chains that can generate variable responses

  • Assessments that can generate variable scores (e.g. llm-as-judge )

Reference

# installation
# !pip install -qU langsmith langchain-teddynotew
# Configuration file for managing API KEY as environment variable
from dotenv import load_dotenv

# Load API KEY information
load_dotenv()
 True 
# Set up LangSmith tracking. https://smith.langchain.com
# !pip install -qU langchain-teddynote
from langchain_teddynote import logging

# Enter a project name.
logging.langsmith("CH16-Evaluations")

Define functions for RAG performance testing

Repeat evaluation for RAG using GPT model

Repeat evaluation for RAG using Ollama model

Last updated