09. Experimental (Experiment) evaluation comparison

Experimental (Experiment) evaluation comparison

The Compare feature provided by LangSmith makes it easy to compare experimental results.

Reference

Copy

#installation
# !pip install -qU langsmith langchain-teddynote

Copy

# Configuration file for managing API KEY as environment variable
from dotenv import load_dotenv

# Load API KEY information
load_dotenv()

Copy

 True 

Copy

# Set up LangSmith tracking. https://smith.langchain.com
# !pip install -qU langchain-teddynote
from langchain_teddynote import logging

# Enter a project name.
logging.langsmith("CH16-Evaluations")

Copy

Define functions for RAG performance testing

We will create a RAG system to use for testing.

Copy

Copy

Copy

Utilize the GPT-4o-mini model and the Ollama model to generate functions that generate answers to your questions.

Copy

Evaluate answers using the GPT-4o-mini model and Ollama model.

Proceed for each of the two chains.

Copy

Use a comparative view to examine the results.

How to make a comparison view

  1. On Dataset's Experiment tab, select the experiment you want to compare.

  2. Click the "Compare" button at the bottom.

  3. A comparison view

Last updated