In DeepEval's latest release, we are introducing a slight change in how a conversation is evaluated.
Previously we assumed a conversation as as a list of LLMTestCase
s, which might necessarily be the case. Now a conversational test case is made up of a list of Turn
s instead, which follows OpenAI's standard messages
format:
from deepeval.test_case import Turn
turns = [Turn(role="user", content="...")]
Docs here: https://deepeval.com/docs/evaluation-test-cases#conversational-test-case