LangChain Reference home pageLangChain ReferenceLangChain Reference
  • GitHub
  • Main Docs
Deep Agents
LangChain
LangGraph
Integrations
LangSmith
  • Overview
  • MCP Adapters
    • Overview
    • Agents
    • Callbacks
    • Chains
    • Chat models
    • Embeddings
    • Evaluation
    • Globals
    • Hub
    • Memory
    • Output parsers
    • Retrievers
    • Runnables
    • LangSmith
    • Storage
    Standard Tests
    Text Splitters
    ⌘I

    LangChain Assistant

    Ask a question to get started

    Enter to send•Shift+Enter new line

    Menu

    MCP Adapters
    OverviewAgentsCallbacksChainsChat modelsEmbeddingsEvaluationGlobalsHubMemoryOutput parsersRetrieversRunnablesLangSmithStorage
    Standard Tests
    Text Splitters
    Language
    Theme
    Pythonlangchain-classicevaluationqaeval_chainContextQAEvalChainevaluate
    Method●Since v1.0

    evaluate

    Evaluate question answering examples and predictions.

    Copy
    evaluate(
      self,
      examples: list[dict],
      predictions: list[dict],
      question_key: str = 'query',
      context_key: str = 'context',
      prediction_key: str = 'result',
      *,
      callbacks: Callbacks = None
    ) -> list[dict]

    Used in Docs

    • How to add evaluators to an existing experiment (Python only)
    • How to define a code evaluator
    • How to define an LLM-as-a-judge evaluator
    • How to evaluate an LLM application
    • How to evaluate with repetitions
    View source on GitHub