LangGraph gives you stateful agent flows with graph-based orchestration, but you still need a reliable way to verify graph behavior after prompt, model, tool, or node changes. EvalView provides that regression loop with a dedicated LangGraph adapter featuring auto-detection, streaming support, and native thread tracking.
LangGraph agents have unique testing needs beyond simple input/output checks:
EvalView auto-detects LangGraph agents. Start your agent server, then connect and test:
# Start your LangGraph agent
uvicorn main:app --reload --port 8000
# Connect EvalView (auto-detects LangGraph)
evalview connect --endpoint http://localhost:8000/api/chat
# Generate tests from the running agent
evalview generate
# Snapshot approved behavior
evalview snapshot tests/generated --approve-generated
# Check for regressions
evalview check tests/generated
Add regression checks to your CI pipeline so every PR is validated:
- name: Check for regressions
uses: hidai25/eval-view@v0.6.0
with:
openai-api-key: ${{ secrets.OPENAI_API_KEY }}
fail-on: REGRESSION
Multi-turn tests are especially useful for LangGraph because they verify state management across conversation turns:
name: refund-needs-order-number
turns:
- query: "I want a refund"
expected:
output:
contains: ["order number"]
- query: "Order 4812"
expected:
tools: ["lookup_order", "check_policy"]
forbidden_tools: ["delete_order"]
output:
contains: ["refund", "processed"]
CI/CD Integration Guide | Regression Testing Guide | Back to EvalView homepage