Skip to Content
IntegrationsAgent Integrations

Agent Integrations

Connect Lamdis to your AI assistants for automated testing.

Supported Integration Methods

Lamdis can test any AI assistant that communicates via HTTP. The primary integration method is:

HTTP Chat API

The most flexible option. Your assistant exposes an HTTP endpoint that accepts messages and returns responses. Works with:

  • Custom chatbots and agents
  • RAG systems
  • Workflow agents
  • Any AI service with an HTTP API

How Integration Works

  1. Create a Connection in Integrations → Connections 
  2. Configure authentication (API key, OAuth, or custom headers)
  3. Create an Assistant in Assistants  using the connection
  4. Reference the assistant in your test suites

Request/Response Format

For HTTP Chat integrations, Lamdis sends:

{ "message": "User's message", "transcript": [ { "role": "user", "content": "Previous message" }, { "role": "assistant", "content": "Previous response" } ], "persona": "Optional persona description" }

Your assistant should respond with:

{ "reply": "Assistant's response text" }

Getting Started

  1. Create a connection  to your assistant’s endpoint
  2. Set up an assistant  configuration
  3. Create a test suite  and run your first test

Self-Hosted Assistants

For assistants running locally or in your own infrastructure:

  • Ensure the endpoint is accessible from Lamdis (or your self-hosted runner)
  • Use environment-specific connections for dev/staging/prod
  • Configure any required authentication headers

See Connections  for detailed setup instructions.

Last updated on