Skip to main content

What is ingestion?

Ingestion allows you to send LLM usage data to Moda from your existing applications. This is useful if you:
  • Already have an application making direct calls to LLM providers
  • Want to add analytics without changing how you call the APIs

Ways to ingest data

How it works

Your App  ──>  LLM Provider (OpenAI, Anthropic, etc.)

    │  Moda SDK captures telemetry


Moda Ingestion API


Moda Dashboard
  1. Your app makes calls to LLM providers as normal
  2. The Moda SDK captures telemetry in the background
  3. Telemetry is sent to Moda’s ingestion API
  4. Moda validates your API key and processes the data
  5. View insights and analytics in the Moda dashboard

What data is captured?

FieldDescription
Conversation IDGroups related messages together
User messageWhat the user asked
Assistant responseWhat the AI replied
ModelWhich model was used
ProviderLLM provider name (e.g., openai, anthropic)
TimestampWhen the interaction happened
Token usageInput, output, and total tokens consumed
Reasoning tokensTokens used for extended thinking (e.g., Claude)
Content blocksStructured content including tool use, thinking, and images
User IDUser identifier for per-user analytics
EnvironmentDeployment environment (development, staging, production)
Prompt trackingPrompt template ID, name, and version

Endpoints

All ingestion endpoints are hosted at https://moda-ingest.modas.workers.dev.
EndpointUse Case
POST /v1/tracesUsed automatically by the Moda SDK
POST /v1/ingestDirect API integrations (conversations across chat, email, voice, and standard completions)

Batch Limits

These limits apply across all ingestion endpoints:
LimitValue
Max events per request1,000
Max message size100 KB
Max request size5 MB

Privacy

Moda stores conversation content to provide analytics. Make sure this aligns with your privacy policy and data handling requirements.
Do not send sensitive personal information (like passwords or credit card numbers) through LLM calls that are being logged.