Monitoring Mastra with SigNoz

Overview

This guide walks you through setting up monitoring for Mastra using OpenTelemetry and exporting traces to SigNoz. With this integration, you can observe model performance, capture request/response details, giving you real-time visibility into latency, error rates, and usage trends for your Mastra applications.

Instrumenting Mastra in your AI applications with telemetry ensures full observability across your AI workflows, making it easier to debug issues, optimize performance, and understand user interactions.

Prerequisites

  • A SigNoz Cloud account with an active ingestion key
  • Internet access to send telemetry data to SigNoz Cloud
  • Node.js (v20.0 or higher)
  • An API key from a supported Model Provider

Setting Up Mastra Project

You can get the full Mastra installation instructions here.

Step 1: Start the CLI Wizard

Run the following command to start the interactive setup:

npx create-mastra@latest

Step 2: Add Your API key

Add your API key to the .env file:

OPENAI_API_KEY=<your-api-key>

This example uses OpenAI. Each LLM provider uses a unique name. See Model Capabilities for more information.

Setup an Agent

Step 1: Create an agent in your project.

For example, create a file src/mastra/agents/chefAgent.ts:

import { openai } from "@ai-sdk/openai";
import { Agent } from "@mastra/core/agent";

export const chefAgent = new Agent({
  name: "chef-agent",
  instructions:
    "You are Michel, a practical and experienced home chef" +
    "You help people cook with whatever ingredients they have available.",
  model: openai("gpt-4o-mini"),
});

Step 2: Register the Agent with Mastra

In your src/mastra/index.ts file, register the agent with telemetry enabled:

import { Mastra } from "@mastra/core";
import { chefAgent } from "./agents/chefAgent";
import { OtelExporter } from "@mastra/otel-exporter";


export const mastra = new Mastra({
  agents: { chefAgent },
  observability: {
    configs: {
      otel: {
        serviceName: '<service-name>',
        exporters: [
          new OtelExporter({
            provider: {
              signoz: {
                apiKey: process.env.SIGNOZ_INGESTION_KEY,
                region: 'us', // 'us' | 'eu' | 'in',
              }
            },
          })
        ],
      },
    },
  },
});
  • <service_name> is the name of your service
  • SIGNOZ_INGESTION_KEY → Your SigNoz ingestion key. Set the value as an environment variable.
  • region → appropriate region

Step 3: Run Mastra Dev Server and Interact with Agent

npm run dev

You should see some output similar to this:

Execution Output
Output when running the Mastra Dev Server

Visit the playground url, and start chatting with the registered agent. All your interactions should be traced and sent to SigNoz under the traces tab.

View Traces in SigNoz

Your Mastra interactions should now automatically emit traces.

You should be able to view traces in Signoz Cloud under the traces tab:

Mastra Trace View
Mastra Trace View

When you click on a trace in SigNoz, you'll see a detailed view of the trace, including all associated spans, along with their events and attributes.

Mastra Detailed Trace View
Mastra Detailed Trace View

Dashboard

You can also check out our custom Mastra dashboard here which provides specialized visualizations for monitoring your Masrta usage in applications. The dashboard includes pre-built charts specifically tailored for LLM usage, along with import instructions to get started quickly.

Mastra Dashboard
Mastra Dashboard Template

Last updated: October 10, 2025

Edit on GitHub

Was this page helpful?