Getting Started with AI Team

Set up Edge Delta’s AI Team in five steps: connect observability data, review out-of-the-box monitors, integrate alerting and code tools, and start AI-powered investigations.

Overview

This guide walks you through five steps to get Edge Delta’s AI Team fully operational: connect your observability data, review the monitors provisioned for you, integrate your alerting and code tools, and start AI-powered investigations.

AI teammates operate read-only by default. Any write action (creating an incident, merging a PR, updating a ticket) requires your explicit approval before it executes.

AI Team overview AI Team overview

What AI teammates do

Edge Delta ships with five out-of-the-box teammates, each with deep expertise in a specific domain:

  • OnCall AI: Orchestrates investigations, routes requests to the right specialist, and synthesizes findings into actionable summaries.
  • SRE: Detects incidents, correlates logs, metrics, and traces, builds timelines, proposes mitigations, and pages on-call engineers.
  • Code Analyzer: Reviews pull requests for risks and missing tests, investigates CI/CD failures, and correlates code changes with production issues.
  • Security Engineer: Analyzes access patterns, reviews IAM configurations, surfaces code vulnerabilities, and detects leaked secrets.
  • Work Tracker: Tracks work across GitHub, Jira, and Linear, surfaces unowned items, and proposes ticket creation for gaps.

Each teammate activates fully once the right connectors are in place. The steps below walk you through connecting them in the recommended order.

For full details on each teammate’s capabilities, see Specialized Teammates.

Before you begin

You need the following:

  • An Edge Delta account (sign up if you do not have one)
  • Access to a Kubernetes cluster (recommended) or another telemetry source
  • Credentials for PagerDuty, GitHub/GitLab, and Jira/Linear if you use them

Step 1: Connect your observability data

AI teammates need telemetry to reason about your systems. Connecting an observability data source is the required first step.

Kubernetes (fastest path)

The Kubernetes connector is a bundled connector that provisions five telemetry sources in a single action:

SignalWhat it collects
LogsContainer stdout/stderr from pods across namespaces
MetricsCluster and workload metrics (CPU, memory, disk I/O)
TracesDistributed application traces across pod boundaries
Service MapLive inter-service communication topology built from network traffic
EventsPod, Node, Deployment, StatefulSet, and Service lifecycle events

To connect:

  1. Navigate to AI Team > Connectors
  2. Click Browse all Connectors
  3. Click the Kubernetes connector card
  4. Enter a connector name
  5. Click Kubernetes
Screenshot Screenshot
  1. Click Create Connector

The commands appear to install the new pipeline: run these commands on your cluster.

Screenshot Screenshot

With this single connector, the SRE teammate can begin correlating logs, metrics, events, traces, and service map data to surface issues early.

Screenshot Screenshot

Click Go to Connectors. Your new Kubernetes connector appears alongside two default connectors that every account includes: the Edge Delta MCP Connector (lets teammates query your telemetry data) and the Edge Delta Documentation Connector (lets teammates reference Edge Delta docs during investigations).

Screenshot Screenshot

For detailed configuration options, see Kubernetes Connector.

Non-Kubernetes alternatives

If you do not run Kubernetes, connect the telemetry source that matches your environment. Common options include File for local log files, OTLP for OpenTelemetry data, HTTP Pull for REST APIs, and Docker Logs for containerized workloads. See Connectors Overview for the full list.

The key principle: AI teammates need data flowing into Edge Delta before they can analyze anything.

Step 2: Review out-of-the-box monitors

Once telemetry is flowing, Edge Delta automatically provisions two default monitors for every new organization. These monitors detect issues and route alerts to AI teammates or your preferred notification channels without additional configuration.

MonitorTypeWhat it detects
Default Service Anomaly MonitorPattern AnomalyAnomalous log pattern spikes grouped by service, using a baseline comparison
Default Pipeline Liveness MonitorMetric ThresholdPipeline agents that stop reporting heartbeat metrics

You can view these monitors on the Monitors > List tab.

Screenshot Screenshot

To customize thresholds, sensitivity, or notification destinations, click the Actions kebab menu for the monitor and select Edit. See Monitors for details.

Screenshot Screenshot

The Default Service Anomaly Monitor uses a 15-minute spike window with a warning threshold of 5 and an alert threshold of 10. When the alert fires, it automatically notifies the @ai-channel-alerts channel, prompting AI teammates to assess severity, identify which specialists should investigate, and determine whether the issue requires immediate attention.

Screenshot Screenshot

The Default Pipeline Liveness Monitor evaluates agent heartbeat metrics over a 5-minute window and alerts when the value drops below 0. If data is missing for 5 minutes, the monitor evaluates the value as 0, which triggers the alert. This means the monitor fires when a pipeline stops reporting entirely, indicating a potential deployment failure or connectivity issue. Like the Service Anomaly Monitor, it notifies @ai-channel-alerts so AI teammates can assess the impact and recommend next steps.

Screenshot Screenshot

Both monitors route alerts to the AI Team alerts channel by default, so AI teammates begin investigating automatically without any additional configuration.

Beyond these defaults, you can create monitors of seven types: Metric Threshold, Metric Change, Log Threshold, Pattern Anomaly, Composite, Synthetic, and Event.

Route monitors to AI Team

Beyond the default monitors, you can configure any monitor to send notifications directly to AI Team channels. When a monitor fires an alert to an AI Team channel, teammates assigned to that channel begin autonomous investigation.

To connect a monitor to AI Team:

  1. Create or edit a monitor in Monitors > Manage
  2. In the Notifications section, type @ to open the autocomplete menu
  3. Select an AI Team channel from the list (channels appear with the prefix ai-channel-)
  4. Use conditional formatting to control when AI Team receives the notification

The following example routes alert-level notifications to an AI Team channel:

{{#is_alert}}
  **Alert**: {{monitor.name}} triggered
  - Value: {{value}}
  - Threshold: {{alert_threshold}}

  @ai-channel-security-issues
{{/is_alert}}

When the monitor enters the alert state, the security-issues channel receives the notification and assigned teammates begin correlating logs, checking for similar patterns, and assembling an investigation summary.

For detailed configuration options and more examples, see Monitor Notifications.

Step 3: Connect PagerDuty

If your team uses PagerDuty for incident management, connecting it enables AI teammates to receive real-time incident notifications and respond autonomously. There are two parts to set up.

Add the PagerDuty connector

  1. Navigate to AI Team > Connectors > PagerDuty
  2. Complete OAuth authorization (recommended) or configure an API token
  3. Copy the Webhook URL and Webhook Token displayed in the connector configuration
  4. Review the Tools tab to confirm permissions (read-only operations are allowed by default, write operations require your approval)
  5. Click Save
PagerDuty connector configuration showing webhook URL and token PagerDuty connector configuration showing webhook URL and token

The PagerDuty connector is automatically assigned to the SRE teammate.

Configure the PagerDuty webhook

The webhook is what closes the loop. PagerDuty fires an alert, and AI teammates receive it and begin triage automatically.

  1. In PagerDuty, navigate to Integrations > Generic Webhooks (v3)
  2. Click New Webhook
  3. Paste the Edge Delta Webhook URL from the connector configuration
  4. Add a custom header: set Name to Authorization and Value to Bearer <webhook_token> using the token from Edge Delta
  5. Select the event types to send (incident triggered, acknowledged, resolved, and others as needed)
  6. Click Add Webhook

For the complete walkthrough with screenshots of the PagerDuty webhook configuration, see PagerDuty Integration Guide.

Without the webhook, teammates can only use PagerDuty tools reactively when you ask. With the webhook, triage starts the moment an alert fires.

What happens when an alert fires

When an incident triggers in PagerDuty (for example, a CPU utilization spike on a production service), PagerDuty sends the alert payload to Edge Delta through the webhook. The event arrives in the alerts channel, where OnCall AI creates an investigation thread and routes it to the SRE teammate. SRE queries Edge Delta telemetry for correlated logs, metrics, and anomalies during the incident window. It checks PagerDuty for on-call information and related incidents. Within minutes, the thread contains a triage summary with symptoms, impact assessment, root cause hypotheses, and proposed next steps. The entire investigation and its findings are visible to your team, and a summary can be posted back to PagerDuty for the audit trail.

Step 4: Connect a Git repository

Connecting a Git repository enables AI teammates to correlate production issues with recent code changes. This is essential for root cause analysis: when the SRE teammate detects an anomaly, it can check what was recently deployed.

  • GitHub: Navigate to Connectors > GitHub and authenticate via OAuth (GitHub App) or Personal Access Token. See GitHub Connector for details.
  • GitLab: Navigate to Connectors > GitLab and authenticate via OAuth. See GitLab Connector for details.
GitHub connector in the connected state after OAuth authorization

Once connected, the Code Analyzer teammate activates fully. It can review pull requests for risks and missing tests, investigate CI/CD failures by correlating build logs with code changes, and identify deployment-correlated regressions during anomaly investigations. The Security Engineer teammate also uses the Git connector to surface Dependabot alerts and code scanning findings.

Connecting Jira or Linear activates the Work Tracker teammate, which synchronizes work items across your development tools and surfaces gaps where tasks are unowned or untracked.

  • Jira: Navigate to Connectors > Atlassian and authenticate via OAuth or API token. See Atlassian Connector for details.
  • Linear: Navigate to Connectors > Linear and authenticate via OAuth. See Linear Connector for details.

Work Tracker correlates tickets, PRs, alerts, and security findings. It operates read-only by default and proposes ticket creation only with your approval.

Teammate activation summary

Each specialized teammate activates when you connect the relevant data sources. The following table shows which connectors enable each teammate:

TeammateActivates withWhat it does when active
OnCall AIAlways activeOrchestrates investigations, routes requests, synthesizes findings
SREEdge Delta MCP, Elastic, PagerDuty, Kubernetes, AWS, Azure, GCPQueries logs, metrics, traces; investigates incidents; pages on-call
Code AnalyzerGitHub, CircleCIReviews PRs, investigates CI/CD failures, correlates code changes
Security EngineerGitHub, AWS, Jira, CrowdStrike FDRAssesses security posture, validates IAM, reviews access patterns
Work TrackerGitHub, Jira, LinearCreates tickets, synchronizes issues, retrieves runbooks

Note: Connectors are automatically assigned to their corresponding specialized teammates when configured. You can modify assignments in the teammate configuration.

Review your pipelines

After connecting your data sources, Edge Delta has created two pipelines that power the AI Team. You can view them on the Pipelines page.

Screenshot Screenshot

Kubernetes pipeline

The Kubernetes connector created a streaming pipeline that deploys an Edge Delta agent to your cluster. This pipeline collects five signal types through dedicated source nodes:

  • Kubernetes input: Container logs from pods across namespaces (excludes Edge Delta’s own pods to avoid feedback loops). Logs pass through a log-to-pattern processor that generates pattern metrics for anomaly detection.
  • Kubernetes events input: Pod, Node, Deployment, and Service lifecycle events. These also pass through the log-to-pattern processor.
  • Kubernetes traces input: Distributed application traces across pod boundaries.
  • Kubernetes traffic input: Network traffic between services, used to build the live service map.
  • Kubernetes metrics input: Cluster and workload metrics (CPU, memory, disk I/O), excluding low-level cadvisor, node exporter, and kubelet metrics to reduce noise.

All five sources route through a shared processor sequence to an Edge Delta destination, where the data becomes available for AI teammates to query through the Edge Delta MCP Connector.

Screenshot Screenshot

Event connector pipeline

The event connectors (PagerDuty, GitHub, and Atlassian) share a single ingestion pipeline called AI-Connector-Telemetry-Pipeline. Each event connector appears as a separate source node in this pipeline. When you add more event connectors in the future, they are automatically added as new sources to this same pipeline.

Each source routes to two destinations:

  • Edge Delta: Stores the event data for search and analysis
  • AI teammates: Delivers events directly to OnCall AI, which creates investigation threads and routes to the appropriate specialist

This dual-destination design means event data is both archived for historical queries and immediately actionable by AI teammates.

Screenshot Screenshot

For more details on how connectors create and use pipelines, see AI Team Overview.

Start your first conversation

With connectors in place, you can begin using AI Team:

  1. Click on the Chat tab.
  2. Select a channel (for example alerts or security-issues).
  3. Type your question in the message box (for example, “Show me recent anomalies in my logs” or “Help me understand my pipeline configuration”).
  4. Press Enter or click Send.

OnCall AI orchestrates the response by routing your request to the appropriate specialized teammates in the thread.

Teammates also respond to incoming events from connectors. For example, when a PagerDuty webhook sends an incident notification, AI teammates can investigate and respond in dedicated channels:

Screenshot Screenshot

Use message enhancement

Message Enhancement uses AI to expand simple queries into more detailed, context-rich prompts. For example, typing "What caused the recent spike in errors?" and clicking the Enhance Message button transforms your query into something like: "Analyze recent error patterns in my telemetry data, identify the root causes of any spikes, correlate with deployment events or infrastructure changes, and provide actionable recommendations to prevent future occurrences." You can review and edit the enhanced version before sending. Enhanced messages typically result in more accurate, comprehensive responses from AI Team members.

Create custom teammates and channels

Beyond the out-of-the-box teammates and channels, you can create custom teammates and channels for your specific workflows.

To create a custom channel:

  1. In the Chat tab, locate the Channels section
  2. Click the + (plus) icon next to “Channels”
  3. Fill in the channel details:
    • Name: Descriptive name (e.g., “database-ops”, “cost-optimization”)
    • Description: Brief description of the channel’s purpose
    • Select Teammates: Choose which teammates should participate
  4. Click Create

Next steps

Common questions

Q: Do I need a Telemetry Pipeline to use AI Team? A: Event connectors automatically create and share a single ingestion pipeline called OnCall_AI_Connectors, with each event connector added as a source node. Ingestion pipelines are stateless and require no provisioning. Streaming connectors integrate with your existing pipelines or guide you through creating new ones. While you can use AI Team for general assistance without pipelines, the full analytical power of specialized teammates becomes available once they can access your telemetry data through these pipelines.

Q: What’s the difference between configuring a connector through AI Team vs. creating a pipeline manually? A: AI Team provides a streamlined connector-first workflow that automatically creates an ingestion pipeline for event connectors. Ingestion pipelines are stateless and require no provisioning, making them ideal for event-driven workflows. For streaming connectors, it guides you through adding them to pipelines. This makes AI Team an easier entry point for setting up telemetry pipelines, while still giving you full pipeline control through standard processors and configurations.

Q: Which connectors should I configure first? A: Follow the five steps in this guide. Start with the Kubernetes connector (or your primary telemetry source), then add PagerDuty if you use it for alerting, followed by GitHub or GitLab for code analysis.

Q: Can multiple teammates use the same connector? A: Yes, connectors can be shared across multiple teammates. Configure once, assign to many.

Q: How do I control what data AI teammates can see? A: The pipelines created by AI Team connectors support all standard Edge Delta processors. You can add filters, samplers, masking processors, or any other transformation to control exactly what data reaches the AI teammates. This gives you the same level of control as manually configured pipelines.

Q: What’s the difference between a channel and a direct message? A: Channels are shared spaces where multiple teammates collaborate on a topic and provide organization-wide visibility for all actions. Direct messages let you query individual teammates for read-only insights, analysis, or recommendations. If you need a teammate to make changes to your infrastructure (for example, using MCP to update configurations), you must use a channel to ensure all users in your organization have visibility into those actions.

Q: How do I know which teammate to use? A: Start a thread in any channel. OnCall AI automatically routes your request to the appropriate specialist. As you become familiar with each teammate’s capabilities, you can message them directly in channels or DMs.

Q: Which connectors activate which teammates? A: See the Teammate activation summary above. OnCall AI is always active. Other teammates activate when you connect the relevant data sources.

Troubleshooting

Teammate not responding: Check that connectors required for your question are configured and that you have an active Telemetry Pipeline if querying observability data.

Teammate not appearing in channel: Verify the teammate is assigned to the channel via Channel Settings.

Connector authentication failing: Ensure credentials are current and have necessary permissions. Some connectors require specific scopes or roles.

For additional help, visit the Support page or check Troubleshooting.