Rancher Integration
Monitor all your Rancher-managed clusters in one place. Fleet deployments, downstream cluster health, project quotas, and Rancher server metrics with AI-powered correlation.
How It Works
Deploy TigerOps via Rancher Apps
Install the TigerOps Helm chart from the Rancher Apps & Marketplace. The chart deploys the monitoring agent to both the local cluster and all registered downstream clusters via the Rancher multi-cluster app feature.
Enable Rancher Monitoring Integration
If rancher-monitoring is enabled, TigerOps automatically adds a RemoteWrite to the existing Prometheus stack. For clusters without rancher-monitoring, TigerOps deploys its own lightweight scraper.
Configure Fleet & Cluster Scoping
Map Rancher projects and clusters to TigerOps monitoring scopes. Project members see only their workload metrics. Fleet workspace metrics are aggregated at the GitRepo and Bundle level for deployment tracking.
Set Fleet & Cluster Health Alerts
Define alerts for Fleet bundle deployment failures, downstream cluster connectivity loss, and Rancher server API latency. TigerOps sends alerts through your existing notification channels with Rancher cluster context.
What You Get Out of the Box
Fleet Bundle Deployment Tracking
Monitor Fleet GitRepo sync status, Bundle ready/modified/errored counts, and BundleDeployment rollout progress per downstream cluster. Alert when bundles fail to apply or drift from the desired state.
Downstream Cluster Health
Track registered cluster agent connectivity, API server response latency, and cluster condition states for every downstream cluster from the Rancher management plane. Detect disconnected clusters before users report issues.
Rancher Server Component Metrics
Monitor Rancher server API request rates, webhook call latency, Norman and Steve API handler errors, and leader election status for the Rancher management server pods.
Project & Namespace Quota Monitoring
Track Rancher project resource quotas (CPU, memory, services, pods) across all member namespaces. Alert when quota utilization approaches limits to prevent scheduling failures in multi-tenant clusters.
etcd & Control Plane per Cluster
Per-cluster etcd member health, wal fsync latency, and control plane component availability for all downstream clusters. Rancher cluster provisioning logs are correlated with control plane metric anomalies.
AI Cross-Cluster Incident Correlation
When a Fleet bundle fails across multiple clusters, TigerOps AI identifies shared root causes — a bad manifest, a registry outage, or a node resource crunch — and surfaces the pattern across all affected clusters simultaneously.
Helm Values for Rancher Multi-Cluster Deployment
Deploy TigerOps across all Rancher-managed clusters using the multi-cluster app feature and Fleet.
# TigerOps Helm values for Rancher multi-cluster deployment
# helm repo add tigerops https://charts.atatus.net
# helm install tigerops tigerops/tigerops-rancher -f values.yaml \
# -n cattle-monitoring-system --create-namespace
global:
apiKey: "${TIGEROPS_API_KEY}"
remoteWriteEndpoint: https://ingest.atatus.net/api/v1/write
rancher:
enabled: true
# Integrate with existing rancher-monitoring if present
rancherMonitoringIntegration: true
# Rancher management server metrics
serverMetrics: true
# Scrape the Rancher API server
apiServerScrape: true
fleet:
enabled: true
# Monitor GitRepo sync and Bundle deployment status
bundleTracking: true
workspaces: [] # empty = all workspaces
downstreamClusters:
# Deploy agent to all registered downstream clusters
allClusters: true
# Or target specific cluster IDs
# clusterIds:
# - c-m-abc123
# - c-m-def456
projectQuotaMonitoring:
enabled: true
alertThresholds:
cpu: 80
memory: 80
pods: 85
alerts:
bundleDeploymentFailed: true
clusterDisconnected: true
rancherServerLatencyMs: 500Common Questions
Does TigerOps work alongside rancher-monitoring (kube-prometheus-stack)?
Yes. TigerOps integrates with the existing rancher-monitoring Prometheus by adding a RemoteWrite entry. No duplicate scraping occurs. If rancher-monitoring is not installed, TigerOps deploys its own scraper in the cattle-monitoring-system namespace.
How does TigerOps monitor all downstream clusters from a single configuration?
The TigerOps Rancher integration uses the Rancher multi-cluster app feature and Fleet to deploy monitoring agents to downstream clusters. Each agent ships metrics tagged with the Rancher cluster ID so you can filter and compare clusters in TigerOps dashboards.
Can TigerOps track Fleet bundle deployment history over time?
Yes. TigerOps records Bundle and BundleDeployment state transitions over time, building a deployment history timeline per GitRepo. You can view deployment frequency, failure rates, and mean time to ready across your entire Fleet workspace.
Does TigerOps support Rancher on RKE, RKE2, and K3s-provisioned clusters?
Yes. TigerOps supports clusters provisioned by RKE1, RKE2, and K3s within Rancher. The agent adapts its control plane metric collection based on the provisioner detected from Rancher cluster annotations.
How are Rancher project permissions reflected in TigerOps access control?
TigerOps syncs Rancher project membership via the Rancher API. Project owners and members are assigned TigerOps scopes that restrict dashboard and alert access to only the namespaces belonging to their Rancher project. Cluster owners have unrestricted access.
One Dashboard for Every Rancher Cluster and Fleet Deployment
Multi-cluster health, Fleet bundle tracking, and project quota monitoring — unified and AI-correlated across your entire Rancher estate.