If you’ve ever deployed a new database cluster, spun up a caching layer, or added a load balancer, you know it can involve a lot of manual configuration, YAML wrangling, and crossing your fingers that you didn’t miss anything critical.
Automatic Discovery in the Splunk Distribution of the OpenTelemetry Collector helps eliminate some of this toil so you can automatically and easily begin monitoring your infrastructure and applications in Splunk Observability Cloud the moment they come online.
Setting up observability for modern infrastructure can be a lift. You deploy a new database instance, and suddenly you’re writing Collector configurations, setting up service discovery, and manually defining which metrics to collect. Scale that across dozens of services, multiple environments, constant deployments, and you’ve got a recipe for potential observability gaps.
With traditional approaches, you need to:
And the result of all this? Potential blind spots in your observability coverage and time lost to configuration rather than innovation and problem solving.
Automatic Discovery thankfully flips this model on its head. Instead of manually configuring monitoring for every service, you enable Automatic Discovery once, and the Splunk Distribution of the OpenTelemetry Collector automatically detects, configures, and starts monitoring new services as they appear.
Automatic Discovery detects and collects signal data from third-party services such as databases and web servers by automatically generating configuration snippets that you can modify and incorporate into your existing configuration. Think of it as a smart assistant that constantly scans your infrastructure, recognizes common services, and automatically sets up monitoring based on what it finds.
With discovery mode enabled, the Collector performs intelligent detection through a multi-step process.
Observer Extensions Scan Your Environment
Discovery state is maintained in memory by the Collector during runtime. When a service is discovered and monitoring begins, the Collector tracks that endpoint as long as it remains active. If the Collector restarts, it performs a fresh discovery scan of the current environment state at startup, re-evaluating all known observer sources like nodes, pods, or network interfaces. This means it will rediscover any services that are still running but won’t retain information about services that stopped while the Collector was down.
For ephemeral services, like containers or scaled-down pods, the Collector stops attempting to collect metrics from those endpoints. In Splunk Observability Cloud, stopped services typically drop from the active discovery list after the next few polling cycles (usually within minutes).
With everything up and running, you can verify what’s been discovered by checking the Collector logs for discovery events or by viewing the new services and metrics that appear automatically in Splunk Observability Cloud’s Infrastructure Navigator.
If you were to deploy a new cache cluster with traditional manual configuration, before you could get any metrics you would need to:
If you happen to be running multiple cache instances across different namespaces, you’d need to repeat this process for each one.
With Automatic Discovery enabled, you deploy your new cache cluster, and the Collector immediately detects the new pods and automatically begins collecting metrics. Within seconds, you see cache-specific metrics – no manual Collector configuration, no YAML editing, no hoping someone remembers to add monitoring.
This same pattern applies for all your supported infrastructure services. Deploy it, and it’s automatically monitored.
With traditional manual configuration, it’s easy to miss new services or forget to update monitoring when services change.
Instead of waiting for someone to manually configure monitoring, new services become observable the moment they start running. This is critical for fast-moving teams practicing continuous deployment.
Whether you’re managing 10 services or 1,000, Automatic Discovery scales with your infrastructure. Define your rules once, and they apply consistently across all environments.
Manual configurations tend to drift across environments and teams. Automatic Discovery ensures consistent monitoring configuration based on your defined rules, reducing environment-specific quirks.
Development teams can deploy new services with proper monitoring automatically – no observability team intervention required.
Automatic discovery isn’t meant to replace thoughtful observability design – it’s meant to eliminate the tedious manual work that prevents you from focusing on what matters. Here’s how it fits into the bigger picture:
Think of Automatic Discovery as providing the observability foundation that lets you focus on higher-value observability practices like building meaningful dashboards, setting up intelligent alerting, and creating SLIs that matter to your business.
Automatic Discovery shines brightest in dynamic environments where services are constantly being deployed, scaled, and updated. If you’re running Kubernetes workloads, containerized applications, or cloud-native infrastructure, Automatic Discovery can drastically simplify your observability operations.
The setup process is straightforward, but there are important considerations around security, configuration management, and best practices that can make the difference between a smooth rollout and a frustrating experience.
In Part 2 of this series, we’ll walk through the step-by-step process of enabling Automatic Discovery, including how to securely handle credentials, avoid common pitfalls, and configure discovery rules that work for specific infrastructure patterns.
Ready to stop manually configuring observability for every new service? You can explore Automatic Discovery by setting it up with Splunk Observability Cloud – try it free for 14 days and see how it simplifies your setup.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.