Before you can rely on any autonomous remediation measures, you need to close the loop between detection and action by codifying your best practices. This approach to bundling your performance metrics, remediation scripts, and operational workflows doesn’t just speed up MTTR, it depressurizes troubleshooting. Lets take a look at an example:
What happened? Your companies’ ThousandEyes agents flagged an intermittent packet loss on a CDN provider serving U.S. West customers. With many CDN’s fronting HTTP traffic, you don’t want handshake connections (packets) to outright fail rather than just load slowly – so now it's worrisome that real users could abandon their shopping carts and support tickets. Instead of checking all your dashboards and hopping in a war room, both ITOps and NetOps teams bidirectionally see a single, clear incident tied to shared SLAs and business impact. Before your hands even reach the keyboard, either automatically or through predefined logic, ITSI sets off a chain of workflows to stop the cascade of alerts and proactively optimize resources if the problem persists. Here’s a high level flow:
With these steps, ITOps and NetOps teams can eliminate midnight pages for fleeting issues, and surface problems in full-context with recommended remediation steps ready to guide the way.
Catching problems before they impact customers is kind of the ultimate idea behind fast detection & remediation, right? By applying advanced time-series forecasting and anomaly scoring to ThousandEyes network tests and application KPIs, ITOps and NetOps teams can turn historical data into forward-looking insights. This could mean getting an early warning sign of capacity bottlenecks, degrading performance, or routing instabilities as well as taking preemptive action to prevent the problems from occurring in the first place. Lets make a prediction ourselves on what this could look like for some of you later this year:
8AM: *Click – You flip on the local Wisconsin news, feeling excited but anxious to host so many relatives for the upcoming holiday break, especially so given how exhausted you were during this time last year. The meteorologists on TV warn that a ‘great blizzard will slam the Northeast U.S. on December 24th with up to 32 inches of snow’... ‘Oh no!’ you think to yourself – ‘I’ll probably be up all night trying to keep services running…’ – but this year you're prepared.
Now at your computer, you start forecasting network telemetry from ThousandEyes into ITSI, and see historical data from similar storms show an 180% surge in e-commerce traffic between 10PM and Midnight CST. With insight into the service performance impact of last-minute shoppers, you prepare to run more tests and pre-scale resources.
Try forecasting network telemetry from ThousandEyes in ITSI:
Rather than ITOps teams running to add resources during the storm, teams can pre‑scale services, validate CDN failover settings, and run synthetic smoke tests - transforming a risky deployment into a smooth, controlled event.
Lets bring everything we’ve already covered together for one last scenario, and imagine a mid‑day network degradation somewhere in Northeast Asia:
By the time teams log in for their shift, the incident is already mitigated – and customer experience remains uninterrupted.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.