Alerting

How to receive a alert when job gets fail?

arunpt92
Loves-to-Learn Lots

index="db_oracle-prod" source="AzureVOCprod" status=4
| eval MSGStatus=case(status=1,"CREATED", status=2,"RUNNING", status=3,"CANCELLED", status=4,"Failed", status=5,"PENDING", status=6,"ENDED UNEXPECTEDLY", status=7,"SUCCEEDED",status=8,"STOPPING", status=9,"COMPLETED")
| join package_name [inputlookup Azure_VOC.csv]
| eval STARTTime=strptime((strftime(now(),"%Y-%m-%d")),"%Y-%m-%d") - strptime(start_time,"%Y-%m-%d")
| where STARTTime=0
| stats count by Azure_Pipeline_name, start_time, end_time, MSGStatus

 

receiving record every 15mins instead we should have only failure ones based on timeframe

Labels (1)
0 Karma

richgalloway
SplunkTrust
SplunkTrust

What are the earliest and latest settings for the alert?  How often does it run?  Is throttling enabled?

---
If this reply helps you, Karma would be appreciated.
0 Karma

arunpt92
Loves-to-Learn Lots

Our first job starts at 11.30AM IST and ends at 6.30PM IST. Should get an alert if job gets fails. But we are getting on data getting record every 15mins instead we should have only 1 or 2 records

 

0 Karma

richgalloway
SplunkTrust
SplunkTrust

I wasn't asking for the time range of when the alert runs, but the time range the alert searches.  It will in the dispatch.earliest and dispatch.latest fields in savedsearches.conf or rest /services/saved/searches.

If the alert runs every 15 minutes and searches more than 15 minutes of events then results will be repeated.

---
If this reply helps you, Karma would be appreciated.
0 Karma
Get Updates on the Splunk Community!

Data Management Digest – December 2025

Welcome to the December edition of Data Management Digest! As we continue our journey of data innovation, the ...

Index This | What is broken 80% of the time by February?

December 2025 Edition   Hayyy Splunk Education Enthusiasts and the Eternally Curious!    We’re back with this ...

Unlock Faster Time-to-Value on Edge and Ingest Processor with New SPL2 Pipeline ...

Hello Splunk Community,   We're thrilled to share an exciting update that will help you manage your data more ...