<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: process control chart e.g. upper/lower control limit. in Splunk Search</title>
    <link>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30575#M178040</link>
    <description>&lt;P&gt;Thank you, if you have any other suggested reading please let me know!&lt;/P&gt;</description>
    <pubDate>Mon, 12 Aug 2013 16:00:16 GMT</pubDate>
    <dc:creator>bigtyma</dc:creator>
    <dc:date>2013-08-12T16:00:16Z</dc:date>
    <item>
      <title>process control chart e.g. upper/lower control limit.</title>
      <link>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30573#M178038</link>
      <description>&lt;P&gt;I have been asked to help a co-worker create a process control chart to understand an applications response time.&lt;/P&gt;

&lt;P&gt;The following three events are generated for each test. &lt;/P&gt;

&lt;P&gt;INFO=Signon_Screen RESPONSE_TIME=2.1000 &lt;BR /&gt;
INFO=Signon_Dept_Screen RESPONSE_TIME=0.6000&lt;BR /&gt;
INFO=Citrix_Login_Comp RESPONSE_TIME=7.6000&lt;/P&gt;

&lt;P&gt;The link below is a step in the right direction but I am having trouble getting this to work.&lt;/P&gt;

&lt;P&gt;&lt;A href="http://splunk-base.splunk.com/answers/73300/which-search-is-faster-reusing-a-calculation-in-an-if-clause-or-using-the-defined-variable-from-the-original-eval" target="_blank"&gt;http://splunk-base.splunk.com/answers/73300/which-search-is-faster-reusing-a-calculation-in-an-if-clause-or-using-the-defined-variable-from-the-original-eval&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 28 Sep 2020 14:32:22 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30573#M178038</guid>
      <dc:creator>bigtyma</dc:creator>
      <dc:date>2020-09-28T14:32:22Z</dc:date>
    </item>
    <item>
      <title>Re: process control chart e.g. upper/lower control limit.</title>
      <link>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30574#M178039</link>
      <description>&lt;P&gt;Simplest solution :&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;index=_internal sourcetype="splunk_web_access" uri="/en-GB/api/shelper" | table spent | rename spent as val 
| autoregress val 
| eval rangeMR=abs(val-val_p1) 
| eventstats count as numR avg(val) as AVG sum(rangeMR) as mrAVG
| eval mrAVG=mrAVG/(numR-1) 
| eval UCL=AVG+2.66*mrAVG 
| eval LCL=AVG-2.66*mrAVG 
| table val_p1 val AVG LCL UCL
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;The pitfalls are worth mentioning though.&lt;/P&gt;

&lt;P&gt;a) A process is said to be in control if its datapoints are within 6 sigma. The naive approach is +-3 sigma. However, for data with a hard floor ( i.e. 0 seconds, you end up with a LCL &amp;lt; 0 which is nonesense. Then you might want to set the LCL to 1 sigma and the UCL to 5 sigma. In other words you need to know your data and process.&lt;/P&gt;

&lt;P&gt;b) The approach was invented when they used to sample ball bearings produced over the day, and sample batches. In technology, the number of samples is huge. 3 sigma is ~ 1 in a thousand. In high volume processing your process will be out of control a lot more frequently than your heldesk can investigate. Backups, anything that creates network latency is going to kill you unless the network is only a small part of the process.&lt;/P&gt;

&lt;P&gt;c) Process charts sample over all data. This means it is possible for a steady stream of highly regular samples to push the chart out of control in the past. Again - this shows the roots in batch sampling for a shifts work - not live monitoring.&lt;/P&gt;

&lt;P&gt;d) The occasional spike means nothing in todays world. Running your data through a fourier transform to look for regular spikes is far more fun and informative. If you want to use control charts for monitoring, you need to also look at loss functions ( read Taguchi ) for a mechanism to infer impact.&lt;/P&gt;

&lt;P&gt;Unless you know what you are doing, using SPC charts to monitor data is a road to frustration.&lt;/P&gt;

&lt;P&gt;This technique has massive value in tuning software components that you have written and control. Its utterly useless in a meta sense like citrix logons.&lt;/P&gt;

&lt;P&gt;The technique dates back to a time where the number of moving parts to create an item was low, and each manufacturing process could be tuned - and that was the point - if you minimize variance - you reduce the monetary loss to the company in returned defective product.&lt;/P&gt;

&lt;P&gt;I seriously doubt you - or anyone else working in IT can name ( let alone influence ) all the moving parts in a citrix logon.&lt;/P&gt;</description>
      <pubDate>Sun, 11 Aug 2013 18:14:40 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30574#M178039</guid>
      <dc:creator>jonuwz</dc:creator>
      <dc:date>2013-08-11T18:14:40Z</dc:date>
    </item>
    <item>
      <title>Re: process control chart e.g. upper/lower control limit.</title>
      <link>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30575#M178040</link>
      <description>&lt;P&gt;Thank you, if you have any other suggested reading please let me know!&lt;/P&gt;</description>
      <pubDate>Mon, 12 Aug 2013 16:00:16 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Search/process-control-chart-e-g-upper-lower-control-limit/m-p/30575#M178040</guid>
      <dc:creator>bigtyma</dc:creator>
      <dc:date>2013-08-12T16:00:16Z</dc:date>
    </item>
  </channel>
</rss>

