<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Reading 1000+ overwritten json files on time interval in Splunk Dev</title>
    <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305338#M4011</link>
    <description>&lt;P&gt;After adding "initCrcLength=1048576" this issue got resolved but when sources got overwritten, the unique source count got reduced in search head.&lt;/P&gt;</description>
    <pubDate>Wed, 13 Sep 2017 12:51:39 GMT</pubDate>
    <dc:creator>rajkumar3v</dc:creator>
    <dc:date>2017-09-13T12:51:39Z</dc:date>
    <item>
      <title>Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305329#M4002</link>
      <description>&lt;P&gt;I have 1000+ json files located in a directory and those files will be overwritten by every day. the file name starting with same characters as shown below,&lt;/P&gt;

&lt;P&gt;1000010496,1000011820,1000013553,1000010097,1000010362...&lt;/P&gt;

&lt;P&gt;my issue is that splunk forwarder is not reading all the files. I have tried flushing fishbucket,deleted indexed data,crcSalt,adding timestamp in filename and none of this have helped me to get entire data. even very less count of source files are showing in splunk. how to read this 1000+ files repeatedly without missing data?&lt;/P&gt;

&lt;P&gt;json files starts like below,&lt;/P&gt;

&lt;P&gt;$result = [&lt;BR /&gt;
            {&lt;BR /&gt;
              'advisory_type' =&amp;gt; 'Security Advisory',&lt;BR /&gt;
              'date' =&amp;gt; '10/12/17',&lt;BR /&gt;
              'advisory_name' =&amp;gt; 'CL-SA-2017:0061',&lt;BR /&gt;
             }          ....&lt;BR /&gt;
            ....&lt;/P&gt;

&lt;P&gt;Thanks in advance.&lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 15:32:01 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305329#M4002</guid>
      <dc:creator>rajkumar3v</dc:creator>
      <dc:date>2020-09-29T15:32:01Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305330#M4003</link>
      <description>&lt;P&gt;Have you ever had all of them indexed (like on the initial start of the forwarder, not just re-reading the files after they are updated)?&lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2017 09:03:05 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305330#M4003</guid>
      <dc:creator>cpetterborg</dc:creator>
      <dc:date>2017-08-30T09:03:05Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305331#M4004</link>
      <description>&lt;P&gt;no, at the time of  first indexing, splunk dint read all the files. it listed only 356 sources instead of 1300 sources..&lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2017 09:23:35 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305331#M4004</guid>
      <dc:creator>rajkumar3v</dc:creator>
      <dc:date>2017-08-30T09:23:35Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305332#M4005</link>
      <description>&lt;P&gt;Do you get a full list of the files when you run this on the forwarder?:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;splunk list monitor
&lt;/CODE&gt;&lt;/PRE&gt;</description>
      <pubDate>Wed, 30 Aug 2017 09:54:49 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305332#M4005</guid>
      <dc:creator>cpetterborg</dc:creator>
      <dc:date>2017-08-30T09:54:49Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305333#M4006</link>
      <description>&lt;P&gt;yes, am getting full list but in search head getting 229 unique sources only. i think splunk will be monitoring the paths specified in monitoring and its not reading files to avoid re-indexing same filename or content. &lt;/P&gt;</description>
      <pubDate>Wed, 30 Aug 2017 10:54:25 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305333#M4006</guid>
      <dc:creator>rajkumar3v</dc:creator>
      <dc:date>2017-08-30T10:54:25Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305334#M4007</link>
      <description>&lt;P&gt;anyone have a solution, please post it..&lt;/P&gt;</description>
      <pubDate>Mon, 04 Sep 2017 05:29:26 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305334#M4007</guid>
      <dc:creator>rajkumar3v</dc:creator>
      <dc:date>2017-09-04T05:29:26Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305335#M4008</link>
      <description>&lt;P&gt;If you have purchased Splunk and have a valid support contract, I'd submit a case to Splunk support.&lt;/P&gt;

&lt;P&gt;Also, if you are not running the latest version of Splunk, you may want to upgrade.&lt;/P&gt;

&lt;P&gt;And finally, if there are empty JSON files, they will not show up in the indexers on in searches because there is no data to index. Check for empty files.&lt;/P&gt;</description>
      <pubDate>Mon, 04 Sep 2017 12:53:41 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305335#M4008</guid>
      <dc:creator>cpetterborg</dc:creator>
      <dc:date>2017-09-04T12:53:41Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305336#M4009</link>
      <description>&lt;P&gt;The problem is that you have too many files/directories to sort through and splunk is getting bogged down tracking everything. You need to make sure that there is a housekeeping process ( &lt;CODE&gt;logrotate&lt;/CODE&gt; can do this ) that is deleting the older log files so they do not hang around "forever". This will only get worse.  Splunk forwarders start to really bog down when having to track and sort through thousands of files and once you cannot make the rounds before you are scheduled to go back around and check (I have no idea what the numbers are for this), then you are in a never-ending cycle of fail and ever-worsening delays.  Also, check your &lt;CODE&gt;inodes&lt;/CODE&gt;; you need user &lt;CODE&gt;splunk&lt;/CODE&gt; to be &lt;CODE&gt;ulimit unlimited&lt;/CODE&gt;.&lt;/P&gt;</description>
      <pubDate>Sun, 10 Sep 2017 20:22:08 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305336#M4009</guid>
      <dc:creator>woodcock</dc:creator>
      <dc:date>2017-09-10T20:22:08Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305337#M4010</link>
      <description>&lt;P&gt;I find these helpful for &lt;EM&gt;up&lt;/EM&gt;-setting the limits. For RHEL6 and earlier:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;cd /etc/security
cat  &amp;gt;&amp;gt;limits.conf &amp;lt;&amp;lt;EOF

*       hard    nofile  102400
*       soft    nofile  10240
*       hard    nproc   16384
*       soft    nproc   16384
EOF
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;And for RHEL7+:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;mkdir -p /etc/systemd/system/splunk.service.d
cat &amp;gt;&amp;gt; /etc/systemd/system/splunk.service.d/filelimit.conf &amp;lt;&amp;lt;EOF

[Service]
LimitNOFILE=10240
EOF
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;Reboot afterwards.&lt;/P&gt;

&lt;P&gt;These can be found around Answers and Docs, but for quick reference here I've provided them. Other versions of Linux will vary, but these are typical for most people to use. &lt;STRONG&gt;Check your version to ensure that these would work for you!!&lt;/STRONG&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 10 Sep 2017 21:08:09 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305337#M4010</guid>
      <dc:creator>cpetterborg</dc:creator>
      <dc:date>2017-09-10T21:08:09Z</dc:date>
    </item>
    <item>
      <title>Re: Reading 1000+ overwritten json files on time interval</title>
      <link>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305338#M4011</link>
      <description>&lt;P&gt;After adding "initCrcLength=1048576" this issue got resolved but when sources got overwritten, the unique source count got reduced in search head.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Sep 2017 12:51:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Splunk-Dev/Reading-1000-overwritten-json-files-on-time-interval/m-p/305338#M4011</guid>
      <dc:creator>rajkumar3v</dc:creator>
      <dc:date>2017-09-13T12:51:39Z</dc:date>
    </item>
  </channel>
</rss>

