<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Segfault errors on a indexer in cluster in Deployment Architecture</title>
    <link>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422924#M15075</link>
    <description>&lt;P&gt;We have this issue very frequently which appeared to have started right after the last upgrade.&lt;BR /&gt;
Below kernel logs shows the frequency, Splunk process on the indexer appears running without restart so it appears to be from search processes.  &lt;/P&gt;

&lt;HR /&gt;

&lt;P&gt;*Linux splunkindexer1  2.6.32-754.9.1.el6.x86_64 #1 SMP Wed Dec 21 10:08:21 PST 2018 x86_64 x86_64 x86_64 GNU/Linux&lt;BR /&gt;
-bash-4.1$ cat /var/log/messages | grep -i kernel| tail&lt;BR /&gt;
Jul 31 08:16:24 splunkindexer1 kernel: splunkd[3149]: segfault at 7ff425810057 ip 000055ad21554260 sp 00007ff4047f8068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:19:34 splunkindexer1 kernel: splunkd[7907]: segfault at 7ff42540e057 ip 000055ad21554260 sp 00007ff4043f6068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:20:30 splunkindexer1 kernel: splunkd[22411]: segfault at 7ff42560f057 ip 000055ad21554260 sp 00007ff4045f7068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:21:07 splunkindexer1 kernel: splunkd[30162]: segfault at 7ff42580f057 ip 000055ad21554260 sp 00007ff4047f7068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;/P&gt;

&lt;H2&gt;Jul 31 08:51:34 splunkindexer1 kernel: splunkd[4092]: segfault at 7ff4224104f7 ip 000055ad21554260 sp 00007ff4013f8508 error 4 in splunkd[55ad1f3d2000+2e2b000]*  &lt;/H2&gt;

&lt;P&gt;This is from one of the crash logs.&lt;/P&gt;

&lt;HR /&gt;

&lt;P&gt;Received fatal signal 11 (Segmentation fault).&lt;BR /&gt;
 Cause:&lt;BR /&gt;
   No memory mapped at address [0x00000261CB7ECF].&lt;BR /&gt;
 Crashing thread: BatchSearch&lt;BR /&gt;
 "SNIP"&lt;/P&gt;

&lt;P&gt;Backtrace (PIC build):&lt;BR /&gt;
  [0x000056345C300260] st_decode_from_vbe + 0 (splunkd + 0x2182260)&lt;BR /&gt;
  [0x000056345C2EC4DA] ? (splunkd + 0x216E4DA)&lt;BR /&gt;
  [0x000056345C2EC7EF] _seek + 143 (splunkd + 0x216E7EF)&lt;BR /&gt;
  [0x000056345C2EF4A9] and_literals + 713 (splunkd + 0x21714A9)&lt;BR /&gt;
  [0x000056345C2F3316] ? (splunkd + 0x2175316)&lt;BR /&gt;
"SNIP" &lt;/P&gt;

&lt;P&gt;Last errno: 2&lt;BR /&gt;
Threads running: 11&lt;BR /&gt;
Runtime: 52652.730678s&lt;BR /&gt;
argv: [splunkd -p 8089 restart splunkd]&lt;BR /&gt;
Process renamed: [splunkd pid=3960] splunkd -p 8089 restart splunkd [process-runner]&lt;/P&gt;

&lt;H2&gt;Process renamed: [splunkd pid=3960] search --id=remote_sh1_scheduler_&lt;EM&gt;d5331&lt;/EM&gt;&lt;EM&gt;search&lt;/EM&gt;_RMD561462962f68d150_at_1562933700_3076_AAAAAAAA-1111-2222-AAAA-ADAAA6256C5C --maxbuckets=0 --ttl=60 --maxout=0 --maxtime=0 --lookups=1 --streaming --sidtype=normal --outCsv=true --acceptSrsLevel=1 --user=d5331 --pro --roles=power:user&lt;/H2&gt;</description>
    <pubDate>Wed, 30 Sep 2020 01:33:50 GMT</pubDate>
    <dc:creator>sylim_splunk</dc:creator>
    <dc:date>2020-09-30T01:33:50Z</dc:date>
    <item>
      <title>Segfault errors on a indexer in cluster</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422924#M15075</link>
      <description>&lt;P&gt;We have this issue very frequently which appeared to have started right after the last upgrade.&lt;BR /&gt;
Below kernel logs shows the frequency, Splunk process on the indexer appears running without restart so it appears to be from search processes.  &lt;/P&gt;

&lt;HR /&gt;

&lt;P&gt;*Linux splunkindexer1  2.6.32-754.9.1.el6.x86_64 #1 SMP Wed Dec 21 10:08:21 PST 2018 x86_64 x86_64 x86_64 GNU/Linux&lt;BR /&gt;
-bash-4.1$ cat /var/log/messages | grep -i kernel| tail&lt;BR /&gt;
Jul 31 08:16:24 splunkindexer1 kernel: splunkd[3149]: segfault at 7ff425810057 ip 000055ad21554260 sp 00007ff4047f8068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:19:34 splunkindexer1 kernel: splunkd[7907]: segfault at 7ff42540e057 ip 000055ad21554260 sp 00007ff4043f6068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:20:30 splunkindexer1 kernel: splunkd[22411]: segfault at 7ff42560f057 ip 000055ad21554260 sp 00007ff4045f7068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;BR /&gt;
Jul 31 08:21:07 splunkindexer1 kernel: splunkd[30162]: segfault at 7ff42580f057 ip 000055ad21554260 sp 00007ff4047f7068 error 4 in splunkd[55ad1f3d2000+2e2b000]&lt;/P&gt;

&lt;H2&gt;Jul 31 08:51:34 splunkindexer1 kernel: splunkd[4092]: segfault at 7ff4224104f7 ip 000055ad21554260 sp 00007ff4013f8508 error 4 in splunkd[55ad1f3d2000+2e2b000]*  &lt;/H2&gt;

&lt;P&gt;This is from one of the crash logs.&lt;/P&gt;

&lt;HR /&gt;

&lt;P&gt;Received fatal signal 11 (Segmentation fault).&lt;BR /&gt;
 Cause:&lt;BR /&gt;
   No memory mapped at address [0x00000261CB7ECF].&lt;BR /&gt;
 Crashing thread: BatchSearch&lt;BR /&gt;
 "SNIP"&lt;/P&gt;

&lt;P&gt;Backtrace (PIC build):&lt;BR /&gt;
  [0x000056345C300260] st_decode_from_vbe + 0 (splunkd + 0x2182260)&lt;BR /&gt;
  [0x000056345C2EC4DA] ? (splunkd + 0x216E4DA)&lt;BR /&gt;
  [0x000056345C2EC7EF] _seek + 143 (splunkd + 0x216E7EF)&lt;BR /&gt;
  [0x000056345C2EF4A9] and_literals + 713 (splunkd + 0x21714A9)&lt;BR /&gt;
  [0x000056345C2F3316] ? (splunkd + 0x2175316)&lt;BR /&gt;
"SNIP" &lt;/P&gt;

&lt;P&gt;Last errno: 2&lt;BR /&gt;
Threads running: 11&lt;BR /&gt;
Runtime: 52652.730678s&lt;BR /&gt;
argv: [splunkd -p 8089 restart splunkd]&lt;BR /&gt;
Process renamed: [splunkd pid=3960] splunkd -p 8089 restart splunkd [process-runner]&lt;/P&gt;

&lt;H2&gt;Process renamed: [splunkd pid=3960] search --id=remote_sh1_scheduler_&lt;EM&gt;d5331&lt;/EM&gt;&lt;EM&gt;search&lt;/EM&gt;_RMD561462962f68d150_at_1562933700_3076_AAAAAAAA-1111-2222-AAAA-ADAAA6256C5C --maxbuckets=0 --ttl=60 --maxout=0 --maxtime=0 --lookups=1 --streaming --sidtype=normal --outCsv=true --acceptSrsLevel=1 --user=d5331 --pro --roles=power:user&lt;/H2&gt;</description>
      <pubDate>Wed, 30 Sep 2020 01:33:50 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422924#M15075</guid>
      <dc:creator>sylim_splunk</dc:creator>
      <dc:date>2020-09-30T01:33:50Z</dc:date>
    </item>
    <item>
      <title>Re: Segfault errors on a indexer in cluster</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422925#M15076</link>
      <description>&lt;P&gt;This could have been caused by some corrupted buckets when searches run against them.&lt;BR /&gt;
You may want to fix the buckets and try the same search to see if it fixes it.&lt;BR /&gt;
Follow the steps below to get list of buckets suspected corrupted.&lt;/P&gt;

&lt;P&gt;*** How to get the list of corrupt buckets ***&lt;BR /&gt;
1. @the indexer, cd to $SPLUNK_HOME/var/log/splunk&lt;BR /&gt;
2. Run below&lt;BR /&gt;
$ grep "MAP:" crash-2019-07-31*.log |grep "/opt/splunk/storage"&lt;BR /&gt;
    "/opt/splunk/storage" varies according to your deployment set up and is taken from the line below in crash log.&lt;BR /&gt;
    &lt;EM&gt;crash-2019-07-31-00:15:17.log: &lt;BR /&gt;
MAP: 7f00e9cdb000-7f00ea000000 r--s 00000000 fd:03 563872524                  /opt/splunk/storage/hot/myindex1/rb_1560184689_1559942722_7530_AAAAAAAA-BBBB-1111-8C82-ABAD1EDD033D/1560184689-1560184620-11473276039248555956.tsidx&lt;/EM&gt;&lt;BR /&gt;
3. It will return the problematic buckets. From the above example, the bucket location is &lt;EM&gt;/opt/splunk/storage/hot/myindex1/rb_1560184689_1559942722_7530_AAAAAAAA-BBBB-1111-8C82-ABAD1EDD033D&lt;/EM&gt;&lt;/P&gt;

&lt;P&gt;*** How to fix the corrupted buckets ***&lt;BR /&gt;
Rebuilding the bucket using fsck should fix the problem. Follow the steps to rebuild buckets:&lt;BR /&gt;
0. @CM, splunk enable maintenance-mode&lt;BR /&gt;
1. @Anonymous, splunk offline&lt;BR /&gt;
2. @Anonymous, for all the buckets from above, run splunk fsck repair --one-bucket --bucket-path="path_from_above"&lt;BR /&gt;
i.e:&lt;BR /&gt;
&lt;EM&gt;splunk fsck repair --one-bucket --bucket-path=/opt/splunk/storage/hot/myindex1/rb_1560184689_1559942722_7530_AAAAAAAA-BBBB-1111-8C82-ABAD1EDD033D&lt;/EM&gt;&lt;BR /&gt;
3. @Anonymous, splunk start&lt;BR /&gt;
4. @CM, splunk disable maintenance-mode&lt;/P&gt;

&lt;P&gt;If this is not helping improve the situation please contact Splunk Support with details of deployment architecture and a drag from the indexer.&lt;/P&gt;</description>
      <pubDate>Wed, 30 Sep 2020 01:33:53 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422925#M15076</guid>
      <dc:creator>sylim_splunk</dc:creator>
      <dc:date>2020-09-30T01:33:53Z</dc:date>
    </item>
    <item>
      <title>Re: Segfault errors on a indexer in cluster</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422926#M15077</link>
      <description>&lt;P&gt;That error means that a process (splunkd) has attempted to access memory that is not assigned to it. I believe this is/was a known bug in Splunk 7.1.x and below.&lt;/P&gt;</description>
      <pubDate>Wed, 31 Jul 2019 17:23:17 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422926#M15077</guid>
      <dc:creator>codebuilder</dc:creator>
      <dc:date>2019-07-31T17:23:17Z</dc:date>
    </item>
    <item>
      <title>Re: Segfault errors on a indexer in cluster</title>
      <link>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422927#M15078</link>
      <description>&lt;P&gt;Known issue (SPL-153976) and fixed as part of 7.1.3&lt;/P&gt;</description>
      <pubDate>Wed, 31 Jul 2019 17:26:41 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Deployment-Architecture/Segfault-errors-on-a-indexer-in-cluster/m-p/422927#M15078</guid>
      <dc:creator>codebuilder</dc:creator>
      <dc:date>2019-07-31T17:26:41Z</dc:date>
    </item>
  </channel>
</rss>

