<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: advice for when you have more than 100 automatically extracted fields? in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115958#M24229</link>
    <description>&lt;P&gt;One always learns. This is an amazing command. Modifing limits.conf under [kv] with limit = 300 helped my on my all-in-one lab box. Cannot make it work on my SH-IDX deployment. For now this helped me a lot! Thx.&lt;/P&gt;</description>
    <pubDate>Tue, 05 May 2020 11:28:18 GMT</pubDate>
    <dc:creator>tomasmoser</dc:creator>
    <dc:date>2020-05-05T11:28:18Z</dc:date>
    <item>
      <title>advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115951#M24222</link>
      <description>&lt;P&gt;It seems that if you have a lot of fields being extracted automatically, like via &lt;CODE&gt;INDEXED_EXTRACTIONS=csv&lt;/CODE&gt; or via automatic kv extraction,   beyond any fields that are explicitly mentioned in your search, Splunk 6.0 will only allow itself to automatically extract about 100 more fields. &lt;/P&gt;

&lt;P&gt;This really prevents certain commands like  &lt;CODE&gt;fieldsummary&lt;/CODE&gt; or &lt;CODE&gt;transpose&lt;/CODE&gt; from working properly. &lt;/P&gt;

&lt;P&gt;Basically if you have more than 100 fields, you don't know what their names are, and you need to get that list of field names or values in Splunk,  you wont be able to using the Splunk search language.  Whatever search language you use - &lt;CODE&gt;fieldsummary&lt;/CODE&gt; or &lt;CODE&gt;stats first(*) as * | transpose&lt;/CODE&gt;, Splunk will ignore &lt;EM&gt;some&lt;/EM&gt; of your extraction rules each time and you'll always end up with an incomplete list of fields. Which one it ignores seems somewhat random, so certain fields will be appearing and dissappearing from your results over time.&lt;/P&gt;

&lt;P&gt;Does anyone know if anything can be set in limits.conf or in the search language to override this behavior? &lt;/P&gt;

&lt;P&gt;I believe it is related to these other posts &lt;BR /&gt;
&lt;A href="http://answers.splunk.com/answers/82252/how-many-field-extract-in-splunk"&gt;http://answers.splunk.com/answers/82252/how-many-field-extract-in-splunk&lt;/A&gt;&lt;BR /&gt;
&lt;A href="http://answers.splunk.com/answers/117884/fields-not-automatically-extracting"&gt;http://answers.splunk.com/answers/117884/fields-not-automatically-extracting&lt;/A&gt;&lt;/P&gt;

&lt;P&gt;Also, I've already tried setting &lt;CODE&gt;maxcols&lt;/CODE&gt; in the &lt;CODE&gt;[kv]&lt;/CODE&gt; stanza in limits.conf and it has no effect here - probably because that key only affects keys being generated by kv and autokv configurations, and this is generated via &lt;CODE&gt;INDEXED_EXTRACTIONS=csv&lt;/CODE&gt;&lt;/P&gt;

&lt;P&gt;There is one very limited workaround that I have found - if you mention all the fields in your search explicitly somehow, like with an enormous &lt;CODE&gt;fields&lt;/CODE&gt; clause,  this forces Splunk to extract them, no matter how many of them there are.   However that doesn't help situations where you really don't know in advance or with certainty what they will be.  &lt;/P&gt;

&lt;P&gt;In my particular case there is a csv sourcetype where we don't know in advance what the fields are going to be.  But we need the field list and using the search language itself to get it is the only way we have.    Depending on various factors there are hundreds of fields that might be present in the sourcetype, and on any one customer environment there &lt;EM&gt;will&lt;/EM&gt; be about 100 - 140 fields present. &lt;/P&gt;</description>
      <pubDate>Tue, 01 Apr 2014 23:13:16 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115951#M24222</guid>
      <dc:creator>sideview</dc:creator>
      <dc:date>2014-04-01T23:13:16Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115952#M24223</link>
      <description>&lt;P&gt;For completeness' sake, does &lt;CODE&gt;fields *&lt;/CODE&gt; or &lt;CODE&gt;fields + *&lt;/CODE&gt; do anything in this case?&lt;/P&gt;</description>
      <pubDate>Wed, 02 Apr 2014 07:26:55 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115952#M24223</guid>
      <dc:creator>martin_mueller</dc:creator>
      <dc:date>2014-04-02T07:26:55Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115953#M24224</link>
      <description>&lt;P&gt;Good point. I forgot to mention that, but no they have no effect on this problem.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Apr 2014 16:23:28 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115953#M24224</guid>
      <dc:creator>sideview</dc:creator>
      <dc:date>2014-04-02T16:23:28Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115954#M24225</link>
      <description>&lt;P&gt;Bugger....&lt;/P&gt;</description>
      <pubDate>Wed, 02 Apr 2014 16:47:32 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115954#M24225</guid>
      <dc:creator>martin_mueller</dc:creator>
      <dc:date>2014-04-02T16:47:32Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115955#M24226</link>
      <description>&lt;P&gt;I am using Splunk 6.1.4.  I had the same issue, tried modifying maxcols, and it did not work for me.&lt;/P&gt;

&lt;P&gt;So I took a peek at $SPLUNK_HOME/etc/system/default/limits.conf and observed a couple more parameters to change.&lt;/P&gt;

&lt;P&gt;I modified $SPLUNK_HOME/etc/system/local/limits.conf to contain the following:&lt;/P&gt;

&lt;BLOCKQUOTE&gt;
&lt;PRE&gt;&lt;CODE&gt;[kv]
# when non-zero, the point at which kv should stop creating new columns
maxcols  = 1800
# maximum number of keys auto kv can generate
limit    = 500
# truncate _raw to to this size and then do auto KV
maxchars = 102400
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;That seems to have worked for me.&lt;/P&gt;
&lt;/BLOCKQUOTE&gt;</description>
      <pubDate>Fri, 17 Oct 2014 21:26:43 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115955#M24226</guid>
      <dc:creator>veganjay</dc:creator>
      <dc:date>2014-10-17T21:26:43Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115956#M24227</link>
      <description>&lt;P&gt;I came across this issue as well.&lt;BR /&gt;
On 6.2 our default/limits.conf had the limit set to 50, we increased this in our local/limits.conf to 150 but after a splunk restart that didn't seem to have fixed it.&lt;/P&gt;

&lt;P&gt;As a work around I used &lt;CODE&gt;| fields -* | extract pairdelim=",", kvdelim="=", auto=f, limit=200, mv_add=t&lt;/CODE&gt; to drop the automatically extracted fields and re-extract them all. This also has the advantage that Multiple fields are extracted properly rather than only the 1st found occurrence.&lt;/P&gt;</description>
      <pubDate>Thu, 21 May 2015 14:19:31 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115956#M24227</guid>
      <dc:creator>Harveyj</dc:creator>
      <dc:date>2015-05-21T14:19:31Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115957#M24228</link>
      <description>&lt;P&gt;&lt;CODE&gt;+1&lt;/CODE&gt; for you.&lt;/P&gt;

&lt;P&gt;Despite being old, this answer is still quickly found through Google and addressed my problem. The alternative option that I used based on this answer is to explicitly table out all expected fields instead.&lt;/P&gt;

&lt;P&gt;So, with a limit of 100, rare_field will not show up with:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;| index=yourindex
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;Even if you change "Interesting Fields" to "All Fields" (instead of &amp;gt;1% coverage). However, this search will return it where applicable:&lt;/P&gt;

&lt;PRE&gt;&lt;CODE&gt;| index=yourindex
| table rare_field
&lt;/CODE&gt;&lt;/PRE&gt;

&lt;P&gt;If you can identify all of your important fields, this will also work without changing limits.conf&lt;/P&gt;

&lt;P&gt;Cheers,&lt;BR /&gt;
Jacob&lt;/P&gt;</description>
      <pubDate>Fri, 01 May 2020 16:49:39 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115957#M24228</guid>
      <dc:creator>jacobpevans</dc:creator>
      <dc:date>2020-05-01T16:49:39Z</dc:date>
    </item>
    <item>
      <title>Re: advice for when you have more than 100 automatically extracted fields?</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115958#M24229</link>
      <description>&lt;P&gt;One always learns. This is an amazing command. Modifing limits.conf under [kv] with limit = 300 helped my on my all-in-one lab box. Cannot make it work on my SH-IDX deployment. For now this helped me a lot! Thx.&lt;/P&gt;</description>
      <pubDate>Tue, 05 May 2020 11:28:18 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/advice-for-when-you-have-more-than-100-automatically-extracted/m-p/115958#M24229</guid>
      <dc:creator>tomasmoser</dc:creator>
      <dc:date>2020-05-05T11:28:18Z</dc:date>
    </item>
  </channel>
</rss>

