<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic TCP Buffer Options in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/TCP-Buffer-Options/m-p/214707#M42253</link>
    <description>&lt;P&gt;We are using a Cloud Foundry for an Internal Cloud Implementation. We are migrating applications and hence using TCP streaming versus the Splunk Forwarding Agent to stream logs. The Cloud Support Team is observing errors logged because the TCP stream is backing up. I am using a vanilla TCP configuration. Are there extra configurations I can add to the stanza to increase the buffer to prevent messages from dropping?&lt;/P&gt;

&lt;P&gt;Sample Vanilla TCP configuration:&lt;BR /&gt;
[tcp://3301]&lt;BR /&gt;
connection_host = dns&lt;BR /&gt;
index = index_name&lt;BR /&gt;
sourcetype = rfc5424_syslog&lt;/P&gt;

&lt;P&gt;Logs from the Cloud Foundry Doppler Service:&lt;BR /&gt;
{"timestamp":1474911116.219854593,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too full","data":{"appId":"d41c5c78-955a-4148-b575-cf868dc0b6fe","destination":"syslog://tlaloga1.dev.prod.travp.net:3303","dropped":99,"total_dropped":496},"file":"/var/vcap/data/compile/doppler/loggregator/src/truncatingbuffer/truncating_buffer.go","line":112,"method":"truncatingbuffer.(*TruncatingBuffer).forwardMessage"}&lt;BR /&gt;
{"timestamp":1474911142.510904074,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too full","data":{"appId":"d41c5c78-955a-4148-b575-cf868dc0b6fe","destination":"syslog://tlaloga1.dev.prod.travp.net:3303","dropped":99,"total_dropped":595},"file":"/var/vcap/data/compile/doppler/loggregator/src/truncatingbuffer/truncating_buffer.go","line":112,"method":"truncatingbuffer.(*TruncatingBuffer).forwardMessage"}&lt;BR /&gt;
{"timestamp":1474911161.138316393,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too &lt;/P&gt;</description>
    <pubDate>Tue, 29 Sep 2020 11:10:15 GMT</pubDate>
    <dc:creator>ezajac</dc:creator>
    <dc:date>2020-09-29T11:10:15Z</dc:date>
    <item>
      <title>TCP Buffer Options</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/TCP-Buffer-Options/m-p/214707#M42253</link>
      <description>&lt;P&gt;We are using a Cloud Foundry for an Internal Cloud Implementation. We are migrating applications and hence using TCP streaming versus the Splunk Forwarding Agent to stream logs. The Cloud Support Team is observing errors logged because the TCP stream is backing up. I am using a vanilla TCP configuration. Are there extra configurations I can add to the stanza to increase the buffer to prevent messages from dropping?&lt;/P&gt;

&lt;P&gt;Sample Vanilla TCP configuration:&lt;BR /&gt;
[tcp://3301]&lt;BR /&gt;
connection_host = dns&lt;BR /&gt;
index = index_name&lt;BR /&gt;
sourcetype = rfc5424_syslog&lt;/P&gt;

&lt;P&gt;Logs from the Cloud Foundry Doppler Service:&lt;BR /&gt;
{"timestamp":1474911116.219854593,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too full","data":{"appId":"d41c5c78-955a-4148-b575-cf868dc0b6fe","destination":"syslog://tlaloga1.dev.prod.travp.net:3303","dropped":99,"total_dropped":496},"file":"/var/vcap/data/compile/doppler/loggregator/src/truncatingbuffer/truncating_buffer.go","line":112,"method":"truncatingbuffer.(*TruncatingBuffer).forwardMessage"}&lt;BR /&gt;
{"timestamp":1474911142.510904074,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too full","data":{"appId":"d41c5c78-955a-4148-b575-cf868dc0b6fe","destination":"syslog://tlaloga1.dev.prod.travp.net:3303","dropped":99,"total_dropped":595},"file":"/var/vcap/data/compile/doppler/loggregator/src/truncatingbuffer/truncating_buffer.go","line":112,"method":"truncatingbuffer.(*TruncatingBuffer).forwardMessage"}&lt;BR /&gt;
{"timestamp":1474911161.138316393,"process_id":17528,"source":"doppler","log_level":"warn","message":"TB: Output channel too &lt;/P&gt;</description>
      <pubDate>Tue, 29 Sep 2020 11:10:15 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/TCP-Buffer-Options/m-p/214707#M42253</guid>
      <dc:creator>ezajac</dc:creator>
      <dc:date>2020-09-29T11:10:15Z</dc:date>
    </item>
    <item>
      <title>Re: TCP Buffer Options</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/TCP-Buffer-Options/m-p/214708#M42254</link>
      <description>&lt;P&gt;Rather than increase the buffer size you can scale the number of Dopplers available. Below is a link to a guide to how many Dopplers and Traffic Controllers to plan for. &lt;/P&gt;

&lt;P&gt;&lt;A href="https://discuss.pivotal.io/hc/en-us/articles/225564028-How-to-Calculate-the-Loggregators-Message-Throughput,Rather"&gt;https://discuss.pivotal.io/hc/en-us/articles/225564028-How-to-Calculate-the-Loggregators-Message-Throughput,Rather&lt;/A&gt; than scaling the buffer size the approach we recommend is scaling the number of dopplers. See  &lt;A href="https://discuss.pivotal.io/hc/en-us/articles/225564028-How-to-Calculate-the-Loggregators-Message-Throughput"&gt;https://discuss.pivotal.io/hc/en-us/articles/225564028-How-to-Calculate-the-Loggregators-Message-Throughput&lt;/A&gt; for a guide to scaling more dopplers (and possibly Traffic Controllers) as well. &lt;/P&gt;</description>
      <pubDate>Wed, 30 Nov 2016 18:10:16 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/TCP-Buffer-Options/m-p/214708#M42254</guid>
      <dc:creator>ahev</dc:creator>
      <dc:date>2016-11-30T18:10:16Z</dc:date>
    </item>
  </channel>
</rss>

