<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: s3 - Multiple directories data to be ingested into Splunk Cloud in Getting Data In</title>
    <link>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750510#M119236</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/263698"&gt;@shoaibalimir&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When you assessed and didnt get the required outcome - what is the issue you had specifically?&lt;/P&gt;&lt;P&gt;Is this a one-time ingestion of historic files already in S3, or are you wanting to ingest on an ongoing basis (I assume the latter?).&lt;/P&gt;&lt;P&gt;Personally I would avoid Generic-S3 as it relies on checkpoint files and can get messy quickly. SQS based S3 is the way to go here I believe.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Check out&amp;nbsp;&lt;A href="https://splunk.github.io/splunk-add-on-for-amazon-web-services/SQS-basedS3/" target="_blank"&gt;https://splunk.github.io/splunk-add-on-for-amazon-web-services/SQS-basedS3/&lt;/A&gt;&amp;nbsp;for more details on setting up SQS-based-S3 input. Its also worth nothing that the dynamic parts of the path shouldnt be a problem. If you have requirements to put them into specific indexes depending on the dynamic values then you can configure this when you setup the event notification (&lt;A href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/enable-event-notifications.html" target="_blank"&gt;https://docs.aws.amazon.com/AmazonS3/latest/userguide/enable-event-notifications.html&lt;/A&gt;) and will probably need multiple SQS. Alternatively you could use props/transforms to route to the correct index at ingest time.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-unicode-emoji" title=":glowing_star:"&gt;🌟&lt;/span&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Did this answer help you?&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;If so, please consider:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Adding karma to show it was useful&lt;/LI&gt;&lt;LI&gt;Marking it as the solution if it resolved your issue&lt;/LI&gt;&lt;LI&gt;Commenting if you need any clarification&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Your feedback encourages the volunteers in this community to continue contributing&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Sat, 26 Jul 2025 14:22:17 GMT</pubDate>
    <dc:creator>livehybrid</dc:creator>
    <dc:date>2025-07-26T14:22:17Z</dc:date>
    <item>
      <title>s3 - Multiple directories data to be ingested into Splunk Cloud</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750448#M119226</link>
      <description>&lt;P&gt;Hi Community,&lt;/P&gt;&lt;P&gt;I'm exploring ways to ingest data into Splunk Cloud from a Amazon s3 Bucket which has multiple directories and multiple files to be ingested onto Splunk.&lt;/P&gt;&lt;P&gt;Now, I have assessed the Generic s3, SQS-s3 and the Data Manager Inputs for AWS available on Splunk but am not getting the required outcome.&lt;/P&gt;&lt;P&gt;My use case is given below:&lt;/P&gt;&lt;P&gt;There's a s3 bucket named as exampledatastore, in that there's a directory named as&amp;nbsp;statichexcodedefinition, in that there're multiple message Ids and Dates.&lt;/P&gt;&lt;P&gt;The s3 example structure is given below:&lt;/P&gt;&lt;P&gt;s3://exampledatastore/statichexcodedefinition/{messageId}/functionname/{date}/* - functionnameattribute&lt;/P&gt;&lt;P&gt;Where the {messageId} and the {date} values are dynamic. And I have a start date to begin with but the messageId varies.&lt;/P&gt;&lt;P&gt;Please can you assist me on this on how to get the data into Splunk.&lt;BR /&gt;&lt;BR /&gt;Many Thanks!&lt;/P&gt;</description>
      <pubDate>Mon, 28 Jul 2025 10:23:44 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750448#M119226</guid>
      <dc:creator>shoaibalimir</dc:creator>
      <dc:date>2025-07-28T10:23:44Z</dc:date>
    </item>
    <item>
      <title>Re: s3 - Multiple directories data to be ingested into Splunk Cloud</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750510#M119236</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/263698"&gt;@shoaibalimir&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When you assessed and didnt get the required outcome - what is the issue you had specifically?&lt;/P&gt;&lt;P&gt;Is this a one-time ingestion of historic files already in S3, or are you wanting to ingest on an ongoing basis (I assume the latter?).&lt;/P&gt;&lt;P&gt;Personally I would avoid Generic-S3 as it relies on checkpoint files and can get messy quickly. SQS based S3 is the way to go here I believe.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Check out&amp;nbsp;&lt;A href="https://splunk.github.io/splunk-add-on-for-amazon-web-services/SQS-basedS3/" target="_blank"&gt;https://splunk.github.io/splunk-add-on-for-amazon-web-services/SQS-basedS3/&lt;/A&gt;&amp;nbsp;for more details on setting up SQS-based-S3 input. Its also worth nothing that the dynamic parts of the path shouldnt be a problem. If you have requirements to put them into specific indexes depending on the dynamic values then you can configure this when you setup the event notification (&lt;A href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/enable-event-notifications.html" target="_blank"&gt;https://docs.aws.amazon.com/AmazonS3/latest/userguide/enable-event-notifications.html&lt;/A&gt;) and will probably need multiple SQS. Alternatively you could use props/transforms to route to the correct index at ingest time.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-unicode-emoji" title=":glowing_star:"&gt;🌟&lt;/span&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Did this answer help you?&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;If so, please consider:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Adding karma to show it was useful&lt;/LI&gt;&lt;LI&gt;Marking it as the solution if it resolved your issue&lt;/LI&gt;&lt;LI&gt;Commenting if you need any clarification&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Your feedback encourages the volunteers in this community to continue contributing&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sat, 26 Jul 2025 14:22:17 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750510#M119236</guid>
      <dc:creator>livehybrid</dc:creator>
      <dc:date>2025-07-26T14:22:17Z</dc:date>
    </item>
    <item>
      <title>Re: s3 - Multiple directories data to be ingested into Splunk Cloud</title>
      <link>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750552#M119249</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.splunk.com/t5/user/viewprofilepage/user-id/170906"&gt;@livehybrid&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'll assess again with the SQS-s3 Connector, and I'll need to ingest both historic data as well as ongoing data stream.&lt;/P&gt;&lt;P&gt;By the initial observations I think I'll need to use multiple SQS-s3 Connectors or would need to use Lambda to process those into single SQS-s3 Connector.&lt;/P&gt;&lt;P&gt;Please let me know if there's any other alternative to this assumption.&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Mon, 28 Jul 2025 10:23:14 GMT</pubDate>
      <guid>https://community.splunk.com/t5/Getting-Data-In/s3-Multiple-directories-data-to-be-ingested-into-Splunk-Cloud/m-p/750552#M119249</guid>
      <dc:creator>shoaibalimir</dc:creator>
      <dc:date>2025-07-28T10:23:14Z</dc:date>
    </item>
  </channel>
</rss>

