"A few hours offset" can be a big contributing factor here. How big of an offset are we talking here? If you haven't changed the value of maxHotBuckets for the index, it defaults to auto, which the indexers use to the set the value to 3. If the timestamp of the data is all over the place (Some in the future, some really old data), Splunk ingests it but force roll a hot bucket to create a new one for the data with unusual timestamp and if it happens frequently, you find a lot of small hot buckets being created in a short span of time. Please check if the data coming from AWS is being accepted and stored "in the future" by running searches with index=yourindex sourcetype=aws:* earliest=+5m latest=+7d. If the volume is considerably large, this could be a big contributor to the error. Please look for errors like "Accepted time is suspiciously far away from the previous event's time". This can tell you if the events that got ingested have timestamps far back in the past and Splunk ended creating hot buckets for them. Creating a custom props.conf to define TIME_PREFIX, MAX_TIMESTAMP_LOOKAHEAD and TIME_FORMAT along with line breaking to ensure that Splunk reads the timestamp properly from your data. Doing a sanity check of the data itself. I've seen some log sources in some environments, where timestamp within the log source was all over the place all the time. Ended up with DATETIME_CONFIG = current to resolve the problem. Hope this helps, ###If it helps, kindly consider accepting as an answer/upvoting###
... View more