Deployment Architecture

Remove excess buckets via REST API

danman06
Explorer

We have some problems with excess buckets from time to time, and I am writing a python program to check for excess buckets. Then, I want to take that list and purge them. In the documentation, I see the endpoint /cluster/master/control/control/prune_index. My question is, is this the correct API end point to call? Do I need to pass it a list of buckets? Or am I off track?

0 Karma

jiaqya
Builder

how to avoid using user:pass if possible ?

0 Karma

harsmarvania57
Ultra Champion

Hi @danman06,

Yes, that is correct. REST API endpoint for removing excess buckets. Based on the documentation, if you want to remove excess buckets from certain indexes, you need to supply index name in that REST API. If you do not provide any index name, Splunk will remove excess buckets from all indexes.

index   Optional. The index from which to remove excess bucket copies. If not specified, the POST operation clears excess bucket copies across all indexes.

With Index name

curl -k -u admin:pass https://localhost:8089/services/cluster/master/control/control/prune_index -d index="my_index"

Without index (Which means it will remove excess bucket from all indexes)

curl -k -u admin:pass https://localhost:8089/services/cluster/master/control/control/prune_index

danman06
Explorer

Excellent, I'll do some testing and thanks for the prompt reply.

0 Karma
Got questions? Get answers!

Join the Splunk Community Slack to learn, troubleshoot, and make connections with fellow Splunk practitioners in real time!

Meet up IRL or virtually!

Join Splunk User Groups to connect and learn in-person by region or remotely by topic or industry.

Get Updates on the Splunk Community!

Index This | What travels the world but is also stuck in place?

April 2026 Edition  Hayyy Splunk Education Enthusiasts and the Eternally Curious!   We’re back with this ...

Discover New Use Cases: Unlock Greater Value from Your Existing Splunk Data

Realizing the full potential of your Splunk investment requires more than just understanding current usage; it ...

Continue Your Journey: Join Session 2 of the Data Management and Federation Bootcamp ...

As data volumes continue to grow and environments become more distributed, managing and optimizing data ...