Getting Data In

strip sensitive data before indexing

Path Finder

Hi all,

We run several tools in our environment for network inspection and the logging it provides logs things like NTLM creds, HTTP Basic Auth etc. We'd like to strip this data off before indexing, or at least 'sanitize' it so we dont index usernames/passwords. Can anyone point me in the right direction to solve this?

Example data:

"signature":"ET POLICY HTTP traffic on port 443 (PROPFIND)",
"category":"Potentially Bad Traffic",
"http_user_agent":"Mozilla\/5.0 (Linux) mirall\/2.3.3",
"payload_printable":"PROPFIND \/owncloud\/remote.php\/webdav\/ HTTP\/1.1\r\n
Depth: 0\r\n
Authorization: Basic REDACTEDBASE64PASSWORDHERE==\r\n
User-Agent: Mozilla\/5.0 (Linux) mirall\/2.3.3\r\n
Accept: \/\r\n
Content-Type: text\/xml; charset=utf-8\r\nCookie: oc_sessionPassphrase=redacted \r\nContent-Length: 105\r\n
Connection: Keep-Alive\r\n
Accept-Encoding: gzip, deflate\r\nAccept-Language: en-US,*\r\n
Host:\r\n\r\n\n\n \n \n <\/d:prop>\n<\/d:propfind>\n",

0 Karma
1 Solution




If this reply helps you, Karma would be appreciated.
Get Updates on the Splunk Community!

Build Scalable Security While Moving to Cloud - Guide From Clayton Homes

 Clayton Homes faced the increased challenge of strengthening their security posture as they went through ...

Mission Control | Explore the latest release of Splunk Mission Control (2.3)

We’re happy to announce the release of Mission Control 2.3 which includes several new and exciting features ...

Cloud Platform | Migrating your Splunk Cloud deployment to Python 3.7

Python 2.7, the last release of Python 2, reached End of Life back on January 1, 2020. As part of our larger ...