Splunk Enterprise Security

"Concurrent Login Attempts Detected" CS results have hostnames and IPs showing as previous_src and src

wgawhh5hbnht
Communicator

We're getting false positives on the correlated search, "Concurrent Login Attempts Detected", because the previous_src and the src are the hostname and IP of the same system.

In ES if I use Investigate Asset Artifacts and enter the hostname, the IP is returned, and vice versa. To me that means the asset lookups are functioning properly, right?

Anyone know how to correct for this?

0 Karma
1 Solution

wgawhh5hbnht
Communicator

I opened a case with Splunk support & they said this an expected behavior, it either needs more work or there is something specific that's causing hundreds of these alerts per day. To reduce the noise I've eliminated all Domain Controller IPs & Hostnames. Next I added multiple looks for the IPs, this is done based on both src & previous_src:

| lookup dnslookup clientip as src OUTPUT clienthost as src_fqdn
| lookup dnslookup clientip as previous_src OUTPUT clienthost as previous_src_fqdn

I then remove any matches between src_fqdn & previous_src_fqdn:

| where (src_fqdn!=previous_src_fqdn)

I needed to remove the Fully Qualified Domain Name (FQDN) to match possible src/previous_src:

| rex field=src_fqdn "(?<src_host>^([^.]+))"
| rex field=previous_src_fqdn "(?<previous_src_host>^([^.]+))"

Then I added multiple where clauses to eliminate duplicates:

| where (src_host!=src) AND (src_host!=previous_src) AND (previous_src_host!=src) AND (previous_src_host!=previous_src)

While all of this has reduced the noise from thousands of alerts per day to hundreds, it unfortunately has added a lot of overhead. If anyone has any suggestions to speed this up, I would appreciate any assistance.

View solution in original post

0 Karma

wgawhh5hbnht
Communicator

I opened a case with Splunk support & they said this an expected behavior, it either needs more work or there is something specific that's causing hundreds of these alerts per day. To reduce the noise I've eliminated all Domain Controller IPs & Hostnames. Next I added multiple looks for the IPs, this is done based on both src & previous_src:

| lookup dnslookup clientip as src OUTPUT clienthost as src_fqdn
| lookup dnslookup clientip as previous_src OUTPUT clienthost as previous_src_fqdn

I then remove any matches between src_fqdn & previous_src_fqdn:

| where (src_fqdn!=previous_src_fqdn)

I needed to remove the Fully Qualified Domain Name (FQDN) to match possible src/previous_src:

| rex field=src_fqdn "(?<src_host>^([^.]+))"
| rex field=previous_src_fqdn "(?<previous_src_host>^([^.]+))"

Then I added multiple where clauses to eliminate duplicates:

| where (src_host!=src) AND (src_host!=previous_src) AND (previous_src_host!=src) AND (previous_src_host!=previous_src)

While all of this has reduced the noise from thousands of alerts per day to hundreds, it unfortunately has added a lot of overhead. If anyone has any suggestions to speed this up, I would appreciate any assistance.

0 Karma
Get Updates on the Splunk Community!

More Ways To Control Your Costs With Archived Metrics | Register for Tech Talk

Tuesday, May 14, 2024  |  11AM PT / 2PM ET Register to Attend Join us for this Tech Talk and learn how to ...

.conf24 | Personalize your .conf experience with Learning Paths!

Personalize your .conf24 Experience Learning paths allow you to level up your skill sets and dive deeper ...

Threat Hunting Unlocked: How to Uplevel Your Threat Hunting With the PEAK Framework ...

WATCH NOWAs AI starts tackling low level alerts, it's more critical than ever to uplevel your threat hunting ...