Hi,
In "splunk_app_db_connect" I've defined this input configuration:
[ALERT_SNO_MISMATCH]
connection = PDBAPP_SYSTEM_SCAN
description = Search all 5 minutes for "SNO mismatch"
disabled = 0
index = temp
index_time_mode = dbColumn
input_timestamp_column_number = 1
input_timestamp_format = yyyy-MM-dd HH:mm:ss.SSS
interval = */5 * * * *
mode = batch
query = select to_char(ORIGINATING_TIMESTAMP,'YYYY-MM-DD HH24:MI:SS.FF3') AS TIME, MESSAGE_TEXT\
from v$diag_alert_ext\
where component_id like '%rdbms%'\
and message_text like '%SNO mismatch for LOCAL TRAN%'\
and originating_timestamp>sysdate-1/288;
query_timeout = 300
sourcetype = csv
fetch_size = 10000
Running the SQL statement in SQL explorer I receive such a result set:
TIME | MESSAGE_TEXT |
2020-10-21 15:30:08.379 | SNO mismatch for LOCAL TRAN 14.4.82391 |
2020-10-21 15:31:07.907 | SNO mismatch for LOCAL TRAN 11.30.78254 |
2020-10-21 15:31:08.709 | SNO mismatch for LOCAL TRAN 27.33.68134 |
2020-10-21 15:31:42.296 | SNO mismatch for LOCAL TRAN 20.28.52198 |
But as repetitive running db_input-job it fails always with this error:
2020-10-21 15:30:33.144 +0200 [QuartzScheduler_Worker-21] ERROR org.easybatch.core.job.BatchJob - Unable to process Record: {header=[number=1, source="ALERT_SNO_MISMATCH", creationDate="Wed Oct 21 15:30:33 CEST 2020"], payload=[HikariProxyResultSet@1360101480 wrapping oracle.jdbc.driver.ForwardOnlyResultSet@409d22a]}
2020-10-21 15:30:33.144 +0200 [QuartzScheduler_Worker-21] ERROR org.easybatch.core.job.BatchJob - Unable to process Record: {header=[number=2, source="ALERT_SNO_MISMATCH", creationDate="Wed Oct 21 15:30:33 CEST 2020"], payload=[HikariProxyResultSet@1360101480 wrapping oracle.jdbc.driver.ForwardOnlyResultSet@409d22a]}
It does this error for each row. At the end nothing got stored in index.
Any idea what I'm doing wrong?
Thanks in advance, Aaron
Were you able to find a solution for this?