Reporting

encoding of exported CSV

vanvan
Path Finder

Hi everyone,

I have produced a search, which formats events in a table with couple of columns. The data and column names use Cyrillic words and in the GUI these look just fine. However, when I try to export the table as CSV (via the "Export To" option) the data and column names are encoded incorrectly and are not readable. 

Is there a setting which I can change so that this problem is fixed?

 

I've searched the other topics here in Communities, but didn't find an asnwer, e.g.:

https://community.splunk.com/t5/Splunk-Search/Why-are-special-characters-replaced-with-UTF-8-after-e...

https://community.splunk.com/t5/Getting-Data-In/Korean-character-is-broken-when-I-export-the-query-r... 

Any help is appreciated,

Thanks!

Labels (2)
Tags (3)
0 Karma

scelikok
Champion

Hi @vanvan,

If you see the exported CSV file readable in Notepad++/Sublime but in Excel, below fix should work for you. Excel needs BOM characters at the beginning of file to understand UTF8.

Please update readall function as below and restart Splunk service. This is for Splunk 8.0 and later.

$SPLUNK_HOME/lib/python3.7/site-packages/splunk/rest/__init__.py

def readall(self, blocksize=32768):
        """
        Returns a generator reading blocks of data from the response
        until all data has been read
        """
        response = self.response
        import codecs
        counter = 0;
        while True:
            data = response.read(blocksize)
            if not data:
                break
            if counter == 0:
                data = b"".join((codecs.BOM_UTF8, data))
                counter += 1
            yield data

 

If this reply helps you an upvote is appreciated.

vanvan
Path Finder

Thanks!

This is interesting, I'll try it.

0 Karma
.conf21 Now Fully Virtual!
Register for FREE Today!

We've made .conf21 totally virtual and totally FREE! Our completely online experience will run from 10/19 through 10/20 with some additional events, too!