nosql - Loading CSV (or TSV) into MarkLogic with automatic encoding -


Using MLCP (MarkLogic Content) I have not successfully loaded a very clean (plain English, no fancy symbols or images) CSV file in Markloadic I HTML code (MarkLogic Content) Pump so that it's the first row in column names , And I have learned that when I try to load something that is not clear (i.e. mixed with other languages ​​and encoding) it fails

I have read from the injection guide () that the encoding is not controllable with the MLCP, so I decided to give the Java API and XMPP Extra.

While using the Java API and I'm getting: Invalid UTF-8 escape sequence on line 1549 - Document UTF-8 is not encoded

If I load it with Xdmp If I try, with automatic encoding in the question console or in a stream on the information studio, it loads without a problem, but MarkLogic does not take the first row as column names, but its full form Il am is that I do not take it as a document.

Is there any way to load the CSV file without the encoding problem and take it as the column name in the first line?

Thanks in advance.

Record loader can:

  CONFIGURATION_CLASSNAME = com Marklogic.recordloader.xcc.DelimitedDataConfiguration FIELD_DELIMITER =, RECORD_NAME = my-root-element-name  

play re-enter Dloader.sh those properties and your CSV file (Files) with Recordloader hopes that the first row will be a list of headers, and that will change into the names of those elements. Adjust yourself to adjust my-root-element-name and set the INPUT_ENCODING to whatever encoding is required.

Look for more configuration options


Comments

Popular posts from this blog

import - Python ImportError: No module named wmi -

Editing Python Class in Shell and SQLAlchemy -

lua - HowTo create a fuel bar -