So I am trying to import 26 million rows in around 300 csv files (using bash to execute the bulk import code).
I come across a limitation in the number of csv's I can reference before I'm told , that the command is too long...
And some of those csv's have null values in columns.. in the normal load csv , I know how to deal with those, but with the bulk import (neo4j-admin.bat) I do not.
note that in regexp you need to use .* instead of * for "any character"
the null values are skipped during import
if the --ignore-empty-strings setting is set to true:
--ignore-empty-strings <true/false>
Whether or not empty string fields, i.e. "" from input source are ignored, i.e.
treated as null. Default value: false
You are a kind man for answering one of my questions again Michael.. :) Thanks.. I'll give it a shot right now. (that helps me having to adjust the data export out of oracle that I'm thought I'd have to labour through).
I don't know of such a limit in bash. Did you use teh regexps for the files?
And you can put all the command line options into a file too:
--f <file name>
File containing all arguments, used as an alternative to supplying all arguments
on the command line directly.Each argument can be on a separate line or multiple
arguments per line separated by space.Arguments containing spaces needs to be
quoted.Supplying other arguments in addition to this file argument is not
supported.
Thats a depreciated feature and therefore no longer documentated on neo4j (though in some blogs)
It works on a test example for me , so yeah, that fixes (till you get rid of the neo4j-import , my null value on bulk ) and the import using the file fixes my other issue.