I am trying to import 2.3 TB of data onto an EC2 Box with 8 Cores 120 GB Ram and a 8 TB SSD. I have been able to load smaller datasets but am now scaling up to a larger dataset. The command to invoke the import is
~/../../../usr/bin/neo4j-admin import \
--nodes "import/uids-header.csv,import/uid_no.*" \
--nodes "import/age-header.csv,import/age_no.*" \
--nodes "import/gender-header.csv,import/gender_no.*" \
--nodes "import/ip-header.csv,import/ip_no.*" \
--nodes "import/device-header.csv,import/device_no.*" \
--nodes "import/os-header.csv,import/os_no.*" \
--nodes "import/browser-header.csv,import/browser_no.*" \
--nodes "import/identitylink-header.csv,import/idlink_no.*" \
--nodes "import/opti-header.csv,import/opti_no.*" \
--nodes "import/bluekai-header.csv,import/bk_no.*" \
--nodes "import/acxiom-header.csv,import/axm_no.*" \
--nodes "import/adobe-header.csv,import/adb_no.*" \
--nodes "import/lr-header.csv,import/lr_no.*" \
--nodes "import/viant-header.csv,import/vnt_no.*" \
--nodes "import/ga-header.csv,import/ggl_no.*" \
--nodes "import/segment-header.csv,import/seg_no.*" \
--nodes "import/email-header.csv,import/email_no.*" \
--nodes "import/country-header.csv,import/cntry_no.*" \
--nodes "import/citystate-header.csv,import/city_no.*" \
--relationships:OBSERVED_WITH "import/rels-header.csv,import/opti_li.*,import/idlink_li.*,import/bk_li.*,import/axm_li.*,import/adb_li.*,import/lr_li.*,import/vnt_li.*,import/ggl_li.*,import/seg_li.*,import/email_li.*" \
--relationships:VISITED_ON "import/rels-header.csv,import/device_li.*,import/os_li.*,import/browser_li.*" \
--relationships:VISITED_FROM "import/rels-header.csv,import/city_li.*,import/cntry_li.*,import/ip_li.*" \
--relationships:IDENTIFIED_AS "import/rels-header.csv,import/gender_li.*,import/age_li.*" \
--ignore-duplicate-nodes=true \
--ignore-missing-nodes=true \
--delimiter="~" \
--max-memory=95%
Please provide the following information if you ran into a more serious issue:
- neo4j version: Community 3.4.9
- neo4j.log and debug.log
There is insufficient memory for the Java Runtime Environment to continue.
Native memory allocation (mmap) failed to map 224919552 bytes for committing reserved memory. Possible reasons:
The system is out of physical RAM or swap space
In 32 bit mode, the process size limit was hit
Possible solutions:
Reduce memory load on the system
Increase physical memory or swap space
Check if swap backing store is full
Use 64 bit Java on a 64 bit OS
Decrease Java heap size (-Xmx/-Xms)
Decrease number of Java threads
Decrease Java thread stack sizes (-Xss)
Set larger code cache with -XX:ReservedCodeCacheSize=
This output file may be truncated or incomplete.
Out of Memory Error (os_linux.cpp:2657), pid=7928, tid=0x00007fbd2e13c700
JRE version: OpenJDK Runtime Environment (8.0_181-b13) (build 1.8.0_181-b13)
Java VM: OpenJDK 64-Bit Server VM (25.181-b13 mixed mode linux-amd64 compressed oops)
Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again