You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: modules/ROOT/pages/import.adoc
+19-13Lines changed: 19 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,7 +12,16 @@ You should use this tool when:
12
12
13
13
* Import performance is important because you have a large amount of data (millions/billions of entities).
14
14
* The database can be taken offline and you have direct access to one of the servers hosting your Neo4j DBMS.
15
-
* The database is either empty.
15
+
* The database is empty and you have to perform the initial data load.
16
+
* You need to update your graph with large amount of data.
17
+
In this case, importing data incrementally can be more performant that transactional insertion.
18
+
+
19
+
[NOTE]
20
+
====
21
+
The incremental import can be done either within a single command or in stages.
22
+
For details, see <<_incremental_import_in_a_single_command>> and <<incremental-import-stages>>.
23
+
====
24
+
+
16
25
* The CSV data is clean/fault-free (nodes are not duplicated and relationships' start and end nodes exist).
17
26
This tool can handle data faults but performance is not optimized.
18
27
If your data has a lot of faults, it is recommended to clean it using a dedicated tool before import.
@@ -686,20 +695,17 @@ Incremental import into an existing database.
686
695
687
696
=== Usage and limitations
688
697
689
-
[WARNING]
690
-
====
691
-
To safely use the importer in a clustered environment please use the following process.
698
+
The importer works well on standalone servers.
692
699
693
-
Use the importer to import the database into one server in the cluster.
694
-
This server can then be the source for other cluster members to copy the database from using the xref:clustering/databases.adoc#cluster-designated-seeder[designated seeder] approach.
695
-
696
-
To perform an incremental import in a clustered environment first alter the database topology to 1 primary using xref:procedures.adoc#procedure_dbms_recreateDatabase[`dbms.recreateDatabase()`].
697
-
Then stop the database using xref::database-administration/standard-databases/start-stop-databases.adoc#manage-databases-stop[STOP DATABASE].
698
-
Perform the incremental import against the server hosting the database.
699
-
Then start the database using xref::database-administration/standard-databases/start-stop-databases.adoc#manage-databases-start[START DATABASE].
700
-
Lastly, alter the database back to the desired topology using xref::database-administration/standard-databases/alter-databases.adoc#[ALTER DATABASE].
701
-
====
700
+
To safely perform an incremental import in a clustered environment, follow these steps:
702
701
702
+
. Run the incremental import command on a single server in the cluster.
703
+
This server can then be used as the xref:clustering/databases.adoc#cluster-designated-seeder[designated seeder] from which other cluster members can copy the database.
704
+
. Reconfigure the database topology to a single primary by running the xref:procedures.adoc#procedure_dbms_recreateDatabase[`dbms.recreateDatabase()`] procedure.
705
+
. Then stop the database using xref::database-administration/standard-databases/start-stop-databases.adoc#manage-databases-stop[STOP DATABASE].
706
+
. Perform the incremental import on the server that hosts the database.
707
+
. Then start the database with xref::database-administration/standard-databases/start-stop-databases.adoc#manage-databases-start[START DATABASE].
708
+
. Lastly, restore the desired database topology using xref::database-administration/standard-databases/alter-databases.adoc#[ALTER DATABASE].
703
709
The incremental import command can be used to add:
0 commit comments