DuckDB implementation of the LDBC Social Network Benchmark's Interactive workload.
Grab DuckDB:
scripts/get.sh
The data sets need to be generated before loading it to the database. No preprocessing is required. To generate the data sets for DuckDB, use the same settings as for PostgreSQL, i.e. the Hadoop-based Datagen's CsvMergeForeign
serializer classes.
Set the following environment variable based on your data source:
export DUCKDB_CSV_DIR=`pwd`/../postgres/test-data
Load the data set as follows:
scripts/load.sh
The instructions below explain how to run the benchmark driver in one of the three modes (create validation parameters, validate, benchmark). For more details on the driver modes, check the "Driver modes" section of the main README.
-
Edit the
driver/benchmark.properties
file. Make sure that theldbc.snb.interactive.scale_factor
,ldbc.snb.interactive.updates_dir
,ldbc.snb.interactive.parameters_dir
properties are set correctly and are in sync. -
Run the script:
driver/create-validation-parameters.sh
-
Edit the
driver/validate.properties
file. Make sure that thevalidate_database
property points to the file you would like to validate against. -
Run the script:
driver/validate.sh
-
Edit the
driver/benchmark.properties
file. Make sure that theldbc.snb.interactive.scale_factor
,ldbc.snb.interactive.updates_dir
, andldbc.snb.interactive.parameters_dir
properties are set correctly and are in sync. -
Run the script:
driver/benchmark.sh
scripts/backup-database.sh
and scripts/restore-database.sh
scripts to achieve this.