The IMPORT
statement imports the following types of data into CockroachDB:
This page has reference information about the IMPORT
statement. For instructions and working examples showing how to migrate data from other databases and formats, see the Migration Overview.
IMPORT
only works for creating new tables. It does not support adding data to existing tables. Also, IMPORT
cannot be used within a transaction.
Required privileges
Only members of the admin
role can run IMPORT
. By default, the root
user belongs to the admin
role.
Synopsis
Import a table from CSV
Import a database or table from dump file
Parameters
For import from CSV
Parameter | Description |
---|---|
table_name |
The name of the table you want to import/create. |
table_elem_list |
The table schema you want to use. |
CREATE USING file_location |
If not specifying the table schema inline via table_elem_list , this is the URL of a CSV file containing the table schema. |
file_location |
The URL of a CSV file containing the table data. This can be a comma-separated list of URLs to CSV files. For an example, see Import a table from multiple CSV files below. |
WITH kv_option_list |
Control your import's behavior with these options. |
For import from dump file
Parameter | Description |
---|---|
table_name |
The name of the table you want to import/create. Use this when the dump file contains a specific table. Leave out TABLE table_name FROM when the dump file contains an entire database. |
import_format |
PGDUMP or MYSQLDUMP |
file_location |
The URL of a dump file you want to import. |
WITH kv_option_list |
Control your import's behavior with these options. |
Import file URLs
URLs for the files you want to import must use the format shown below. For examples, see Example file URLs.
[scheme]://[host]/[path]?[parameters]
Location | Scheme | Host | Parameters |
---|---|---|---|
Amazon S3 | s3 |
Bucket name | AWS_ACCESS_KEY_ID , AWS_SECRET_ACCESS_KEY , AWS_SESSION_TOKEN |
Azure | azure |
N/A (see Example file URLs | AZURE_ACCOUNT_KEY , AZURE_ACCOUNT_NAME |
Google Cloud 1 | gs |
Bucket name | AUTH (optional; can be default , implicit , or specified ), CREDENTIALS |
HTTP 2 | http |
Remote host | N/A |
NFS/Local 3 | nodelocal |
Empty or nodeID 4 (see Example file URLs) |
N/A |
S3-compatible services 5 | s3 |
Bucket name | AWS_ACCESS_KEY_ID , AWS_SECRET_ACCESS_KEY , AWS_SESSION_TOKEN , AWS_REGION 6 (optional), AWS_ENDPOINT |
If you write to nodelocal
storage in a multi-node cluster, individual data files will be written to the extern
directories of arbitrary nodes and will likely not work as intended. To work correctly, each node must have the --external-io-dir
flag point to the same NFS mount or other network-backed, shared storage.
The location parameters often contain special characters that need to be URI-encoded. Use Javascript's encodeURIComponent function or Go language's url.QueryEscape function to URI-encode the parameters. Other languages provide similar functions to URI-encode special characters.
If your environment requires an HTTP or HTTPS proxy server for outgoing connections, you can set the standard HTTP_PROXY
and HTTPS_PROXY
environment variables when starting CockroachDB.
1If the
AUTH
parameter is not specified, thecloudstorage.gs.default.key
cluster setting will be used if it is non-empty, otherwise theimplicit
behavior is used. If theAUTH
parameter isimplicit
, all GCS connections use Google's default authentication strategy. If theAUTH
parameter isdefault
, thecloudstorage.gs.default.key
cluster setting must be set to the contents of a service account file which will be used during authentication. New in v19.1: If theAUTH
parameter isspecified
, GCS connections are authenticated on a per-statement basis, which allows the JSON key object to be sent in theCREDENTIALS
parameter. The JSON key object should be base64-encoded (using the standard encoding in RFC 4648).2 You can create your own HTTP server with Caddy or nginx. A custom root CA can be appended to the system's default CAs by setting the
cloudstorage.http.custom_ca
cluster setting, which will be used when verifying certificates from HTTPS URLs.3 The file system backup location on the NFS drive is relative to the path specified by the
--external-io-dir
flag set while starting the node. If the flag is set todisabled
, then imports from local directories and NFS drives are disabled.4 New in v19.1: The host component of NFS/Local can either be empty or the
nodeID
. If thenodeID
is specified, it is currently ignored (i.e., any node can be sent work and it will look in its local input/output directory); however, thenodeID
will likely be required in the future.5 A custom root CA can be appended to the system's default CAs by setting the
cloudstorage.http.custom_ca
cluster setting, which will be used when verifying certificates from an S3-compatible service.6 The
AWS_REGION
parameter is optional since it is not a required parameter for most S3-compatible services. Specify the parameter only if your S3-compatible service requires it.
Example file URLs
Location | Example |
---|---|
Amazon S3 | s3://acme-co/employees.sql?AWS_ACCESS_KEY_ID=123&AWS_SECRET_ACCESS_KEY=456 |
Azure | azure://employees.sql?AZURE_ACCOUNT_KEY=123&AZURE_ACCOUNT_NAME=acme-co |
Google Cloud | gs://acme-co/employees.sql |
HTTP | http://localhost:8080/employees.sql |
NFS/Local | nodelocal:///path/employees , nodelocal://2/path/employees Note: If you write to nodelocal storage in a multi-node cluster, individual data files will be written to the extern directories of arbitrary nodes and will likely not work as intended. To work correctly, each node must have the --external-io-dir flag point to the same NFS mount or other network-backed, shared storage. |
Import options
You can control the IMPORT
process's behavior using any of the following key-value pairs as a kv_option
.
Key | Context | Value | Required? | Example |
---|---|---|---|---|
delimiter |
CSV | The unicode character that delimits columns in your rows. Default: , . |
No | To use tab-delimited values: IMPORT TABLE foo (..) CSV DATA ('file.csv') WITH delimiter = e'\t' |
comment |
CSV | The unicode character that identifies rows to skip. | No | IMPORT TABLE foo (..) CSV DATA ('file.csv') WITH comment = '#' |
nullif |
CSV | The string that should be converted to NULL. | No | To use empty columns as NULL: IMPORT TABLE foo (..) CSV DATA ('file.csv') WITH nullif = '' |
skip |
CSV | The number of rows to be skipped while importing a file. Default: '0' . |
No | To import CSV files with column headers: IMPORT ... CSV DATA ('file.csv') WITH skip = '1' |
decompress |
General | The decompression codec to be used: gzip , bzip , auto , or none . Default: 'auto' , which guesses based on file extension (.gz , .bz , .bz2 ). none disables decompression. |
No | IMPORT ... WITH decompress = 'bzip' |
skip_foreign_keys |
Postgres, MySQL | Ignore foreign key constraints in the dump file's DDL. Off by default. May be necessary to import a table with unsatisfied foreign key constraints from a full database dump. | No | IMPORT TABLE foo FROM MYSQLDUMP 'dump.sql' WITH skip_foreign_keys |
max_row_size |
Postgres | Override limit on line size. Default: 0.5MB. This setting may need to be tweaked if your Postgres dump file has extremely long lines, for example as part of a COPY statement. |
No | IMPORT PGDUMP DATA ... WITH max_row_size = '5MB' |
For examples showing how to use these options, see the Examples section below.
For instructions and working examples showing how to migrate data from other databases and formats, see the Migration Overview.
Requirements
Prerequisites
Before using IMPORT
, you should have:
- The schema of the table you want to import.
- The data you want to import, preferably hosted on cloud storage. This location must be equally accessible to all nodes using the same import file location. This is necessary because the
IMPORT
statement is issued once by the client, but is executed concurrently across all nodes of the cluster. For more information, see the Import file location section below.
Import targets
Imported tables must not exist and must be created in the IMPORT
statement. If the table you want to import already exists, you must drop it with DROP TABLE
.
You can specify the target database in the table name in the IMPORT
statement. If it's not specified there, the active database in the SQL session is used.
Create table
Your IMPORT
statement must reference a CREATE TABLE
statement representing the schema of the data you want to import. You have several options:
Specify the table's columns explicitly from the SQL client. For an example, see Import a table from a CSV file below.
Load a file that already contains a
CREATE TABLE
statement. For an example, see Import a Postgres database dump below.
We also recommend specifying all secondary indexes you want to use in the CREATE TABLE
statement. It is possible to add secondary indexes later, but it is significantly faster to specify them during import.
By default, the Postgres and MySQL import formats support foreign keys. However, the most common dependency issues during import are caused by unsatisfied foreign key relationships that cause errors like pq: there is no unique constraint matching given keys for referenced table tablename
. You can avoid these issues by adding the skip_foreign_keys
option to your IMPORT
statement as needed. Ignoring foreign constraints will also speed up data import.
Available storage
Each node in the cluster is assigned an equal part of the imported data, and so must have enough temp space to store it. In addition, data is persisted as a normal table, and so there must also be enough space to hold the final, replicated data. The node's first-listed/default store
directory must have enough available storage to hold its portion of the data.
On cockroach start
, if you set --max-disk-temp-storage
, it must also be greater than the portion of the data a node will store in temp space.
Import file location
We strongly recommend using cloud/remote storage (Amazon S3, Google Cloud Platform, etc.) for the data you want to import.
Local files are supported; however, they must be accessible to all nodes in the cluster using identical Import file URLs.
To import a local file, you have the following options:
Option 1. Run a local file server to make the file accessible from all nodes.
Option 2. Make the file accessible from each local node's store:
- Create an
extern
directory on each node's store. The pathname will differ depending on the--store
flag passed tocockroach start
(if any), but will look something like/path/to/cockroach-data/extern/
. - Copy the file to each node's
extern
directory. - Assuming the file is called
data.sql
, you can access it in yourIMPORT
statement using the following import file URL:'nodelocal:///data.sql'
.
- Create an
Table users and privileges
Imported tables are treated as new tables, so you must GRANT
privileges to them.
Performance
All nodes are used during the import job, which means all nodes' CPU and RAM will be partially consumed by the IMPORT
task in addition to serving normal traffic.
Viewing and controlling import jobs
After CockroachDB successfully initiates an import, it registers the import as a job, which you can view with SHOW JOBS
.
After the import has been initiated, you can control it with PAUSE JOB
, RESUME JOB
, and CANCEL JOB
.
If initiated correctly, the statement returns when the import is finished or if it encounters an error. In some cases, the import can continue after an error has been returned (the error message will tell you that the import has resumed in background).
IMPORT
job will cause it to restart from the beginning.Examples
Import a table from a CSV file
To manually specify the table schema:
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/customers.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]&AWS_SESSION_TOKEN=[placeholder]')
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv')
;
To use a file to specify the table schema:
Amazon S3:
> IMPORT TABLE customers
CREATE USING 's3://acme-co/customers-create-table.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]'
CSV DATA ('s3://acme-co/customers.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
;
Azure:
> IMPORT TABLE customers
CREATE USING 'azure://acme-co/customer-create-table.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co'
CSV DATA ('azure://acme-co/customer-import-data.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
;
Google Cloud:
> IMPORT TABLE customers
CREATE USING 'gs://acme-co/customers-create-table.sql'
CSV DATA ('gs://acme-co/customers.csv')
;
Import a table from multiple CSV files
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA (
's3://acme-co/customers.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]',
's3://acme-co/customers2.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder',
's3://acme-co/customers3.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]',
's3://acme-co/customers4.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]',
);
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA (
'azure://acme-co/customer-import-data1.1.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co',
'azure://acme-co/customer-import-data1.2.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co',
'azure://acme-co/customer-import-data1.3.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co',
'azure://acme-co/customer-import-data1.4.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co',
'azure://acme-co/customer-import-data1.5.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co',
);
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA (
'gs://acme-co/customers.csv',
'gs://acme-co/customers2.csv',
'gs://acme-co/customers3.csv',
'gs://acme-co/customers4.csv',
);
Import a table from a TSV file
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/customers.tsv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
WITH
delimiter = e'\t'
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.tsv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
WITH
delimiter = e'\t'
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.tsv')
WITH
delimiter = e'\t'
;
Skip commented lines
The comment
option determines which Unicode character marks the rows in the data to be skipped.
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/customers.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
WITH
comment = '#'
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
WITH
comment = '#'
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv')
WITH
comment = '#'
;
Skip first n lines
The skip
option determines the number of header rows to skip when importing a file.
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/customers.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
WITH
skip = '2'
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
WITH
skip = '2'
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv')
WITH
skip = '2'
;
Use blank characters as NULL
The nullif
option defines which string should be converted to NULL
.
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/employees.csv?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
WITH
nullif = ''
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
WITH
nullif = ''
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv')
WITH
nullif = ''
;
Import a compressed CSV file
CockroachDB chooses the decompression codec based on the filename (the common extensions .gz
or .bz2
and .bz
) and uses the codec to decompress the file during import.
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/employees.csv.gz?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv.gz?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv.gz')
;
Optionally, you can use the decompress
option to specify the codec to be used for decompressing the file during import:
Amazon S3:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('s3://acme-co/employees.csv.gz?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]')
WITH
decompress = 'gzip'
;
Azure:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('azure://acme-co/customer-import-data.csv.gz.latest?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co')
WITH
decompress = 'gzip'
;
Google Cloud:
> IMPORT TABLE customers (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
name TEXT,
INDEX name_idx (name)
)
CSV DATA ('gs://acme-co/customers.csv.gz')
WITH
decompress = 'gzip'
;
Import a Postgres database dump
Amazon S3:
> IMPORT PGDUMP 's3://your-external-storage/employees.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]';
Azure:
> IMPORT PGDUMP 'azure://acme-co/employees.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co';
Google Cloud:
> IMPORT PGDUMP 'gs://acme-co/employees.sql';
For the commands above to succeed, you need to have created the dump file with specific flags to pg_dump
. For more information, see Migrate from Postgres.
Import a table from a Postgres database dump
Amazon S3:
> IMPORT TABLE employees FROM PGDUMP 's3://your-external-storage/employees-full.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]' WITH skip_foreign_keys;
Azure:
> IMPORT TABLE employees FROM PGDUMP 'azure://acme-co/employees.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co' WITH skip_foreign_keys;
Google Cloud:
> IMPORT TABLE employees FROM PGDUMP 'gs://acme-co/employees.sql' WITH skip_foreign_keys;
If the table schema specifies foreign keys into tables that do not exist yet, the WITH skip_foreign_keys
shown may be needed. For more information, see the list of import options.
For the command above to succeed, you need to have created the dump file with specific flags to pg_dump
. For more information, see Migrate from Postgres.
Import a CockroachDB dump file
Cockroach dump files can be imported using the IMPORT PGDUMP
.
Amazon S3:
> IMPORT PGDUMP 's3://your-external-storage/employees-full.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]';
Azure:
> IMPORT PGDUMP 'azure://acme-co/employees.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co';
Google Cloud:
> IMPORT PGDUMP 'gs://acme-co/employees.sql';
For more information, see SQL Dump (Export).
Import a MySQL database dump
Amazon S3:
> IMPORT MYSQLDUMP 's3://your-external-storage/employees-full.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]';
Azure:
> IMPORT MYSQLDUMP 'azure://acme-co/employees.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co';
Google Cloud:
> IMPORT MYSQLDUMP 'gs://acme-co/employees.sql';
For more detailed information about importing data from MySQL, see Migrate from MySQL.
Import a table from a MySQL database dump
Amazon S3:
> IMPORT TABLE employees FROM MYSQLDUMP 's3://your-external-storage/employees-full.sql?AWS_ACCESS_KEY_ID=[placeholder]&AWS_SECRET_ACCESS_KEY=[placeholder]' WITH skip_foreign_keys
Azure:
> IMPORT TABLE employees FROM MYSQLDUMP 'azure://acme-co/employees.sql?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co' WITH skip_foreign_keys
Google Cloud:
> IMPORT TABLE employees FROM MYSQLDUMP 'gs://acme-co/employees.sql' WITH skip_foreign_keys
If the table schema specifies foreign keys into tables that do not exist yet, the WITH skip_foreign_keys
shown may be needed. For more information, see the list of import options.
For more detailed information about importing data from MySQL, see Migrate from MySQL.
Known limitation
IMPORT
can sometimes fail with a "context canceled" error, or can restart itself many times without ever finishing. If this is happening, it is likely due to a high amount of disk contention. This can be mitigated by setting the kv.bulk_io_write.max_rate
cluster setting to a value below your max disk write speed. For example, to set it to 10MB/s, execute:
> SET CLUSTER SETTING kv.bulk_io_write.max_rate = '10MB';