docs-v2/content/influxdb3/cloud-serverless/guides/migrate-data/migrate-1x-to-serverless.md

14 KiB

title description menu weight aliases alt_links
Migrate data from InfluxDB 1.x to InfluxDB Cloud Serverless To migrate data from a TSM-powered InfluxDB 1.x (OSS or Enterprise) to an InfluxDB Cloud Serverless organization, export the data as line protocol and write the exported data to an bucket in your InfluxDB Cloud Serverless organization.
influxdb_cloud_serverless
name parent
Migrate from 1.x to Serverless Migrate data
103
/influxdb3/cloud-serverless/write-data/migrate-data/migrate-1x-to-iox/
/influxdb3/cloud-serverless/guides/migrate-data/migrate-1x-to-iox/
cloud-dedicated clustered
/influxdb3/cloud-dedicated/guides/migrate-data/migrate-1x-to-cloud-dedicated/ /influxdb3/clustered/guides/migrate-data/migrate-1x-to-clustered/

To migrate data from an InfluxDB 1.x OSS or Enterprise instance to InfluxDB Cloud Serverless powered by the v3 storage engine, export the data as line protocol and write the exported data to a bucket in your InfluxDB Cloud Serverless organization. Because full data migrations will likely exceed your organizations' limits and adjustable quotas, migrate your data in batches.

[!Important] All write requests are subject to your InfluxDB Cloud Serverless organization's rate limits and adjustable quotas.

Before you migrate

Before you migrate from InfluxDB 1.x to {{< product-name >}}, there are schema design practices supported by the TSM storage engine that are not supported in the InfluxDB 3 storage engine. Specifically, InfluxDB 3 enforces the following schema restrictions:

  • You can't use duplicate names for tags and fields
  • Measurements can contain up to 200 columns where each column represents time, a field, or a tag.

For more information, see Schema restrictions.

If your schema does not adhere to these restrictions, you must update your schema before migrating to {{< product-name >}}.

{{< expand-wrapper >}} {{% expand "Fix duplicate tag and field names" %}}

If your current schema in InfluxDB 1.x includes tags and fields with the same name, rename either the duplicate tag key or field key. The simplest way to do this is to directly modify the line protocol exported in step 1 of Migrate data below before writing it to {{< product-name >}}.

For example, the following line protocol includes both a tag and field named temp.

home,room=Kitchen,temp=F co=0i,hum=56.6,temp=71.0 1672531200000000000

To be able to write this line protocol to {{< product-name >}}, update the temp tag key to tempScale:

home,room=Kitchen,tempScale=F co=0i,hum=56.6,temp=71.0 1672531200000000000

{{% /expand %}} {{% expand "Fix measurements with more than 200 total columns" %}}

If in your current schema, the total number of tags, fields, and time columns in a single measurement exceeds 200, you need to update your schema before migrating to {{< product-name >}}. Because tags are metadata used to identify specific series, we recommend splitting groups of fields across multiple measurements.

In your InfluxDB 1.x instance:

  1. Create a new database to store the modified data in.
  2. Identify what fields could be grouped together in a measurement and not exceed the 200 column limit.
  3. Query each group of fields from the existing database and write them into a unique measurement in the new database:

{{< flex >}} {{% flex-content "half" %}}

Write one set of fields to a new measurement
SELECT 
  field1,
  field2,
  field3,
  field4
INTO
  "new-database".."new-measurement-1"
FROM
  "example-measurement"
GROUP BY *

{{% /flex-content %}} {{% flex-content "half" %}}

Write another set of fields to new measurement
SELECT 
  field5,
  field6,
  field7,
  field8
INTO
  "new-database".."new-measurement-2"
FROM
  "example-measurement"
GROUP BY *

{{% /flex-content %}} {{< /flex >}}

When exporting your data as line protocol in step 1 of Migrate data below, only export the newly created database that contains measurements with separate groups of fields.

If duplicating data is not feasible

If duplicating data into a new database isn't feasible, you can directly modify the exported line protocol to group certain fields into unique measurements. For example:

example-measurement field1=0,field2=0,field3=0,field4=0,field5=0,field6=0,field7=0,field8=0 1672531200000000000

Would become:

new-measurement-1 field1=0,field2=0,field3=0,field4=0 1672531200000000000
new-measurement-2 field5=0,field6=0,field7=0,field8=0 1672531200000000000

{{% /expand %}} {{< /expand-wrapper >}}

Tools to use

The migration process uses the following tools:

{{% note %}}

InfluxDB 1.x and 2.x CLIs are unique

If both the InfluxDB 1.x and 2.x influx CLIs are installed in your $PATH, rename one of the the binaries to ensure you're executing commands with the correct CLI. {{% /note %}}

Migrate data

  1. Export data from your InfluxDB 1.x instance as line protocol.

    Use the InfluxDB 1.x influx_inspect export utility to export data as line protocol and store it in a file. Include the following:

    • ({{< req "Required" >}}) -lponly flag to export line protocol without InfluxQL DDL or DML.
    • ({{< req "Required" >}}) -out flag with a path to an output file. Default is ~/.influxdb/export. Any subsequent export commands without the output file defined will overwrite the existing export file.
    • -compress flag to use gzip to compress the output.
    • -datadir flag with the path to your InfluxDB 1.x data directory. Only required if the data directory is at a non-default location. For information about default locations, see InfluxDB OSS 1.x file system layout or InfluxDB Enterprise 1.x file system layout.
    • -waldir flag with the path to your InfluxDB 1.x wal directory. Only required if the wal directory is at a non-default location. For information about default locations, see InfluxDB OSS 1.x file system layout or InfluxDB Enterprise 1.x file system layout.
    • -database flag with a specific database name to export. By default, all databases are exported.
    • -retention flag with a specific retention policy to export. By default, all retention policies are exported.
    • -start flag with an RFC3339 timestamp that defines the earliest time to export. Default is 1677-09-20T16:27:54-07:44.
    • -end flag with an RFC3339 timestamp that defines the latest time to export. Default is 2262-04-11T16:47:16-07:00.

    {{% note %}} We recommend exporting each database and retention policy combination separately to easily write the exported data into corresponding InfluxDB {{< current-version >}} buckets. {{% /note %}}

    Export all data in a database and retention policy to a file
    influx_inspect export \
      -lponly \
      -database example-db \
      -retention example-rp \
      -out path/to/export-file.lp
    
    View more export command examples:

    {{< expand-wrapper >}} {{% expand "Export all data to a file" %}}

influx_inspect export \
  -lponly \
  -out path/to/export-file.lp.gzip

{{% /expand %}}

{{% expand "Export all data to a compressed file" %}}

influx_inspect export \
  -lponly \
  -compress \
  -out path/to/export-file.lp.gzip

{{% /expand %}}

{{% expand "Export data within time bounds to a file" %}}

influx_inspect export \
  -lponly \
  -start 2020-01-01T00:00:00Z \
  -end 2023-01-01T00:00:00Z \
  -out path/to/export-file.lp

{{% /expand %}}

{{% expand "Export a database and all its retention policies to a file" %}}

influx_inspect export \
  -lponly \
  -database example-db \
  -out path/to/export-file.lp

{{% /expand %}}

{{% expand "Export a specific database and retention policy to a file" %}}

influx_inspect export \
  -lponly \
  -database example-db \
  -retention example-rp \
  -out path/to/export-file.lp

{{% /expand %}}

{{% expand "Export all data from non-default data and wal directories" %}}

influx_inspect export \
  -lponly \
  -datadir path/to/influxdb/data/ \
  -waldir path/to/influxdb/wal/ \
  -out path/to/export-file.lp

{{% /expand %}} {{< /expand-wrapper >}}

  1. Create InfluxDB Cloud Serverless buckets for each InfluxDB 1.x database and retention policy combination. InfluxDB {{< current-version >}} combines InfluxDB 1.x databases and retention policies into buckets--named locations for time series data with specified retention periods.

    {{< expand-wrapper >}} {{% expand "View example 1.x databases and retention policies as InfluxDB Cloud buckets" %}} If you have the following InfluxDB 1.x data structure:

  • example-db (database)
    • autogen (retention policy)
    • historical-1mo (retention policy)
    • historical-6mo (retention policy)
    • historical-1y (retention policy)

You would create the following InfluxDB {{< current-version >}} buckets:

  • example-db/autogen
  • example-db/historical-1mo
  • example-db/historical-6mo
  • example-db/historical-1y

{{% /expand %}} {{< /expand-wrapper >}}

Use the **InfluxDB 2.x `influx` CLI** or the **InfluxDB {{< current-version >}} user interface (UI)**
to create a bucket.

{{< tabs-wrapper >}}

{{% tabs %}} influx CLI InfluxDB UI {{% /tabs %}}

{{% tab-content %}}

Use the influx bucket create command to create a new bucket.

Provide the following:

influx bucket create \
  --name example-db/autogen \
  --retention 7d

{{% /tab-content %}} {{% tab-content %}}

  1. Go to cloud2.influxdata.com in a browser to log in and access the InfluxDB UI.

  2. Navigate to Load Data > Buckets using the left navigation bar.

{{< nav-icon "load data" >}}

  1. Click + {{< caps >}}Create bucket{{< /caps >}}.

  2. Provide a bucket name (for example: example-db/autogen) and select a retention period. Supported retention periods depend on your InfluxDB Cloud Serverless plan.

  3. Click {{< caps >}}Create{{< /caps >}}.

{{% /tab-content %}} {{< /tabs-wrapper >}}

  1. Write the exported line protocol to your InfluxDB Cloud Serverless organization.

    Use the InfluxDB 2.x CLI to write data to InfluxDB Cloud Serverless. While you can use the /api/v2/write API endpoint to write data directly, the influx write command lets you define the rate at which data is written to avoid exceeding your organization's rate limits.

    Use the influx write command and include the following:

    • InfluxDB Cloud Serverless connection and authentication credentials
    • -b, --bucket flag to identify the target bucket.
    • -f, --file flag with the path to the line protocol file to import.
    • -rate-limit flag with a rate limit that matches your InfluxDB Cloud organization's write rate limit.
    • --compression flag to identify the compression type of the import file. Options are none or gzip. Default is none.

    {{< cli/influx-creds-note >}}

    {{< code-tabs-wrapper >}} {{% code-tabs %}} Uncompressed Compressed {{% /code-tabs %}} {{% code-tab-content %}}

influx write \
  --bucket example-db/autogen \
  --file path/to/export-file.lp \
  --rate-limit "300 MB / 5 min"

{{% /code-tab-content %}} {{% code-tab-content %}}

influx write \
  --bucket example-db/autogen \
  --file path/to/export-file.lp.gzip \
  --rate-limit "300 MB / 5 min" \
  --compression gzip

{{% /code-tab-content %}} {{< /code-tabs-wrapper >}}

Repeat for each export file and target bucket.