diff --git a/api-docs/cloud/ref.yml b/api-docs/cloud/ref.yml
index 3fac6d59b..00d799ea3 100644
--- a/api-docs/cloud/ref.yml
+++ b/api-docs/cloud/ref.yml
@@ -2395,11 +2395,6 @@ components:
- invalid
readOnly: true
type: string
- maxLength:
- description: Max length in bytes for a body of line-protocol.
- format: int32
- readOnly: true
- type: integer
message:
description: Message is a human-readable message.
readOnly: true
@@ -2407,7 +2402,6 @@ components:
required:
- code
- message
- - maxLength
Link:
description: URI of resource.
format: uri
@@ -9989,7 +9983,7 @@ paths:
description: Unexpected error
summary: List scripts
tags:
- - Invokable Scripts
+ - Invocable Scripts
post:
operationId: PostScripts
requestBody:
@@ -10011,7 +10005,7 @@ paths:
description: Unexpected error
summary: Create a script
tags:
- - Invokable Scripts
+ - Invocable Scripts
/api/v2/scripts/{scriptID}:
delete:
description: Deletes a script and all associated records.
@@ -10031,9 +10025,9 @@ paths:
description: Unexpected error
summary: Delete a script
tags:
- - Invokable Scripts
+ - Invocable Scripts
get:
- description: Uses script ID to retrieve details of an invokable script.
+ description: Uses script ID to retrieve details of an invocable script.
operationId: GetScriptsID
parameters:
- description: The script ID.
@@ -10054,10 +10048,10 @@ paths:
description: Unexpected error
summary: Retrieve a script
tags:
- - Invokable Scripts
+ - Invocable Scripts
patch:
description: >
- Updates properties (`name`, `description`, and `script`) of an invokable
+ Updates properties (`name`, `description`, and `script`) of an invocable
script.
operationId: PatchScriptsID
parameters:
@@ -10086,7 +10080,7 @@ paths:
description: Unexpected error
summary: Update a script
tags:
- - Invokable Scripts
+ - Invocable Scripts
/api/v2/scripts/{scriptID}/invoke:
post:
description: >-
@@ -10116,7 +10110,7 @@ paths:
description: Unexpected error
summary: Invoke a script
tags:
- - Invokable Scripts
+ - Invocable Scripts
/api/v2/setup:
get:
description: >-
@@ -12127,6 +12121,10 @@ paths:
format.
+ InfluxDB Cloud enforces rate and size limits different from InfluxDB
+ OSS. For details, see Responses.
+
+
For more information and examples, see the following:
- [Write data with the InfluxDB
@@ -12289,10 +12287,19 @@ paths:
`bucket`, and name.
'413':
content:
+ application/json:
+ examples:
+ dataExceedsSizeLimitOSS:
+ summary: InfluxDB OSS response
+ value: >
+ {"code":"request too large","message":"unable to read data:
+ points batch is too large"}
+ schema:
+ $ref: '#/components/schemas/LineProtocolLengthError'
text/html:
examples:
dataExceedsSizeLimit:
- summary: Cloud response
+ summary: InfluxDB Cloud response
value: |
413 Request Entity Too Large
@@ -12304,13 +12311,21 @@ paths:
schema:
type: string
- description: >-
- Request entity too large. The payload exceeded the 50MB size limit.
- InfluxDB rejected the batch and did not write any data.
+ description: >
+ The request payload is too large. InfluxDB rejected the batch and
+ did not write any data.
+
+ #### InfluxDB Cloud:
+ - returns this error if the payload exceeds the 50MB size limit.
+ - returns `Content-Type: text/html` for this error.
+
+ #### InfluxDB OSS:
+ - returns this error only if the [Go (golang) `ioutil.ReadAll()`](https://pkg.go.dev/io/ioutil#ReadAll) function raises an error.
+ - returns `Content-Type: application/json` for this error.
'429':
description: >-
- The token is temporarily over quota. The Retry-After header
- describes when to try the write again.
+ InfluxDB Cloud only. The token is temporarily over quota. The
+ Retry-After header describes when to try the write again.
headers:
Retry-After:
description: >-
@@ -12387,7 +12402,7 @@ tags:
- DBRPs
- Delete
- DemoDataBuckets
- - Invokable Scripts
+ - Invocable Scripts
- Labels
- Limits
- NotificationEndpoints
@@ -12519,7 +12534,7 @@ x-tagGroups:
- DBRPs
- Delete
- DemoDataBuckets
- - Invokable Scripts
+ - Invocable Scripts
- Labels
- Limits
- NotificationEndpoints
diff --git a/api-docs/v2.1/ref.yml b/api-docs/v2.1/ref.yml
index 60ddded7c..fd6309f73 100644
--- a/api-docs/v2.1/ref.yml
+++ b/api-docs/v2.1/ref.yml
@@ -2294,11 +2294,6 @@ components:
- invalid
readOnly: true
type: string
- maxLength:
- description: Max length in bytes for a body of line-protocol.
- format: int32
- readOnly: true
- type: integer
message:
description: Message is a human-readable message.
readOnly: true
@@ -2306,7 +2301,6 @@ components:
required:
- code
- message
- - maxLength
Link:
description: URI of resource.
format: uri
@@ -12717,7 +12711,8 @@ paths:
To write data into InfluxDB, you need the following:
- - **organization** – _See [View
+
+ - **organization name or ID** – _See [View
organizations](https://docs.influxdata.com/influxdb/v2.1/organizations/view-orgs/#view-your-organization-id)
for instructions on viewing your organization ID._
@@ -12736,6 +12731,10 @@ paths:
format.
+ InfluxDB Cloud enforces rate and size limits different from InfluxDB
+ OSS. For details, see Responses.
+
+
For more information and examples, see the following:
- [Write data with the InfluxDB
@@ -12757,8 +12756,8 @@ paths:
schema:
default: identity
description: >-
- The header value specifies that the line protocol in the request
- body is encoded with gzip or not encoded with identity.
+ The content coding. Use `gzip` for compressed data or `identity`
+ for unmodified, uncompressed data.
enum:
- gzip
- identity
@@ -12848,9 +12847,7 @@ paths:
application/json:
examples:
measurementSchemaFieldTypeConflict:
- summary: >-
- Example of a field type conflict thrown by an explicit
- bucket schema
+ summary: Field type conflict thrown by an explicit bucket schema
value:
code: invalid
message: >-
@@ -12901,13 +12898,52 @@ paths:
'413':
content:
application/json:
+ examples:
+ dataExceedsSizeLimitOSS:
+ summary: InfluxDB OSS response
+ value: >
+ {"code":"request too large","message":"unable to read data:
+ points batch is too large"}
schema:
$ref: '#/components/schemas/LineProtocolLengthError'
+ text/html:
+ examples:
+ dataExceedsSizeLimit:
+ summary: InfluxDB Cloud response
+ value: |
+
+ 413 Request Entity Too Large
+
+ 413 Request Entity Too Large
+
+ nginx
+
+
+ schema:
+ type: string
description: >
- All request data was rejected and not written. InfluxDB OSS only
- returns this error if the [Go (golang)
- `ioutil.ReadAll()`](https://pkg.go.dev/io/ioutil#ReadAll) function
- raises an error.
+ The request payload is too large. InfluxDB rejected the batch and
+ did not write any data.
+
+ #### InfluxDB Cloud:
+ - returns this error if the payload exceeds the 50MB size limit.
+ - returns `Content-Type: text/html` for this error.
+
+ #### InfluxDB OSS:
+ - returns this error only if the [Go (golang) `ioutil.ReadAll()`](https://pkg.go.dev/io/ioutil#ReadAll) function raises an error.
+ - returns `Content-Type: application/json` for this error.
+ '429':
+ description: >-
+ InfluxDB Cloud only. The token is temporarily over quota. The
+ Retry-After header describes when to try the write again.
+ headers:
+ Retry-After:
+ description: >-
+ A non-negative decimal integer indicating the seconds to delay
+ after the response is received.
+ schema:
+ format: int32
+ type: integer
'500':
content:
application/json:
diff --git a/api-docs/v2.1/swaggerV1Compat.yml b/api-docs/v2.1/swaggerV1Compat.yml
index 583c4b0d8..d0ab5fd9d 100644
--- a/api-docs/v2.1/swaggerV1Compat.yml
+++ b/api-docs/v2.1/swaggerV1Compat.yml
@@ -29,7 +29,7 @@ paths:
schema:
type: string
required: true
- description: The bucket to write to. If the specified bucket does not exist, a bucket is created with a default 3 day retention policy.
+ description: The bucket to write to. If none exist a bucket will be created with a default 3 day retention policy.
- in: query
name: rp
schema:
@@ -54,7 +54,7 @@ paths:
"204":
description: Write data is correctly formatted and accepted for writing to the bucket.
"400":
- description: Line protocol was not in correct format, and no points were written. Response can be used to determine the first malformed line in the line-protocol body. All data in body was rejected and not written.
+ description: Line protocol poorly formed and no points were written. Response can be used to determine the first malformed line in the body line-protocol. All data in body was rejected and not written.
content:
application/json:
schema:
diff --git a/content/chronograf/v1.9/guides/dashboard-template-variables.md b/content/chronograf/v1.9/guides/dashboard-template-variables.md
index 68eb43b91..64638f670 100644
--- a/content/chronograf/v1.9/guides/dashboard-template-variables.md
+++ b/content/chronograf/v1.9/guides/dashboard-template-variables.md
@@ -104,7 +104,7 @@ Use dot or bracket notation to reference the variable key inside of the `v` reco
```js
from(bucket: v.bucket)
- |> range(start: v.timeRangeStart, stop: v.timeRangeStart)
+ |> range(start: v.timeRangeStart, stop: v.timeRangeStop)
|> filter(fn: (r) => r._field == v["Field key"])
|> aggregateWindow(every: v.windowPeriod, fn: v.aggregateFunction)
```
diff --git a/content/enterprise_influxdb/v1.7/guides/fine-grained-authorization.md b/content/enterprise_influxdb/v1.7/guides/fine-grained-authorization.md
index 6f31cad66..17da646f0 100644
--- a/content/enterprise_influxdb/v1.7/guides/fine-grained-authorization.md
+++ b/content/enterprise_influxdb/v1.7/guides/fine-grained-authorization.md
@@ -15,6 +15,12 @@ Use fine-grained authorization (FGA) in InfluxDB Enterprise to control user acce
You must have [admin permissions](/influxdb/v1.7/administration/authentication_and_authorization/#admin-user-management) to set up FGA.
+{{% warn %}}
+#### FGA does not apply to Flux
+FGA does not restrict actions performed by Flux queries (both read and write).
+If using FGA, we recommend [disabling Flux](/enterprise_influxdb/v{{< current-version >}}/flux/installation/).
+{{% /warn %}}
+
## Set up fine-grained authorization
1. [Enable authentication](/influxdb/v1.7/administration/authentication_and_authorization/#set-up-authentication) in your InfluxDB configuration file.
diff --git a/content/enterprise_influxdb/v1.8/guides/fine-grained-authorization.md b/content/enterprise_influxdb/v1.8/guides/fine-grained-authorization.md
index c98fb1aa4..28b95be36 100644
--- a/content/enterprise_influxdb/v1.8/guides/fine-grained-authorization.md
+++ b/content/enterprise_influxdb/v1.8/guides/fine-grained-authorization.md
@@ -17,6 +17,12 @@ Use fine-grained authorization (FGA) in InfluxDB Enterprise to control user acce
You must have [admin permissions](/influxdb/v1.8/administration/authentication_and_authorization/#admin-user-management) to set up FGA.
+{{% warn %}}
+#### FGA does not apply to Flux
+FGA does not restrict actions performed by Flux queries (both read and write).
+If using FGA, we recommend [disabling Flux](/enterprise_influxdb/v{{< current-version >}}/flux/installation/).
+{{% /warn %}}
+
## Set up fine-grained authorization
1. [Enable authentication](/influxdb/v1.8/administration/authentication_and_authorization/#set-up-authentication) in your InfluxDB configuration file.
diff --git a/content/enterprise_influxdb/v1.9/administration/manage/security/authentication_and_authorization-api.md b/content/enterprise_influxdb/v1.9/administration/manage/security/authentication_and_authorization-api.md
index 2f9945c19..412cda691 100644
--- a/content/enterprise_influxdb/v1.9/administration/manage/security/authentication_and_authorization-api.md
+++ b/content/enterprise_influxdb/v1.9/administration/manage/security/authentication_and_authorization-api.md
@@ -10,7 +10,7 @@ weight: 42
---
- [Overview](#overview)
-- [API examples](#user-and-privilege-management-over-the-influxd-meta-api)
+- [API examples](#user-and-privilege-management-over-the-influxdb-enterprise-meta-api)
## Overview
diff --git a/content/enterprise_influxdb/v1.9/administration/monitor/logs.md b/content/enterprise_influxdb/v1.9/administration/monitor/logs.md
index 3da60a964..a6bfa05f2 100644
--- a/content/enterprise_influxdb/v1.9/administration/monitor/logs.md
+++ b/content/enterprise_influxdb/v1.9/administration/monitor/logs.md
@@ -130,6 +130,179 @@ See [Structured logging](/enterprise_influxdb/v1.9/administration/logs/#structur
## Tracing
-Logging has been enhanced, starting in InfluxDB 1.5, to provide tracing of important InfluxDB operations. Tracing is useful for error reporting and discovering performance bottlenecks.
+Logging has been enhanced to provide tracing of important InfluxDB operations.
+Tracing is useful for error reporting and discovering performance bottlenecks.
-See [Tracing](/enterprise_influxdb/v1.9/administration/logs/#tracing) in the InfluxDB OSS documentation.
+### Logging keys used in tracing
+
+#### Tracing identifier key
+
+The `trace_id` key specifies a unique identifier for a specific instance of a trace.
+You can use this key to filter and correlate all related log entries for an operation.
+
+All operation traces include consistent starting and ending log entries, with the same message (`msg`) describing the operation (e.g., "TSM compaction"), but adding the appropriate `op_event` context (either `start` or `end`).
+For an example, see [Finding all trace log entries for an InfluxDB operation](#finding-all-trace-log-entries-for-an-influxdb-operation).
+
+**Example:** `trace_id=06R0P94G000`
+
+#### Operation keys
+
+The following operation keys identify an operation's name, the start and end timestamps, and the elapsed execution time.
+
+##### `op_name`
+Unique identifier for an operation.
+You can filter on all operations of a specific name.
+
+**Example:** `op_name=tsm1_compact_group`
+
+##### `op_event`
+Specifies the start and end of an event.
+The two possible values, `(start)` or `(end)`, are used to indicate when an operation started or ended.
+For example, you can grep by values in `op_name` AND `op_event` to find all starting operation log entries.
+For an example of this, see [Finding all starting log entries](#finding-all-starting-operation-log-entries).
+
+**Example:** `op_event=start`
+
+##### `op_elapsed`
+Duration of the operation execution.
+Logged with the ending trace log entry.
+Valid duration units are `ns`, `µs`, `ms`, and `s`.
+
+**Example:** `op_elapsed=352ms`
+
+
+#### Log identifier context key
+
+The log identifier key (`log_id`) lets you easily identify _every_ log entry for a single execution of an `influxd` process.
+There are other ways a log file could be split by a single execution, but the consistent `log_id` eases the searching of log aggregation services.
+
+**Example:** `log_id=06QknqtW000`
+
+#### Database context keys
+
+- **db\_instance**: Database name
+- **db\_rp**: Retention policy name
+- **db\_shard\_id**: Shard identifier
+- **db\_shard\_group**: Shard group identifier
+
+### Tooling
+
+Here are a couple of popular tools available for processing and filtering log files output in `logfmt` or `json` formats.
+
+#### hutils
+
+The [hutils](https://blog.heroku.com/hutils-explore-your-structured-data-logs) utility collection, provided by Heroku, provides tools for working with `logfmt`-encoded logs, including:
+
+- **lcut**: Extracts values from a `logfmt` trace based on a specified field name.
+- **lfmt**: Prettifies `logfmt` lines as they emerge from a stream, and highlights their key sections.
+- **ltap**: Accesses messages from log providers in a consistent way to allow easy parsing by other utilities that operate on `logfmt` traces.
+- **lviz**: Visualizes `logfmt` output by building a tree out of a dataset combining common sets of key-value pairs into shared parent nodes.
+
+#### lnav (Log File Navigator)
+
+The [lnav (Log File Navigator)](http://lnav.org) is an advanced log file viewer useful for watching and analyzing your log files from a terminal.
+The lnav viewer provides a single log view, automatic log format detection, filtering, timeline view, pretty-print view, and querying logs using SQL.
+
+### Operations
+
+The following operations, listed by their operation name (`op_name`) are traced in InfluxDB internal logs and available for use without changes in logging level.
+
+#### Initial opening of data files
+
+The `tsdb_open` operation traces include all events related to the initial opening of the `tsdb_store`.
+
+
+#### Retention policy shard deletions
+
+The `retention.delete_check` operation includes all shard deletions related to the retention policy.
+
+#### TSM snapshotting in-memory cache to disk
+
+The `tsm1_cache_snapshot` operation represents the snapshotting of the TSM in-memory cache to disk.
+
+#### TSM compaction strategies
+
+The `tsm1_compact_group` operation includes all trace log entries related to TSM compaction strategies and displays the related TSM compaction strategy keys:
+
+- **tsm1\_strategy**: level or full
+- **tsm1\_level**: 1, 2, or 3
+- **tsm\_optimize**: true or false
+
+#### Series file compactions
+
+The `series_partition_compaction` operation includes all trace log entries related to series file compactions.
+
+#### Continuous query execution (if logging enabled)
+
+The `continuous_querier_execute` operation includes all continuous query executions, if logging is enabled.
+
+#### TSI log file compaction
+
+The `tsi1_compact_log_file` operation includes all trace log entries related to log file compactions.
+
+#### TSI level compaction
+
+The `tsi1_compact_to_level` operation includes all trace log entries for TSI level compactions.
+
+
+### Tracing examples
+
+#### Finding all trace log entries for an InfluxDB operation
+
+In the example below, you can see the log entries for all trace operations related to a "TSM compaction" process.
+Note that the initial entry shows the message "TSM compaction (start)" and the final entry displays the message "TSM compaction (end)".
+{{% note %}}
+ Log entries were grepped using the `trace_id` value and then the specified key values were displayed using `lcut` (an `hutils` tool).
+ {{% /note %}}\]
+
+```
+$ grep "06QW92x0000" influxd.log | lcut ts lvl msg strategy level
+2018-02-21T20:18:56.880065Z info TSM compaction (start) full
+2018-02-21T20:18:56.880162Z info Beginning compaction full
+2018-02-21T20:18:56.880185Z info Compacting file full
+2018-02-21T20:18:56.880211Z info Compacting file full
+2018-02-21T20:18:56.880226Z info Compacting file full
+2018-02-21T20:18:56.880254Z info Compacting file full
+2018-02-21T20:19:03.928640Z info Compacted file full
+2018-02-21T20:19:03.928687Z info Finished compacting files full
+2018-02-21T20:19:03.928707Z info TSM compaction (end) full
+```
+
+
+#### Finding all starting operation log entries
+
+To find all starting operation log entries, you can grep by values in `op_name` AND `op_event`.
+In the following example, the grep returned 101 entries, so the result below only displays the first entry.
+In the example result entry, the timestamp, level, strategy, trace_id, op_name, and op_event values are included.
+
+```
+$ grep -F 'op_name=tsm1_compact_group' influxd.log | grep -F 'op_event=start'
+ts=2018-02-21T20:16:16.709953Z lvl=info msg="TSM compaction" log_id=06QVNNCG000 engine=tsm1 level=1 strategy=level trace_id=06QV~HHG000 op_name=tsm1_compact_group op_event=start
+...
+```
+
+Using the `lcut` utility (in hutils), the following command uses the previous `grep` command, but adds an `lcut` command to only display the keys and their values for keys that are not identical in all of the entries.
+The following example includes 19 examples of unique log entries displaying selected keys: `ts`, `strategy`, `level`, and `trace_id`.
+
+```
+$ grep -F 'op_name=tsm1_compact_group' influxd.log | grep -F 'op_event=start' | lcut ts strategy level trace_id | sort -u
+2018-02-21T20:16:16.709953Z level 1 06QV~HHG000
+2018-02-21T20:16:40.707452Z level 1 06QW0k0l000
+2018-02-21T20:17:04.711519Z level 1 06QW2Cml000
+2018-02-21T20:17:05.708227Z level 2 06QW2Gg0000
+2018-02-21T20:17:29.707245Z level 1 06QW3jQl000
+2018-02-21T20:17:53.711948Z level 1 06QW5CBl000
+2018-02-21T20:18:17.711688Z level 1 06QW6ewl000
+2018-02-21T20:18:56.880065Z full 06QW92x0000
+2018-02-21T20:20:46.202368Z level 3 06QWFizW000
+2018-02-21T20:21:25.292557Z level 1 06QWI6g0000
+2018-02-21T20:21:49.294272Z level 1 06QWJ_RW000
+2018-02-21T20:22:13.292489Z level 1 06QWL2B0000
+2018-02-21T20:22:37.292431Z level 1 06QWMVw0000
+2018-02-21T20:22:38.293320Z level 2 06QWMZqG000
+2018-02-21T20:23:01.293690Z level 1 06QWNygG000
+2018-02-21T20:23:25.292956Z level 1 06QWPRR0000
+2018-02-21T20:24:33.291664Z full 06QWTa2l000
+2018-02-21T21:12:08.017055Z full 06QZBpKG000
+2018-02-21T21:12:08.478200Z full 06QZBr7W000
+```
diff --git a/content/enterprise_influxdb/v1.9/guides/fine-grained-authorization.md b/content/enterprise_influxdb/v1.9/guides/fine-grained-authorization.md
index 4f08ed03b..c9cadfbb5 100644
--- a/content/enterprise_influxdb/v1.9/guides/fine-grained-authorization.md
+++ b/content/enterprise_influxdb/v1.9/guides/fine-grained-authorization.md
@@ -16,10 +16,18 @@ related:
Use fine-grained authorization (FGA) in InfluxDB Enterprise to control user access at the database, measurement, and series levels.
-> **Note:** InfluxDB OSS controls access at the database level only.
+{{% note %}}
+**Note:** InfluxDB OSS controls access at the database level only.
+{{% /note %}}
You must have [admin permissions](/enterprise_influxdb/v1.9/administration/authentication_and_authorization/#admin-user-management) to set up FGA.
+{{% warn %}}
+#### FGA does not apply to Flux
+FGA does not restrict actions performed by Flux queries (both read and write).
+If using FGA, we recommend [disabling Flux](/enterprise_influxdb/v{{< current-version >}}/flux/installation/).
+{{% /warn %}}
+
## Set up fine-grained authorization
1. [Enable authentication](/enterprise_influxdb/v1.9/administration/authentication_and_authorization/#set-up-authentication) in your InfluxDB configuration file.
@@ -34,19 +42,25 @@ You must have [admin permissions](/enterprise_influxdb/v1.9/administration/authe
3. Ensure that you can access the **meta node** API (port 8091 by default).
- > In a typical cluster configuration, the HTTP ports for data nodes
- > (8086 by default) are exposed to clients but the meta node HTTP ports are not.
- > You may need to work with your network administrator to gain access to the meta node HTTP ports.
+ {{% note %}}
+In a typical cluster configuration, the HTTP ports for data nodes
+(8086 by default) are exposed to clients but the meta node HTTP ports are not.
+You may need to work with your network administrator to gain access to the meta node HTTP ports.
+ {{% /note %}}
4. _(Optional)_ [Create roles](#manage-roles).
Roles let you grant permissions to groups of users assigned to each role.
- > For an overview of how users and roles work in InfluxDB Enterprise, see [InfluxDB Enterprise users](/enterprise_influxdb/v1.9/features/users/).
+ {{% note %}}
+For an overview of how users and roles work in InfluxDB Enterprise, see [InfluxDB Enterprise users](/enterprise_influxdb/v1.9/features/users/).
+ {{% /note %}}
5. [Set up restrictions](#manage-restrictions).
Restrictions apply to all non-admin users.
- > Permissions (currently "read" and "write") may be restricted independently depending on the scenario.
+ {{% note %}}
+Permissions (currently "read" and "write") may be restricted independently depending on the scenario.
+ {{% /note %}}
7. [Set up grants](#manage-grants) to remove restrictions for specified users and roles.
diff --git a/content/flux/v0.x/stdlib/universe/time.md b/content/flux/v0.x/stdlib/universe/time.md
index 083b43f84..e69b8eb27 100644
--- a/content/flux/v0.x/stdlib/universe/time.md
+++ b/content/flux/v0.x/stdlib/universe/time.md
@@ -48,7 +48,7 @@ time(v: "2021-01-01T00:00:00Z")
#### Convert an integer to a time value
```js
-int(v: 609459200000000000)
+time(v: 609459200000000000)
// Returns 2021-01-01T00:00:00Z
```
diff --git a/content/influxdb/cloud/account-management/change-password.md b/content/influxdb/cloud/account-management/change-password.md
index 3c2614157..6815a2468 100644
--- a/content/influxdb/cloud/account-management/change-password.md
+++ b/content/influxdb/cloud/account-management/change-password.md
@@ -4,6 +4,7 @@ seotitle: Change your InfluxDB Cloud password
description: >
To update your InfluxDB Cloud password, click the **Forgot Password** link on
the [InfluxDB Cloud login page](https://cloud2.influxdata.com/login).
+ Passwords must be at least 8 characters in length, and must not contain common words, personal information, or previous passwords.
menu:
influxdb_cloud:
name: Change your password
@@ -17,3 +18,13 @@ To change or reset your InfluxDB Cloud password:
2. Open the **InfluxCloud: Password Change Requested** email sent to the email
address associated with your InfluxDB Cloud account, click the **Reset Password**
button, and then enter and confirm a new password.
+
+### Password requirements
+
+Passwords must meet the following requirements:
+
+- Must be longer than 8 characters.
+- Must not contain personal information.
+- Must not be a common or previous password.
+
+These requirements follow the National Institute of Standards and Technology (NIST) standards for 2021.
diff --git a/content/influxdb/cloud/account-management/pricing-plans.md b/content/influxdb/cloud/account-management/pricing-plans.md
index 1def55fc9..50081fb0e 100644
--- a/content/influxdb/cloud/account-management/pricing-plans.md
+++ b/content/influxdb/cloud/account-management/pricing-plans.md
@@ -77,7 +77,7 @@ The Usage-Based Plan uses the following pricing vectors to calculate InfluxDB Cl
- Each individual operation—including queries, tasks, alerts, notifications, and Data Explorer activity—is one billable query operation.
- Refreshing a dashboard with multiple cells will incur multiple query operations.
- Failed operations aren’t counted.
-- **Data In** is the amount of data you’re writing into InfluxDB (measured in MB/second).
+- **Data In** is the amount of data you’re writing into InfluxDB (measured in MB).
- **Storage** is the amount of data you’re storing in InfluxDB (measured in GB/hour).
Discover how to [manage InfluxDB Cloud billing](/influxdb/cloud/account-management/billing/).
diff --git a/content/influxdb/v2.1/reference/cli/influx/auth/create.md b/content/influxdb/v2.1/reference/cli/influx/auth/create.md
index 09e71bd80..45a9d1c8c 100644
--- a/content/influxdb/v2.1/reference/cli/influx/auth/create.md
+++ b/content/influxdb/v2.1/reference/cli/influx/auth/create.md
@@ -41,7 +41,7 @@ influx auth create [flags]
| | `--read-orgs` | Grant permission to read organizations | | |
| | `--read-tasks` | Grant permission to read tasks | | |
| | `--read-telegrafs` | Grant permission to read Telegraf configurations | | |
-| | `--read-user` | Grant permission to read organization users | | |
+| | `--read-users` | Grant permission to read organization users | | |
| | `--skip-verify` | Skip TLS certificate verification | | `INFLUX_SKIP_VERIFY` |
| `-t` | `--token` | API token | string | `INFLUX_TOKEN` |
| `-u` | `--user` | Username | string | |
@@ -55,7 +55,7 @@ influx auth create [flags]
| | `--write-orgs` | Grant permission to create and update organizations | | |
| | `--write-tasks` | Grant permission to create and update tasks | | |
| | `--write-telegrafs` | Grant permission to create and update Telegraf configurations | | |
-| | `--write-user` | Grant permission to create and update organization users | | |
+| | `--write-users` | Grant permission to create and update organization users | | |
## Examples
@@ -100,7 +100,7 @@ influx auth create \
--read-orgs \
--read-tasks \
--read-telegrafs \
- --read-user \
+ --read-users \
--write-buckets \
--write-checks \
--write-dashboards \
@@ -110,7 +110,7 @@ influx auth create \
--write-orgs \
--write-tasks \
--write-telegrafs \
- --write-user
+ --write-users
```
### Create an API token with read and write access to specific buckets
@@ -136,5 +136,5 @@ influx auth create \
--read-orgs \
--read-tasks \
--read-telegrafs \
- --read-user
+ --read-users
```
diff --git a/content/influxdb/v2.1/tools/influx-cli.md b/content/influxdb/v2.1/tools/influx-cli.md
index b2ae46ab2..0e1b03d9f 100644
--- a/content/influxdb/v2.1/tools/influx-cli.md
+++ b/content/influxdb/v2.1/tools/influx-cli.md
@@ -154,7 +154,7 @@ brew list | grep influxdb-cli
sudo cp influxdb2-client-{{< latest-patch cli=true >}}-linux-amd64/influx /usr/local/bin/
# arm
- sudo cp influxdb2-client-{{< latest-patch cli=true >}}-linux-amd64/influx /usr/local/bin/
+ sudo cp influxdb2-client-{{< latest-patch cli=true >}}-linux-arm64/influx /usr/local/bin/
```
If you do not move the `influx` binary into your `$PATH`, prefix the executable