--- title: Configure your InfluxDB cluster using Helm description: > Use Helm to configure and deploy your InfluxDB Clustered `AppInstance` resource. menu: influxdb_clustered: name: Use Helm parent: Configure your cluster weight: 230 related: - /influxdb/clustered/admin/users/ --- Manage your InfluxDB Clustered deployments using Kubernetes and apply configuration settings using a YAML configuration file. The [InfluxDB Clustered Helm chart](https://github.com/influxdata/helm-charts/tree/master/charts/influxdb3-clustered) provides an alternative method for deploying your InfluxDB cluster using [Helm](https://helm.sh/). It acts as a wrapper for the InfluxDB `AppInstance` resource. When using Helm, apply configuration options in a a `values.yaml` on your local machine. InfluxData provides the following items: - **`influxdb-docker-config.json`**: an authenticated Docker configuration file. The InfluxDB Clustered software is in a secure container registry. This file grants access to the collection of container images required to install InfluxDB Clustered. --- ## Configuration data When ready to install InfluxDB, have the following information available: - **InfluxDB cluster hostname**: the hostname Kubernetes uses to expose InfluxDB API endpoints - **PostgreSQL-style data source name (DSN)**: used to access your PostgreSQL-compatible database that stores the InfluxDB Catalog. - **Object store credentials** _(AWS S3 or S3-compatible)_ - Endpoint URL - Access key - Bucket name - Region (required for S3, may not be required for other object stores) - **Local storage information** _(for ingester pods)_ - Storage class - Storage size - **OAuth2 provider credentials** - Client ID - JWKS endpoint - Device authorization endpoint - Token endpoint InfluxDB is deployed to a Kubernetes namespace which, throughout the following installation procedure, is referred to as the _target_ namespace. For simplicity, we assume this namespace is `influxdb`, however you may use any name you like. ### AppInstance resource The InfluxDB installation, update, and upgrade processes are driven by editing and applying a [Kubernetes custom resource (CRD)](https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/) called `AppInstance`. The `AppInstance` CRD is included in the InfluxDB Clustered Helm chart and can be configured by applying custom settings in the `values.yaml` included in the chart. The `AppInstance` resource contains key information, such as: - Name of the target namespace - Version of the InfluxDB package - Reference to the InfluxDB container registry pull secrets - Hostname where the InfluxDB API is exposed - Parameters to connect to [external prerequisites](/influxdb/clustered/install/prerequisites/) ### Kubit operator The InfluxDB Clustered Helm chart also includes the [`kubecfg kubit` operator](https://github.com/kubecfg/kubit) (maintained by InfluxData) which simplifies the installation and management of the InfluxDB Clustered package. It manages the application of the jsonnet templates used to install, manage, and update an InfluxDB cluster. ## Configure your cluster 1. [Install Helm](#install-helm) 2. [Create a values.yaml file](#create-a-valuesyaml-file) 3. [Create a namespace for InfluxDB](#create-a-namespace-for-influxdb) 4. [Configure access to the InfluxDB container registry](#configure-access-to-the-influxdb-container-registry) 5. [Set up cluster ingress](#set-up-cluster-ingress) 6. [Modify the configuration file to point to prerequisites](#modify-the-configuration-file-to-point-to-prerequisites) 7. [Provide a custom certificate authority bundle](#provide-a-custom-certificate-authority-bundle) (Optional) ### Install Helm If you haven't already, [install Helm](https://helm.sh/docs/intro/install/) on your local machine. ### Create a values.yaml file Download or copy the base `values.yaml` for the InfluxDB Clustered Helm chart from GitHub and store it locally. For example--if using cURL: ```bash curl -O https://raw.githubusercontent.com/influxdata/helm-charts/master/charts/influxdb3-clustered/values.yaml ``` Or you can copy the default `values.yaml` from GitHub: View values.yaml on GitHub ### Create a namespace for InfluxDB Create a namespace for InfluxDB. For example, using `kubectl`: ```sh kubectl create namespace influxdb ``` If you use a namespace name other than `influxdb`, update the `namespaceOverride` field in your `values.yaml` to use your custom namespace name. ### Configure access to the InfluxDB container registry The provided `influxdb-docker-config.json` grants access to a collection of container images required to run InfluxDB Clustered. Your Kubernetes Cluster needs access to the container registry to pull down and install InfluxDB. When pulling images, there are two main scenarios: - You have a Kubernetes cluster that can pull from the InfluxData container registry. - You run in an environment with no network interfaces ("air-gapped") and you can only access a private container registry. In both scenarios, you need a valid container registry secret file. Use [crane](https://github.com/google/go-containerregistry/tree/main/cmd/crane) to create a container registry secret file. 1. [Install crane](https://github.com/google/go-containerregistry/tree/main/cmd/crane#installation) 2. Use the following command to create a container registry secret file and retrieve the necessary secrets: {{% code-placeholders "PACKAGE_VERSION" %}} ```sh mkdir /tmp/influxdbsecret cp influxdb-docker-config.json /tmp/influxdbsecret/config.json DOCKER_CONFIG=/tmp/influxdbsecret \ crane manifest \ us-docker.pkg.dev/influxdb2-artifacts/clustered/influxdb:PACKAGE_VERSION ``` {{% /code-placeholders %}} --- Replace {{% code-placeholder-key %}}`PACKAGE_VERSION`{{% /code-placeholder-key %}} with your InfluxDB Clustered package version. --- If your Docker configuration is valid and you’re able to connect to the container registry, the command succeeds and the output is the JSON manifest for the Docker image, similar to the following: {{< expand-wrapper >}} {{% expand "View JSON manifest" %}} ```json { "schemaVersion": 2, "config": { "mediaType": "application/vnd.kubecfg.bundle.config.v1+json", "digest": "sha256:6900d2f248e678176c68f3768e7e48958bb96a59232070ff31b3b018cf299aa7", "size": 8598 }, "layers": [ { "mediaType": "application/vnd.kubecfg.bundle.tar+gzip", "digest": "sha256:7c1d62e76287035a9b22b2c155f328fae9beff2c6aa7a09a2dd2697539f41d98", "size": 404059 } ], "annotations": { "org.opencontainers.image.created": "1970-01-01T00:00:00Z", "org.opencontainers.image.revision": "unknown", "org.opencontainers.image.source": "kubecfg pack" } } ``` {{% /expand %}} {{< /expand-wrapper >}} If there’s a problem with the Docker configuration, crane won't retrieve the manifest and the output is similar to the following error: ```sh Error: fetching manifest us-docker.pkg.dev/influxdb2-artifacts/clustered/influxdb:: GET https://us-docker.pkg.dev/v2/token?scope=repository%3Ainfluxdb2-artifacts%2Fclustered%2Finfluxdb%3Apull&service=: DENIED: Permission "artifactregistry.repositories.downloadArtifacts" denied on resource "projects/influxdb2-artifacts/locations/us/repositories/clustered" (or it may not exist) ``` {{< tabs-wrapper >}} {{% tabs %}} [Public registry (non-air-gapped)](#) [Private registry (air-gapped)](#) {{% /tabs %}} {{% tab-content %}} #### Public registry (non-air-gapped) To pull from the InfluxData registry, you need to create a Kubernetes secret in the target namespace. ```sh kubectl create secret docker-registry gar-docker-secret \ --from-file=.dockerconfigjson=influxdb-docker-config.json \ --namespace influxdb ``` If successful, the output is the following: ```text secret/gar-docker-secret created ``` By default, this secret is named `gar-docker-secret`. If you change the name of this secret, you must also change the value of the `imagePullSecrets.name` field in your `values.yaml`. {{% /tab-content %}} {{% tab-content %}} #### Private registry (air-gapped) If your Kubernetes cluster can't use a public network to download container images from our container registry, do the following: 1. Copy the images from the InfluxDB registry to your own private registry. 2. Configure your `AppInstance` resource with a reference to your private registry name. 3. Provide credentials to your private registry. The list of images that you need to copy is included in the package metadata. You can obtain it with any standard OCI image inspection tool. For example: {{% code-placeholders "PACKAGE_VERSION" %}} ```sh DOCKER_CONFIG=/tmp/influxdbsecret \ crane config \ us-docker.pkg.dev/influxdb2-artifacts/clustered/influxdb:PACKAGE_VERSION \ | jq -r '.metadata["oci.image.list"].images[]' \ > /tmp/images.txt ``` {{% /code-placeholders %}} The output is a list of image names, similar to the following: ``` us-docker.pkg.dev/influxdb2-artifacts/idpe/idpe-cd-ioxauth@sha256:5f015a7f28a816df706b66d59cb9d6f087d24614f485610619f0e3a808a73864 us-docker.pkg.dev/influxdb2-artifacts/iox/iox@sha256:b59d80add235f29b806badf7410239a3176bc77cf2dc335a1b07ab68615b870c ... ``` Use `crane` to copy the images to your private registry: {{% code-placeholders "REGISTRY_HOSTNAME" %}} ```sh {{% /tab-content %}} {{< /tabs-wrapper >}} ### Set up cluster ingress {{% note %}} InfluxDB Clustered components use gRPC/HTTP2 protocols. If using an external load balancer, you may need to explicitly enable these protocols on your load balancers. {{% /note %}} The InfluxDB Clustered Helm chart includes the [Kubernetes Nginx Ingress Controller](https://github.com/kubernetes/ingress-nginx). Add a valid TLS Certificate to the cluster as a secret. Provide the paths to the TLS certificate file and key file: {{% code-placeholders "TLS_(CERT|KEY)_PATH" %}} ```sh kubectl create secret tls ingress-tls \ --namespace influxdb \ --cert TLS_CERT_PATH \ --key TLS_KEY_PATH ``` {{% /code-placeholders %}} --- Replace the following: - _{{% code-placeholder-key %}}`TLS_CERT_PATH`{{% /code-placeholder-key %}}: Path to the certificate file on your local machine._ - _{{% code-placeholder-key %}}`TLS_KEY_PATH`{{% /code-placeholder-key %}}: Path to the certificate secret key file on your local machine._ --- Provide the TLS certificate secret to the InfluxDB configuration in the [Configure ingress step](#configure-ingress). ### Modify the configuration file to point to prerequisites Update your `values.yaml` file with credentials necessary to connect your cluster to your prerequisites. - [Configure ingress](#configure-ingress) - [Configure the object store](#configure-the-object-store) - [Configure the catalog database](#configure-the-catalog-database) - [Configure local storage for ingesters](#configure-local-storage-for-ingesters) - [Configure your OAuth2 provider](#configure-your-oauth2-provider) - [Configure the size of your cluster](#configure-the-size-of-your-cluster) #### Configure ingress To configure ingress, provide values for the following fields in your `values.yaml`: - **`ingress.hosts`: Cluster hostnames** Provide the hostnames that Kubernetes should use to expose the InfluxDB API endpoints--for example: `{{< influxdb/host >}}`. _You can provide multiple hostnames. The ingress layer accepts incoming requests for all listed hostnames. This can be useful if you want to have distinct paths for your internal and external traffic._ {{% note %}} You are responsible for configuring and managing DNS. Options include: - Manually managing DNS records - Using [external-dns](https://github.com/kubernetes-sigs/external-dns) to synchronize exposed Kubernetes services and ingresses with DNS providers. {{% /note %}} - **`ingress.tlsSecretName`: TLS certificate secret name** Provide the name of the secret that [contains your TLS certificate and key](#set-up-cluster-ingress). The examples in this guide use the name `ingress-tls`. _The `tlsSecretName` field is optional. You may want to use it if you already have a TLS certificate for your DNS name._ {{< expand-wrapper >}} {{% expand "Use cert-manager and Let's Encrypt to manage TLS certificates" %}} If you instead want to automatically create an [ACME](https://datatracker.ietf.org/doc/html/rfc8555) certificate (for example, using [Let's Encrypt](https://letsencrypt.org/)), refer to the [cert-manager documentation](https://cert-manager.io/docs/usage/ingress/). In `ingress.tlsSecretName`, provide a name for the secret it should create. {{% note %}} If you choose to use cert-manager, it's your responsibility to install and configure it. {{% /note %}} {{% /expand %}} {{< /expand-wrapper >}} {{% code-callout "ingress-tls|cluster-host\.com" "green" %}} ```yaml ingress: hosts: - {{< influxdb/host >}} tlsSecretName: ingress-tls ``` {{% /code-callout %}} #### Configure the object store To connect your InfluxDB cluster to your object store. The information required to connect to your object store depends on your object storage provider. {{< tabs-wrapper >}} {{% tabs %}} [Amazon S3 or S3-compatible](#) [Azure Blob Storage](#) [Google Cloud Storage](#) {{% /tabs %}} {{% tab-content %}} If using Amazon S3 or an S3-compatible object store, provide values for the following fields in your `values.yaml`: - `objectStore` - `bucket`: Object storage bucket name - `s3`: - `endpoint`: Object storage endpoint URL - `allowHttp`: _Set to `true` to allow unencrypted HTTP connections_ - `accessKey.value`: Object storage access key _(can use a `value` literal or `valueFrom` to retrieve the value from a secret)_ - `secretKey.value`: Object storage secret key _(can use a `value` literal or `valueFrom` to retrieve the value from a secret)_ - `region`: Object storage region {{% code-placeholders "S3_(URL|ACCESS_KEY|SECRET_KEY|BUCKET_NAME|REGION)" %}} ```yml objectStore: # Bucket that the Parquet files will be stored in bucket: S3_BUCKET_NAME s3: # URL for S3 Compatible object store endpoint: S3_URL # Set to true to allow communication over HTTP (instead of HTTPS) allowHttp: 'false' # S3 Access Key # This can also be provided as a valueFrom: secretKeyRef: accessKey: value: S3_ACCESS_KEY # S3 Secret Key # This can also be provided as a valueFrom: secretKeyRef: secretKey: value: S3_SECRET_KEY # This value is required for AWS S3, it may or may not be required for other providers. region: S3_REGION ``` {{% /code-placeholders %}} --- Replace the following: - {{% code-placeholder-key %}}`S3_BUCKET_NAME`{{% /code-placeholder-key %}}: Object storage bucket name - {{% code-placeholder-key %}}`S3_URL`{{% /code-placeholder-key %}}: Object storage endpoint URL - {{% code-placeholder-key %}}`S3_ACCESS_KEY`{{% /code-placeholder-key %}}: Object storage access key - {{% code-placeholder-key %}}`S3_SECRET_KEY`{{% /code-placeholder-key %}}: Object storage secret key - {{% code-placeholder-key %}}`S3_REGION`{{% /code-placeholder-key %}}: Object storage region --- {{% /tab-content %}} {{% tab-content %}} If using Azure Blob Storage as your object store, provide values for the following fields in your `values.yaml`: - `objectStore` - `bucket`: Azure Blob Storage bucket name - `azure`: - `accessKey.value`: Azure Blob Storage access key _(can use a `value` literal or `valueFrom` to retrieve the value from a secret)_ - `account.value`: Azure Blob Storage account ID _(can use a `value` literal or `valueFrom` to retrieve the value from a secret)_ {{% code-placeholders "AZURE_(BUCKET_NAME|ACCESS_KEY|STORAGE_ACCOUNT)" %}} ```yml objectStore: # Bucket that the Parquet files will be stored in bucket: AZURE_BUCKET_NAME azure: # Azure Blob Storage Access Key # This can also be provided as a valueFrom: accessKey: value: AZURE_ACCESS_KEY # Azure Blob Storage Account # This can also be provided as a valueFrom: secretKeyRef: account: value: AZURE_STORAGE_ACCOUNT ``` {{% /code-placeholders %}} --- Replace the following: - {{% code-placeholder-key %}}`AZURE_BUCKET_NAME`{{% /code-placeholder-key %}}: Object storage bucket name - {{% code-placeholder-key %}}`AZURE_ACCESS_KEY`{{% /code-placeholder-key %}}: Azure Blob Storage access key - {{% code-placeholder-key %}}`AZURE_STORAGE_ACCOUNT`{{% /code-placeholder-key %}}: Azure Blob Storage account ID --- {{% /tab-content %}} {{% tab-content %}} If using Google Cloud Storage as your object store, provide values for the following fields in your `values.yaml`: - `objectStore` - `bucket`: Google Cloud Storage bucket name - `google`: - `serviceAccountSecret.name`: the Kubernetes Secret name that contains your Google IAM service account credentials - `serviceAccountSecret.key`: the key inside of your Google IAM secret that contains your Google IAM account credentials {{% code-placeholders "GOOGLE_(BUCKET_NAME|IAM_SECRET|CREDENTIALS_KEY)" %}} ```yml objectStore: # Bucket that the Parquet files will be stored in bucket: GOOGLE_BUCKET_NAME google: # This section is not needed if you are using GKE Workload Identity. # It is only required to use explicit service account secrets (JSON files) serviceAccountSecret: # Kubernetes Secret name containing the credentials for a Google IAM # Service Account. name: GOOGLE_IAM_SECRET # The key within the Secret containing the credentials. key: GOOGLE_CREDENTIALS_KEY ``` {{% /code-placeholders %}} --- Replace the following: - {{% code-placeholder-key %}}`GOOGLE_BUCKET_NAME`{{% /code-placeholder-key %}}: Google Cloud Storage bucket name - {{% code-placeholder-key %}}`GOOGLE_IAM_SECRET`{{% /code-placeholder-key %}}: the Kubernetes Secret name that contains your Google IAM service account credentials - {{% code-placeholder-key %}}`GOOGLE_CREDENTIALS_KEY`{{% /code-placeholder-key %}}: the key inside of your Google IAM secret that contains your Google IAM account credentials --- {{% /tab-content %}} {{< /tabs-wrapper >}} #### Configure the catalog database The InfluxDB catalog is a PostgreSQL-compatible relational database that stores metadata about your time series data. To connect your InfluxDB cluster to your PostgreSQL-compatible database, provide values for the following fields in your `values.yaml`: {{% note %}} We recommend storing sensitive credentials, such as your PostgreSQL-compatible DSN, as secrets in your Kubernetes cluster. {{% /note %}} - `catalog.dsn` - `SecretName`: Secret name - `SecretKey`: Key in the secret that contains the DSN {{% code-placeholders "SECRET_(NAME|KEY)" %}} ```yml catalog: # Secret name and key within the secret containing the dsn string to connect # to the catalog dsn: # Kubernetes Secret name containing the dsn for the catalog. SecretName: SECRET_NAME # The key within the Secret containing the dsn. SecretKey: SECRET_KEY ``` {{% /code-placeholders %}} --- Replace the following: - {{% code-placeholder-key %}}`SECRET_NAME`{{% /code-placeholder-key %}}: Name of the secret containing your PostgreSQL-compatible DSN - {{% code-placeholder-key %}}`SECRET_KEY`{{% /code-placeholder-key %}}: Key in the secret that references your PostgreSQL-compatible DSN --- {{% note %}} ##### PostgreSQL instances without TLS or SSL If your PostgreSQL-compatible instance runs without TLS or SSL, you must include the `sslmode=disable` parameter in the DSN. For example: {{% code-callout "sslmode=disable" %}} ``` postgres://username:passw0rd@mydomain:5432/influxdb?sslmode=disable ``` {{% /code-callout %}} {{% /note %}} #### Configure local storage for ingesters InfluxDB ingesters require local storage to store the Write Ahead Log (WAL) for incoming data. To connect your InfluxDB cluster to local storage, provide values for the following fields in your `values.yaml`: - `ingesterStorage` - `storageClassName`: [Kubernetes storage class](https://kubernetes.io/docs/concepts/storage/storage-classes/). This differs based on the Kubernetes environment and desired storage characteristics. - `storage`: Storage size. We recommend a minimum of 2 gibibytes (`2Gi`). {{% code-placeholders "STORAGE_(CLASS|SIZE)" %}} ```yaml ingesterStorage: # (Optional) Set the storage class. This will differ based on the K8s # environment and desired storage characteristics. # If not set, the default storage class will be used. storageClassName: STORAGE_CLASS # Set the storage size (minimum 2Gi recommended) storage: STORAGE_SIZE ``` {{% /code-placeholders %}} --- Replace the following: - {{% code-placeholder-key %}}`STORAGE_CLASS`{{% /code-placeholder-key %}}: [Kubernetes storage class](https://kubernetes.io/docs/concepts/storage/storage-classes/) - {{% code-placeholder-key %}}`STORAGE_SIZE`{{% /code-placeholder-key %}}: Storage size (example: `2Gi`) --- #### Configure your OAuth2 provider InfluxDB Clustered uses OAuth2 to authenticate administrative access to your cluster. To connect your InfluxDB cluster to your OAuth2 provide, provide values for the following fields in your `values.yaml`: - `admin` - `identityProvider`: Identity provider name. _If using Microsoft Entra ID (formerly Azure Active Directory), set the name to `azure`_. - `jwksEndpoint`: JWKS endpoint provide by your identity provider. - `users`: List of OAuth2 users to grant administrative access to your InfluxDB cluster. IDs are provided by your identity provider. Below are examples for **Keycloak**, **Auth0**, and **Microsoft Entra ID**, but other OAuth2 providers should work as well: {{< code-tabs-wrapper >}} {{% code-tabs %}} [Keycloak](#) [Auth0](#) [Microsoft Entra ID](#) {{% /code-tabs %}} {{% code-tab-content %}} {{% code-callout "keycloak" "green" %}} {{% code-placeholders "KEYCLOAK_(HOST|REALM|USER_ID)" %}} ```yaml admin: # The identity provider to be used e.g. "keycloak", "auth0", "azure", etc # Note for Azure Active Directory it must be exactly "azure" identityProvider: keycloak # The JWKS endpoint provided by the Identity Provider jwksEndpoint: |- https://KEYCLOAK_HOST/auth/realms/KEYCLOAK_REALM/protocol/openid-connect/certs # The list of users to grant access to Clustered via influxctl users: # All fields are required but `firstName`, `lastName`, and `email` can be # arbitrary values. However, `id` must match the user ID provided by Keycloak. - id: KEYCLOAK_USER_ID firstName: Marty lastName: McFly email: mcfly@influxdata.com ``` {{% /code-placeholders %}} {{% /code-callout %}} --- Replace the following: - {{% code-placeholder-key %}}`KEYCLOAK_HOST`{{% /code-placeholder-key %}}: Host and port of your Keycloak server - {{% code-placeholder-key %}}`KEYCLOAK_REALM`{{% /code-placeholder-key %}}: Keycloak realm - {{% code-placeholder-key %}}`KEYCLOAK_USER_ID`{{% /code-placeholder-key %}}: Keycloak user ID to grant InfluxDB administrative access to --- {{% /code-tab-content %}} {{% code-tab-content %}} {{% code-callout "auth0" "green" %}} {{% code-placeholders "AUTH0_(HOST|USER_ID)" %}} ```yaml admin: # The identity provider to be used e.g. "keycloak", "auth0", "azure", etc # Note for Azure Active Directory it must be exactly "azure" identityProvider: auth0 # The JWKS endpoint provided by the Identity Provider jwksEndpoint: |- https://AUTH0_HOST/.well-known/openid-configuration # The list of users to grant access to Clustered via influxctl users: - AUTH0_USER_ID ``` {{% /code-placeholders %}} {{% /code-callout %}} --- Replace the following: - {{% code-placeholder-key %}}`AUTH0_HOST`{{% /code-placeholder-key %}}: Host and port of your Auth0 server - {{% code-placeholder-key %}}`AUTH0_USER_ID`{{% /code-placeholder-key %}}: Auth0 user ID to grant InfluxDB administrative access to --- {{% /code-tab-content %}} {{% code-tab-content %}} {{% code-callout "azure" "green" %}} {{% code-placeholders "AZURE_(USER|TENANT)_ID" %}} ```yaml admin: # The identity provider to be used e.g. "keycloak", "auth0", "azure", etc # Note for Azure Active Directory it must be exactly "azure" identityProvider: azure # The JWKS endpoint provided by the Identity Provider jwksEndpoint: |- https://login.microsoftonline.com/AZURE_TENANT_ID/discovery/v2.0/keys # The list of users to grant access to Clustered via influxctl users: - AZURE_USER_ID ``` {{% /code-placeholders %}} {{% /code-callout %}} --- Replace the following: - {{% code-placeholder-key %}}`AZURE_TENANT_ID`{{% /code-placeholder-key %}}: Microsoft Entra tenant ID - {{% code-placeholder-key %}}`AZURE_USER_ID`{{% /code-placeholder-key %}}: Microsoft Entra user ID to grant InfluxDB administrative access to _(See [Find user IDs with Microsoft Entra ID](/influxdb/clustered/install/auth/?t=Microsoft+Entra+ID#find-user-ids-with-microsoft-entra-id))_ --- {{% /code-tab-content %}} {{< /code-tabs-wrapper >}} ##### Add users Finally, add the users you wish to have access to use `influxctl`. Update the `admin.users` field with a list of the users. See [Manage users](/influxdb/clustered/admin/users/) for more details. #### Configure the size of your cluster By default, an InfluxDB cluster is configured with the following: - **3 ingesters**: Ensures redundancy on the write path. - **1 compactor**: While you can have multiple compactors, it is more efficient to scale the compactor vertically (assign more CPU and memory) rather than horizontally (increase the number of compactors). - **1 querier**: The optimal number of queriers depends on the number of concurrent queries you are likely to have and how long they take to execute. The default values provide a good starting point for testing. Once you have your cluster up and running and are looking for scaling recommendations, please [contact the InfluxData Support team](https://support.influxdata.com). We are happy to work with you to identify appropriate scale settings based on your anticipated workload. **To use custom scale settings for your InfluxDB cluster**, modify the following fields in your values.yaml`. If omitted, your cluster will use the default scale settings. - `resources` - `ingester.requests` - `cpu`: CPU resource units to assign to ingesters - `memory`: Memory resource units to assign to ingesters - `replicas`: Number of ingester replicas to provision - `compactor.requests` - `cpu`: CPU resource units to assign to compactors - `memory`: Memory resource units to assign to compactors - `replicas`: Number of compactor replicas to provision - `querier.requests` - `cpu`: CPU resource units to assign to queriers - `memory`: Memory resource units to assign to queriers - `replicas`: Number of querier replicas to provision - `router.requests` - `cpu`: CPU resource units to assign to routers - `memory`: Memory resource units to assign to routers - `replicas`: Number of router replicas to provision ###### Related Kubernetes documentation - [CPU resource units](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu) - [Memory resource units](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory) {{% code-placeholders "(INGESTER|COMPACTOR|QUERIER|ROUTER)_(CPU|MEMORY|REPLICAS)" %}} ```yml # The following settings tune the various pods for their cpu/memory/replicas # based on workload needs. Only uncomment the specific resources you want # to change. Anything left commented will use the package default. resources: # The ingester handles data being written ingester: requests: cpu: INGESTER_CPU memory: INGESTER_MEMORY replicas: INGESTER_REPLICAS # Default is 3 # The compactor reorganizes old data to improve query and storage efficiency. compactor: requests: cpu: COMPACTOR_CPU memory: COMPACTOR_MEMORY replicas: COMPACTOR_REPLICAS # Default is 1 # The querier handles querying data. querier: requests: cpu: QUERIER_CPU memory: QUERIER_MEMORY replicas: QUERIER_REPLICAS # Default is 1 # The router performs some api routing. router: requests: cpu: ROUTER_CPU memory: ROUTER_MEMORY replicas: ROUTER_REPLICAS # Default is 1 ``` {{% /code-placeholders %}} ### Provide a custom certificate authority bundle {note="Optional"} InfluxDB attempts to make TLS connections to the services it depends on; notably the [Catalog](/influxdb/clustered/reference/internals/storage-engine/#catalog), and the [Object store](/influxdb/clustered/reference/internals/storage-engine/#object-store). InfluxDB validates the certificates for all of the connections it makes. **If you host these services yourself and you use a private or otherwise not well-known certificate authority to issue certificates to theses services**, InfluxDB will not recognize the issuer and will be unable to validate the certificates. To allow InfluxDB to validate these certificates, provide a PEM certificate bundle containing your custom certificate authority chain. 1. Use `kubectl` to create a config map containing your PEM bundle. Your certificate authority administrator should provide you with a PEM-formatted certificate bundle file. {{% note %}} This PEM-formatted bundle file is *not* the certificate that InfluxDB uses to host its own TLS endpoints. This bundle establishes a chain of trust for the external services that InfluxDB depends on. {{% /note %}} In the example below, `private_ca.pem` is the certificate bundle file. ```sh kubectl --namespace influxdb create configmap custom-ca --from-file=certs.pem=/path/to/private_ca.pem ``` {{% note %}} It's possible to append multiple certificates into the same bundle. This can help if you need to include intermediate certificates or explicitly include leaf certificates. Leaf certificates should be included before any intermediate certificates they depend on. The root certificate should be last in the bundle. {{% /note %}} 2. Update your `values.yaml` to enable custom egress and refer to your certificate authority config map. Set `useCustomEgress` to `true` and update the `egress` property to refer to that config map. For example: ```yml useCustomEgress: true egress: # # If you're using a custom CA you will need to specify the full custom CA bundle here. # # # # NOTE: the custom CA is currently only honoured for outbound requests used to obtain # # the JWT public keys from your identiy provider (see `jwksEndpoint`). customCertificates: valueFrom: configMapKeyRef: key: ca.pem name: custom-ca ``` {{< page-nav prev="/influxdb/clustered/install/auth/" prevText="Set up authentication" next="/influxdb/clustered/install/licensing" nextText="Install your license" tab="Helm" >}}