fix: switch port to 8086

fix(readme): update the port and setup
pull/19494/head
Stuart Carnie 2020-09-03 14:40:29 -07:00
parent 50237d97ad
commit 7964f86ffc
No known key found for this signature in database
GPG Key ID: 848D9C9718D78B4F
43 changed files with 165 additions and 180 deletions

View File

@ -176,7 +176,7 @@ jobs:
node --version && npm --version node --version && npm --version
docker --version docker --version
google-chrome --version && which google-chrome && chromedriver --version && which chromedriver google-chrome --version && which google-chrome && chromedriver --version && which chromedriver
timeout 300 bash -c 'while [[ "$(curl -s -o /dev/null -w ''%{http_code}'' localhost:9999)" != "200" ]]; do sleep 5; done' || false timeout 300 bash -c 'while [[ "$(curl -s -o /dev/null -w ''%{http_code}'' localhost:8086)" != "200" ]]; do sleep 5; done' || false
- run: - run:
name: Selenium tests name: Selenium tests
command: | command: |

View File

@ -16,13 +16,13 @@ Ideally, test cases would be in the form of `curl` commands.
For example: For example:
```bash ```bash
# write data # write data
curl -XPOST "http://localhost:9999/api/v2/write?org=YOUR_ORG&bucket=YOUR_BUCKET&precision=s" \ curl -XPOST "http://localhost:8086/api/v2/write?org=YOUR_ORG&bucket=YOUR_BUCKET&precision=s" \
--header "Authorization: Token YOURAUTHTOKEN" \ --header "Authorization: Token YOURAUTHTOKEN" \
--data-raw "mem,host=host1 used_percent=23.43234543 1556896326" --data-raw "mem,host=host1 used_percent=23.43234543 1556896326"
# query data # query data
# Bug: expected it to return no data, but data comes back. # Bug: expected it to return no data, but data comes back.
curl http://localhost:9999/api/v2/query?org=my-org -XPOST -sS \ curl http://localhost:8086/api/v2/query?org=my-org -XPOST -sS \
-H 'Authorization: Token YOURAUTHTOKEN' \ -H 'Authorization: Token YOURAUTHTOKEN' \
-H 'Accept: application/csv' \ -H 'Accept: application/csv' \
-H 'Content-type: application/vnd.flux' \ -H 'Content-type: application/vnd.flux' \

View File

@ -41,7 +41,7 @@ cd ui
yarn && yarn start yarn && yarn start
``` ```
If there are no errors, hit [localhost:8080](http://localhost:8080) and follow the prompts to setup your username and password. *Note the port difference: `8080` vs the production `9999`* If there are no errors, hit [localhost:8080](http://localhost:8080) and follow the prompts to setup your username and password. *Note the port difference: `8080` vs the production `8086`*
You're set up to develop Influx locally. Any changes you make to front-end code under the `ui/` directory will be updated after the watcher process (that was initiated by running `yarn start`) sees them and rebuilds the bundle. Any changes to go code will need to be re-compiled by re-running the `go run` command above. You're set up to develop Influx locally. Any changes you make to front-end code under the `ui/` directory will be updated after the watcher process (that was initiated by running `yarn start`) sees them and rebuilds the bundle. Any changes to go code will need to be re-compiled by re-running the `go run` command above.
@ -65,7 +65,7 @@ Tab 1:
go run ./cmd/influxd --assets-path=ui/build go run ./cmd/influxd --assets-path=ui/build
``` ```
This starts the influxdb application server. It handles API requests and can be reached via `localhost:9999`. Any changes to go code will need to be re-compiled by re-running the `go run` command above. This starts the influxdb application server. It handles API requests and can be reached via `localhost:8086`. Any changes to go code will need to be re-compiled by re-running the `go run` command above.
Tab 2: Tab 2:

View File

@ -45,7 +45,7 @@ FROM debian:stretch-slim AS influx
COPY --from=dbuild-all /code/bin/linux/influxd /usr/bin/influxd COPY --from=dbuild-all /code/bin/linux/influxd /usr/bin/influxd
COPY --from=dbuild-all /code/bin/linux/influx /usr/bin/influx COPY --from=dbuild-all /code/bin/linux/influx /usr/bin/influx
EXPOSE 9999 EXPOSE 8086
ENTRYPOINT [ "/usr/bin/influxd" ] ENTRYPOINT [ "/usr/bin/influxd" ]

View File

@ -227,7 +227,7 @@ dshell-image:
@docker image build --build-arg "USERID=$(shell id -u)" -t influxdb:dshell --target dshell . @docker image build --build-arg "USERID=$(shell id -u)" -t influxdb:dshell --target dshell .
dshell: dshell-image dshell: dshell-image
@docker container run --rm -p 9999:9999 -p 8080:8080 -u $(shell id -u) -it -v $(shell pwd):/code -w /code influxdb:dshell @docker container run --rm -p 8086:8086 -p 8080:8080 -u $(shell id -u) -it -v $(shell pwd):/code -w /code influxdb:dshell
# .PHONY targets represent actions that do not create an actual file. # .PHONY targets represent actions that do not create an actual file.
.PHONY: all $(SUBDIRS) run fmt checkfmt tidy checktidy checkgenerate test test-go test-js test-go-race bench clean node_modules vet nightly chronogiraffe dist ping protoc e2e run-e2e influxd libflux flags dshell dclean docker-image-flux docker-image-influx goreleaser .PHONY: all $(SUBDIRS) run fmt checkfmt tidy checktidy checkgenerate test test-go test-js test-go-race bench clean node_modules vet nightly chronogiraffe dist ping protoc e2e run-e2e influxd libflux flags dshell dclean docker-image-flux docker-image-influx goreleaser

View File

@ -42,7 +42,7 @@ Planned activities include:
## Installing from Source ## Installing from Source
We have nightly and weekly versioned Docker images, Debian packages, RPM packages, and tarballs of InfluxDB available at the [InfluxData downloads page](https://portal.influxdata.com/downloads/). We have nightly and weekly versioned Docker images, Debian packages, RPM packages, and tarballs of InfluxDB available at the [InfluxData downloads page](https://portal.influxdata.com/downloads/). We also provide the `influx` command line interface (CLI) client as a separate binary available at the same location.
## Building From Source ## Building From Source
@ -133,17 +133,12 @@ Everything in InfluxDB is organized under a concept of an organization. The API
Buckets represent where you store time series data. Buckets represent where you store time series data.
They're synonymous with what was previously in InfluxDB 1.x a database and retention policy. They're synonymous with what was previously in InfluxDB 1.x a database and retention policy.
The simplest way to get set up is to point your browser to [http://localhost:9999](http://localhost:9999) and go through the prompts. The simplest way to get set up is to point your browser to [http://localhost:8086](http://localhost:8086) and go through the prompts.
**Note**: Port 9999 will be used during the beta phases of development of InfluxDB v2.0. You can also get set up from the CLI using the command `influx setup`:
This should allow a v2.0-beta instance to be run alongside a v1.x instance without interfering on port 8086.
InfluxDB will thereafter continue to use 8086.
You can also get set up from the CLI using the subcommands `influx user`, `influx auth`, `influx org` and `influx bucket`,
or do it all in one breath with `influx setup`:
``` ```bash
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx setup $ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx setup
Welcome to InfluxDB 2.0! Welcome to InfluxDB 2.0!
Please type your primary username: marty Please type your primary username: marty
@ -169,48 +164,50 @@ Confirm? (y/n): y
UserID Username Organization Bucket UserID Username Organization Bucket
033a3f2c5ccaa000 marty InfluxData Telegraf 033a3f2c5ccaa000 marty InfluxData Telegraf
Your token has been stored in /Users/marty/.influxdbv2/credentials Your token has been stored in /Users/marty/.influxdbv2/configs
``` ```
You may get into a development loop where `influx setup` becomes tedious. You can run this command non-interactively using the `-f, --force` flag if you are automating the setup.
Some added flags can help: Some added flags can help:
``` ```bash
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx setup --username marty --password F1uxKapacit0r85 --org InfluxData --bucket telegraf --retention 168 --token where-were-going-we-dont-need-roads --force $ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx setup \
--username marty \
--password F1uxKapacit0r85 \
--org InfluxData \
--bucket telegraf \
--retention 168 \
--token where-were-going-we-dont-need-roads \
--force
``` ```
`~/.influxdbv2/credentials` contains your auth token. Once setup is complete, a configuration profile is created to allow you to interact with your local InfluxDB without passing in credentials each time. You can list and manage those profiles using the `influx config` command.
Most `influx` commands read the token from this file path by default. ```bash
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx config
You may need the organization ID and bucket ID later: Active Name URL Org
* default http://localhost:8086 InfluxData
```
$ influx org find
ID Name
033a3f2c708aa000 InfluxData
```
```
$ influx bucket find
ID Name Retention Organization OrganizationID
033a3f2c710aa000 telegraf 72h0m0s InfluxData 033a3f2c708aa000
``` ```
## Writing Data
Write to measurement `m`, with tag `v=2`, in bucket `telegraf`, which belongs to organization `InfluxData`: Write to measurement `m`, with tag `v=2`, in bucket `telegraf`, which belongs to organization `InfluxData`:
```bash
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx write --bucket telegraf --precision s "m v=2 $(date +%s)"
``` ```
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx write --org InfluxData --bucket telegraf --precision s "m v=2 $(date +%s)"
``` Since you have a default profile set up, you can omit the Organization and Token from the command.
Write the same point using `curl`: Write the same point using `curl`:
``` ```bash
curl --header "Authorization: Token $(cat ~/.influxdbv2/credentials)" --data-raw "m v=2 $(date +%s)" "http://localhost:9999/api/v2/write?org=InfluxData&bucket=telegraf&precision=s" curl --header "Authorization: Token $(bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx auth list --json | jq -r '.[0].token')" \
--data-raw "m v=2 $(date +%s)" \
"http://localhost:8086/api/v2/write?org=InfluxData&bucket=telegraf&precision=s"
``` ```
Read that back with a simple Flux query: Read that back with a simple Flux query:
``` ```bash
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx query -o InfluxData 'from(bucket:"telegraf") |> range(start:-1h)' $ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx query 'from(bucket:"telegraf") |> range(start:-1h)'
Result: _result Result: _result
Table: keys: [_start, _stop, _field, _measurement] Table: keys: [_start, _stop, _field, _measurement]
_start:time _stop:time _field:string _measurement:string _time:time _value:float _start:time _stop:time _field:string _measurement:string _time:time _value:float
@ -218,19 +215,7 @@ Table: keys: [_start, _stop, _field, _measurement]
2019-12-30T22:19:39.043918000Z 2019-12-30T23:19:39.043918000Z v m 2019-12-30T23:17:02.000000000Z 2 2019-12-30T22:19:39.043918000Z 2019-12-30T23:19:39.043918000Z v m 2019-12-30T23:17:02.000000000Z 2
``` ```
Use the fancy REPL: Use the `-r, --raw` option to return the raw flux response from the query. This is useful for moving data from one instance to another as the `influx write` command can accept the Flux response using the `--format csv` option.
```
$ bin/$(uname -s | tr '[:upper:]' '[:lower:]')/influx repl -o InfluxData
> from(bucket:"telegraf") |> range(start:-1h)
Result: _result
Table: keys: [_start, _stop, _field, _measurement]
_start:time _stop:time _field:string _measurement:string _time:time _value:float
------------------------------ ------------------------------ ---------------------- ---------------------- ------------------------------ ----------------------------
2019-12-30T22:22:44.776351000Z 2019-12-30T23:22:44.776351000Z v m 2019-12-30T23:17:02.000000000Z 2
>
```
## Introducing Flux ## Introducing Flux

View File

@ -21,7 +21,7 @@ var DefaultSource = &chronograf.Source{
ID: math.MaxInt32, // Use large number to avoid possible collisions in older chronograf. ID: math.MaxInt32, // Use large number to avoid possible collisions in older chronograf.
Name: "autogen", Name: "autogen",
Type: "influx", Type: "influx",
URL: "http://localhost:9999", URL: "http://localhost:8086",
Default: false, Default: false,
} }

View File

@ -26,7 +26,7 @@ type Config struct {
// DefaultConfig is default config without token // DefaultConfig is default config without token
var DefaultConfig = Config{ var DefaultConfig = Config{
Name: "default", Name: "default",
Host: "http://localhost:9999", Host: "http://localhost:8086",
Active: true, Active: true,
} }

View File

@ -36,12 +36,12 @@ func TestWriteConfigs(t *testing.T) {
"default": Config{ "default": Config{
Token: "token1", Token: "token1",
Org: "org1", Org: "org1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
Active: true, Active: true,
}, },
}, },
result: `[default] result: `[default]
url = "http://localhost:9999" url = "http://localhost:8086"
token = "token1" token = "token1"
org = "org1" org = "org1"
active = true` + commentedStr, active = true` + commentedStr,

View File

@ -30,7 +30,7 @@ func TestCmdConfig(t *testing.T) {
flags: []string{ flags: []string{
"--config-name", "default", "--config-name", "default",
"--org", "org1", "--org", "org1",
"--host-url", "http://localhost:9999", "--host-url", "http://localhost:8086",
"--token", "tok1", "--token", "tok1",
"--active", "--active",
}, },
@ -40,7 +40,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
{ {
@ -48,7 +48,7 @@ func TestCmdConfig(t *testing.T) {
flags: []string{ flags: []string{
"-n", "default", "-n", "default",
"-o", "org1", "-o", "org1",
"-u", "http://localhost:9999", "-u", "http://localhost:8086",
"-t", "tok1", "-t", "tok1",
"-a", "-a",
}, },
@ -58,7 +58,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
{ {
@ -66,7 +66,7 @@ func TestCmdConfig(t *testing.T) {
flags: []string{ flags: []string{
"-n", "default", "-n", "default",
"-o", "org1", "-o", "org1",
"-u", "http://localhost:9999", "-u", "http://localhost:8086",
"-t", "tok1", "-t", "tok1",
"-a", "-a",
}, },
@ -83,7 +83,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
} }
@ -138,7 +138,7 @@ func TestCmdConfig(t *testing.T) {
"--configs-path=" + file, "--configs-path=" + file,
"-n", "default", "-n", "default",
"-o", "org1", "-o", "org1",
"-u", "http://localhost:9999", "-u", "http://localhost:8086",
"-t", "tok1", "-t", "tok1",
"-a", "-a",
} }
@ -152,7 +152,7 @@ func TestCmdConfig(t *testing.T) {
require.True(t, ok) require.True(t, ok)
assert.Equal(t, "default", cfg.Name) assert.Equal(t, "default", cfg.Name)
assert.Equal(t, "org1", cfg.Org) assert.Equal(t, "org1", cfg.Org)
assert.Equal(t, "http://localhost:9999", cfg.Host) assert.Equal(t, "http://localhost:8086", cfg.Host)
assert.Equal(t, "tok1", cfg.Token) assert.Equal(t, "tok1", cfg.Token)
assert.True(t, cfg.Active) assert.True(t, cfg.Active)
}) })
@ -199,7 +199,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: false, Active: false,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
expected: config.Config{ expected: config.Config{
@ -207,7 +207,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
{ {
@ -227,7 +227,7 @@ func TestCmdConfig(t *testing.T) {
Active: false, Active: false,
PreviousActive: true, PreviousActive: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
expected: config.Config{ expected: config.Config{
@ -235,7 +235,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
} }
@ -299,7 +299,7 @@ func TestCmdConfig(t *testing.T) {
flags: []string{ flags: []string{
"--config-name", "default", "--config-name", "default",
"--org", "org1", "--org", "org1",
"--host-url", "http://localhost:9999", "--host-url", "http://localhost:8086",
"--token", "tok1", "--token", "tok1",
"--active", "--active",
}, },
@ -308,7 +308,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
{ {
@ -362,7 +362,7 @@ func TestCmdConfig(t *testing.T) {
flags: []string{ flags: []string{
"-n", "default", "-n", "default",
"-o", "org1", "-o", "org1",
"-u", "http://localhost:9999", "-u", "http://localhost:8086",
"-t", "tok1", "-t", "tok1",
"-a", "-a",
}, },
@ -371,7 +371,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: true, Active: true,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
} }
@ -546,7 +546,7 @@ func TestCmdConfig(t *testing.T) {
Org: "org1", Org: "org1",
Active: false, Active: false,
Token: "tok1", Token: "tok1",
Host: "http://localhost:9999", Host: "http://localhost:8086",
}, },
}, },
}, },
@ -594,7 +594,7 @@ func testConfigInvalidURLs(t *testing.T, cmdName string, cmdFn func(*globalFlags
flags: []string{ flags: []string{
"--config-name", "default", "--config-name", "default",
"--org", "org1", "--org", "org1",
"--host-url", "localhost:9999", "--host-url", "localhost:8086",
"--token", "tok1", "--token", "tok1",
}, },
}, },

View File

@ -196,7 +196,7 @@ func launcherOpts(l *Launcher) []cli.Opt {
{ {
DestP: &l.httpBindAddress, DestP: &l.httpBindAddress,
Flag: "http-bind-address", Flag: "http-bind-address",
Default: ":9999", Default: ":8086",
Desc: "bind address for the REST HTTP API", Desc: "bind address for the REST HTTP API",
}, },
{ {

View File

@ -4301,7 +4301,7 @@ const telegrafCfg = `[agent]
collection_jitter = "0s" collection_jitter = "0s"
flush_interval = "10s" flush_interval = "10s"
[[outputs.influxdb_v2]] [[outputs.influxdb_v2]]
urls = ["http://localhost:9999"] urls = ["http://localhost:8086"]
token = "$INFLUX_TOKEN" token = "$INFLUX_TOKEN"
organization = "rg" organization = "rg"
bucket = "rucket_3" bucket = "rucket_3"

View File

@ -1,7 +1,7 @@
FROM debian:stable-slim FROM debian:stable-slim
COPY influx influxd /usr/bin/ COPY influx influxd /usr/bin/
EXPOSE 9999 EXPOSE 8086
ENV DEBIAN_FRONTEND noninteractive ENV DEBIAN_FRONTEND noninteractive

View File

@ -16,7 +16,7 @@ jobs:
node --version && npm --version node --version && npm --version
docker --version docker --version
google-chrome --version && which google-chrome && chromedriver --version && which chromedriver google-chrome --version && which google-chrome && chromedriver --version && which chromedriver
timeout 300 bash -c 'while [[ "$(curl -s -o /dev/null -w ''%{http_code}'' localhost:9999)" != "200" ]]; do sleep 5; done' || false timeout 300 bash -c 'while [[ "$(curl -s -o /dev/null -w ''%{http_code}'' localhost:8086)" != "200" ]]; do sleep 5; done' || false
- run: - run:
name: Selenium tests name: Selenium tests

View File

@ -55,7 +55,7 @@ For example, to declare the `influx_url` property in the `development` configura
e.g. e.g.
`export E2E_DEVELOPMENT_INFLUX_URL="http://172.17.0.2:9999"` `export E2E_DEVELOPMENT_INFLUX_URL="http://172.17.0.2:8086"`
This feature was added specifically to define passwords and usernames only via the test environment. However, it can be used with any configuration key value such as the INFLUX_URL endpoint. This feature was added specifically to define passwords and usernames only via the test environment. However, it can be used with any configuration key value such as the INFLUX_URL endpoint.
@ -121,7 +121,7 @@ e.g.
sudo docker run -it -v `pwd`/report:/home/e2e/report -v `pwd`/screenshots:/home/e2e/screenshots \ sudo docker run -it -v `pwd`/report:/home/e2e/report -v `pwd`/screenshots:/home/e2e/screenshots \
-v /tmp/e2e/etc:/home/e2e/etc -v /tmp/e2e/downloads:/home/e2e/downloads \ -v /tmp/e2e/etc:/home/e2e/etc -v /tmp/e2e/downloads:/home/e2e/downloads \
-e SELENIUM_REMOTE_URL="http://${SELENOID_HOST}:4444/wd/hub" \ -e SELENIUM_REMOTE_URL="http://${SELENOID_HOST}:4444/wd/hub" \
-e E2E_${ACTIVE_CONF^^}_INFLUX_URL="http://${INFLUX2_HOST}:9999" --detach \ -e E2E_${ACTIVE_CONF^^}_INFLUX_URL="http://${INFLUX2_HOST}:8086" --detach \
--name ${TEST_CONTAINER} e2e-${TEST_CONTAINER}:latest --name ${TEST_CONTAINER} e2e-${TEST_CONTAINER}:latest
``` ```

View File

@ -2,7 +2,7 @@
"active": "development", "active": "development",
"development" : { "development" : {
"config_id" : "development", "config_id" : "development",
"influx_url": "http://localhost:9999", "influx_url": "http://localhost:8086",
"def_ctx": "/", "def_ctx": "/",
"headless": false, "headless": false,
"sel_docker": false, "sel_docker": false,
@ -28,7 +28,7 @@
}, },
"nightly" : { "nightly" : {
"config_id": "nightly", "config_id": "nightly",
"influx_url": "http://localhost:9999", "influx_url": "http://localhost:8086",
"def_ctx": "/", "def_ctx": "/",
"headless": true, "headless": true,
"sel_docker": false, "sel_docker": false,

View File

@ -4,7 +4,7 @@ Feature: Load Data - Scrapers
As a user I want to Read Create Update and Delete Scrapers As a user I want to Read Create Update and Delete Scrapers
So that I can manage the stores used with Influxdbv2 So that I can manage the stores used with Influxdbv2
# Move exercise create scraper popup here # Move exercise create scraper popup here
# N.B. can verify scrapers at endpoint http://localhost:9999/api/v2/scrapers # N.B. can verify scrapers at endpoint http://localhost:8086/api/v2/scrapers
@tested @tested
@ -48,7 +48,7 @@ Scenario: Exercise create Scraper popup
Then the form element error message is "Target URL cannot be empty" Then the form element error message is "Target URL cannot be empty"
Then a form input error icon is shown Then a form input error icon is shown
Then the Create Scrapper popup create button is disabled Then the Create Scrapper popup create button is disabled
When enter the value "http://localhost:9999/metrics" into the Create Scraper popup url input When enter the value "http://localhost:8086/metrics" into the Create Scraper popup url input
Then the form element error message is not shown Then the form element error message is not shown
Then no form input error icon is shown Then no form input error icon is shown
Then the Create Scrapper popup create button is enabled Then the Create Scrapper popup create button is enabled
@ -74,8 +74,8 @@ Scenario Outline: Create Scrapers
Examples: Examples:
| NAME | ENDPOINT | BUCKET | | NAME | ENDPOINT | BUCKET |
| Melnik | http://localhost:9999/metrics | DEFAULT | | Melnik | http://localhost:8086/metrics | DEFAULT |
| Morlaix | http://localhost:9999/metrics | Duchamp | | Morlaix | http://localhost:8086/metrics | Duchamp |
| Brno | http://localhost:10018/bogus | DEFAULT | | Brno | http://localhost:10018/bogus | DEFAULT |
| Brest | http://localhost:10018/bogus | Duchamp | | Brest | http://localhost:10018/bogus | Duchamp |

View File

@ -222,5 +222,5 @@ Scenario Outline: Delete Telegraf Card
|Norimberk | |Norimberk |
|Rakovnik | |Rakovnik |
# N.B. can verify telegrafs at endpoint http://localhost:9999/api/v2/telegrafs # N.B. can verify telegrafs at endpoint http://localhost:8086/api/v2/telegrafs
# TODO - Test installation of telegraf and instructions - check back end # TODO - Test installation of telegraf and instructions - check back end

View File

@ -3,7 +3,7 @@
sleep 30 sleep 30
sudo netstat -tlnp sudo netstat -tlnp
curl -v --connect-timeout 60 --max-time 60 http://localhost:9999/debug/flush curl -v --connect-timeout 60 --max-time 60 http://localhost:8086/debug/flush
git clone https://github.com/influxdata/influxdb.git git clone https://github.com/influxdata/influxdb.git

View File

@ -5,7 +5,7 @@ TEST_CONTAINER=bonitoo_e2e
INFLUX2_CONTAINER=influx2_solo INFLUX2_CONTAINER=influx2_solo
E2E_MAP_DIR=/tmp/e2e E2E_MAP_DIR=/tmp/e2e
INFLUX2_HOST=$(sudo docker inspect -f "{{ .NetworkSettings.IPAddress }}" ${INFLUX2_CONTAINER}) INFLUX2_HOST=$(sudo docker inspect -f "{{ .NetworkSettings.IPAddress }}" ${INFLUX2_CONTAINER})
INFLUX2_URL="http://${INFLUX2_HOST}:9999" INFLUX2_URL="http://${INFLUX2_HOST}:8086"
#TAGS="@influx-influx" #TAGS="@influx-influx"
ACTIVE_CONF=development ACTIVE_CONF=development

View File

@ -42,8 +42,8 @@ pull_docker(){
run_docker_influx(){ run_docker_influx(){
mkdir -p ${LOG_DIR} mkdir -p ${LOG_DIR}
echo "["$(date +"%d.%m.%Y %T")"] starting docker instance ${INSTANCE_NAME}" echo "["$(date +"%d.%m.%Y %T")"] starting docker instance ${INSTANCE_NAME}"
sudo docker run --name ${INSTANCE_NAME} --publish 9999:9999 ${DOCKER_IMAGE} > ${LOG_FILE} 2>&1 & sudo docker run --name ${INSTANCE_NAME} --publish 8086:8086 ${DOCKER_IMAGE} > ${LOG_FILE} 2>&1 &
echo "["$(date +"%d.%m.%Y %T")"] started instance $INSTANCE_NAME listening at port 9999." echo "["$(date +"%d.%m.%Y %T")"] started instance $INSTANCE_NAME listening at port 8086."
echo "logfile at $LOG_FILE" echo "logfile at $LOG_FILE"
sleep 3 sleep 3
echo "\n$(tail -n32 $LOG_FILE)\n" echo "\n$(tail -n32 $LOG_FILE)\n"
@ -56,7 +56,7 @@ run_docker_influx_test_env(){
sudo docker rm $INSTANCE_NAME sudo docker rm $INSTANCE_NAME
sudo docker pull quay.io/influxdb/influx:nightly sudo docker pull quay.io/influxdb/influx:nightly
sudo docker build -t influxdb_test_image . sudo docker build -t influxdb_test_image .
sudo docker run --name $INSTANCE_NAME --publish 9999:9999 influxdb_test_image > ${LOG_FILE} 2>&1 & sudo docker run --name $INSTANCE_NAME --publish 8086:8086 influxdb_test_image > ${LOG_FILE} 2>&1 &
} }

View File

@ -507,7 +507,7 @@ const getAuthorizations = async(userName) => {
return await authsAPI.getAuthorizations(); return await authsAPI.getAuthorizations();
}; };
// http://localhost:9999/api/v2/labels // http://localhost:8086/api/v2/labels
// {"orgID":"8576cb897e0b4ce9","name":"MyLabel","properties":{"description":"","color":"#7CE490"}} // {"orgID":"8576cb897e0b4ce9","name":"MyLabel","properties":{"description":"","color":"#7CE490"}}
const createLabel = async(userName, const createLabel = async(userName,
labelName, labelName,

View File

@ -1,5 +1,5 @@
ping_cancelled=false # Keep track of whether the loop was cancelled, or succeeded ping_cancelled=false # Keep track of whether the loop was cancelled, or succeeded
until nc -z 127.0.0.1 9999; do :; done & until nc -z 127.0.0.1 8086; do :; done &
trap "kill $!; ping_cancelled=true" SIGINT trap "kill $!; ping_cancelled=true" SIGINT
wait $! # Wait for the loop to exit, one way or another wait $! # Wait for the loop to exit, one way or another
trap - INT # Remove the trap, now we're done with it trap - INT # Remove the trap, now we're done with it

View File

@ -1851,7 +1851,7 @@ func TestService_handlePostDashboardLabel(t *testing.T) {
t.Fatalf("failed to unmarshal label mapping: %v", err) t.Fatalf("failed to unmarshal label mapping: %v", err)
} }
url := fmt.Sprintf("http://localhost:9999/api/v2/dashboards/%s/labels", tt.args.dashboardID) url := fmt.Sprintf("http://localhost:8086/api/v2/dashboards/%s/labels", tt.args.dashboardID)
r := httptest.NewRequest("POST", url, bytes.NewReader(b)) r := httptest.NewRequest("POST", url, bytes.NewReader(b))
w := httptest.NewRecorder() w := httptest.NewRecorder()

View File

@ -86,7 +86,7 @@ func TestSessionHandler_handleSignin(t *testing.T) {
h := NewSessionHandler(zaptest.NewLogger(t), b) h := NewSessionHandler(zaptest.NewLogger(t), b)
w := httptest.NewRecorder() w := httptest.NewRecorder()
r := httptest.NewRequest("POST", "http://localhost:9999/api/v2/signin", nil) r := httptest.NewRequest("POST", "http://localhost:8086/api/v2/signin", nil)
r.SetBasicAuth(tt.args.user, tt.args.password) r.SetBasicAuth(tt.args.user, tt.args.password)
h.ServeHTTP(w, r) h.ServeHTTP(w, r)

View File

@ -1179,7 +1179,7 @@ func TestService_handlePostTaskLabel(t *testing.T) {
t.Fatalf("failed to unmarshal label mapping: %v", err) t.Fatalf("failed to unmarshal label mapping: %v", err)
} }
url := fmt.Sprintf("http://localhost:9999/api/v2/tasks/%s/labels", tt.args.taskID) url := fmt.Sprintf("http://localhost:8086/api/v2/tasks/%s/labels", tt.args.taskID)
r := httptest.NewRequest("POST", url, bytes.NewReader(b)) r := httptest.NewRequest("POST", url, bytes.NewReader(b))
w := httptest.NewRecorder() w := httptest.NewRecorder()
@ -1266,7 +1266,7 @@ func TestTaskHandler_CreateTaskWithOrgName(t *testing.T) {
const script = `option task = {name:"x", every:1m} from(bucket:"b-src") |> range(start:-1m) |> to(bucket:"b-dst", org:"o")` const script = `option task = {name:"x", every:1m} from(bucket:"b-src") |> range(start:-1m) |> to(bucket:"b-dst", org:"o")`
url := "http://localhost:9999/api/v2/tasks" url := "http://localhost:8086/api/v2/tasks"
b, err := json.Marshal(influxdb.TaskCreate{ b, err := json.Marshal(influxdb.TaskCreate{
Flux: script, Flux: script,
@ -1402,7 +1402,7 @@ func TestTaskHandler_Sessions(t *testing.T) {
} }
h := newHandler(t, ts) h := newHandler(t, ts)
url := fmt.Sprintf("http://localhost:9999/api/v2/tasks/%s/runs", taskID) url := fmt.Sprintf("http://localhost:8086/api/v2/tasks/%s/runs", taskID)
valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{{Key: "id", Value: taskID.String()}}) valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{{Key: "id", Value: taskID.String()}})
r := httptest.NewRequest("GET", url, nil).WithContext(valCtx) r := httptest.NewRequest("GET", url, nil).WithContext(valCtx)
w := httptest.NewRecorder() w := httptest.NewRecorder()
@ -1495,7 +1495,7 @@ func TestTaskHandler_Sessions(t *testing.T) {
} }
h := newHandler(t, ts) h := newHandler(t, ts)
url := fmt.Sprintf("http://localhost:9999/api/v2/tasks/%s/runs/%s", taskID, runID) url := fmt.Sprintf("http://localhost:8086/api/v2/tasks/%s/runs/%s", taskID, runID)
valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{ valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{
{Key: "id", Value: taskID.String()}, {Key: "id", Value: taskID.String()},
{Key: "rid", Value: runID.String()}, {Key: "rid", Value: runID.String()},
@ -1590,7 +1590,7 @@ func TestTaskHandler_Sessions(t *testing.T) {
} }
h := newHandler(t, ts) h := newHandler(t, ts)
url := fmt.Sprintf("http://localhost:9999/api/v2/tasks/%s/runs/%s/logs", taskID, runID) url := fmt.Sprintf("http://localhost:8086/api/v2/tasks/%s/runs/%s/logs", taskID, runID)
valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{ valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{
{Key: "id", Value: taskID.String()}, {Key: "id", Value: taskID.String()},
{Key: "rid", Value: runID.String()}, {Key: "rid", Value: runID.String()},
@ -1684,7 +1684,7 @@ func TestTaskHandler_Sessions(t *testing.T) {
} }
h := newHandler(t, ts) h := newHandler(t, ts)
url := fmt.Sprintf("http://localhost:9999/api/v2/tasks/%s/runs/%s/retry", taskID, runID) url := fmt.Sprintf("http://localhost:8086/api/v2/tasks/%s/runs/%s/retry", taskID, runID)
valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{ valCtx := context.WithValue(sessionAllPermsCtx, httprouter.ParamsKey, httprouter.Params{
{Key: "id", Value: taskID.String()}, {Key: "id", Value: taskID.String()},
{Key: "rid", Value: runID.String()}, {Key: "rid", Value: runID.String()},

View File

@ -351,8 +351,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"
@ -437,8 +437,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"
@ -527,8 +527,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"
@ -613,8 +613,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"
@ -705,8 +705,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"
@ -791,8 +791,8 @@ func TestTelegrafHandler_handleGetTelegraf(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "no_more_secrets" token = "no_more_secrets"

View File

@ -863,7 +863,7 @@ func TestService_handlePostVariableLabel(t *testing.T) {
t.Fatalf("failed to unmarshal label mapping: %v", err) t.Fatalf("failed to unmarshal label mapping: %v", err)
} }
url := fmt.Sprintf("http://localhost:9999/api/v2/variables/%s/labels", tt.args.variableID) url := fmt.Sprintf("http://localhost:8086/api/v2/variables/%s/labels", tt.args.variableID)
r := httptest.NewRequest("POST", url, bytes.NewReader(b)) r := httptest.NewRequest("POST", url, bytes.NewReader(b))
w := httptest.NewRecorder() w := httptest.NewRecorder()

View File

@ -317,7 +317,7 @@ func TestWriteHandler_handleWrite(t *testing.T) {
r := httptest.NewRequest( r := httptest.NewRequest(
"POST", "POST",
"http://localhost:9999/api/v2/write", "http://localhost:8086/api/v2/write",
strings.NewReader(tt.request.body), strings.NewReader(tt.request.body),
) )

View File

@ -85,7 +85,7 @@ spec:
[agent] [agent]
interval = "10s" interval = "10s"
[[outputs.influxdb_v2]] [[outputs.influxdb_v2]]
urls = ["http://localhost:9999"] urls = ["http://localhost:8086"]
token = "$INFLUX_TOKEN" token = "$INFLUX_TOKEN"
organization = "rg" organization = "rg"
bucket = "rucket_3" bucket = "rucket_3"

View File

@ -32,7 +32,7 @@
"name": "label-2" "name": "label-2"
} }
], ],
"config": "# Configuration for telegraf agent\n [agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n [[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:9999\n urls = [\"http://localhost:9999\"]\n\n ## Token for authentication.\n token = \"$INFLUX_TOKEN\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"rg\"\n\n ## Destination bucket to write into.\n bucket = \"rucket_3\"\n [[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n [[inputs.disk]]\n ## By default stats will be gathered for all mount points.\n ## Set mount_points will restrict the stats to only the specified mount points.\n # mount_points = [\"/\"]\n ## Ignore mount points by filesystem type.\n ignore_fs = [\"tmpfs\", \"devtmpfs\", \"devfs\", \"overlay\", \"aufs\", \"squashfs\"]\n [[inputs.diskio]]\n [[inputs.mem]]\n [[inputs.net]]\n [[inputs.processes]]\n [[inputs.swap]]\n [[inputs.system]]" "config": "# Configuration for telegraf agent\n [agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n [[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:8086\n urls = [\"http://localhost:8086\"]\n\n ## Token for authentication.\n token = \"$INFLUX_TOKEN\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"rg\"\n\n ## Destination bucket to write into.\n bucket = \"rucket_3\"\n [[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n [[inputs.disk]]\n ## By default stats will be gathered for all mount points.\n ## Set mount_points will restrict the stats to only the specified mount points.\n # mount_points = [\"/\"]\n ## Ignore mount points by filesystem type.\n ignore_fs = [\"tmpfs\", \"devtmpfs\", \"devfs\", \"overlay\", \"aufs\", \"squashfs\"]\n [[inputs.diskio]]\n [[inputs.mem]]\n [[inputs.net]]\n [[inputs.processes]]\n [[inputs.swap]]\n [[inputs.system]]"
} }
}, },
{ {
@ -42,7 +42,7 @@
"name": "tele-2" "name": "tele-2"
}, },
"spec": { "spec": {
"config": "# Configuration for telegraf agent\n [agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n [[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:9999\n urls = [\"http://localhost:9999\"]\n\n ## Token for authentication.\n token = \"$INFLUX_TOKEN\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"rg\"\n\n ## Destination bucket to write into.\n bucket = \"rucket_3\"\n [[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n [[inputs.disk]]\n ## By default stats will be gathered for all mount points.\n ## Set mount_points will restrict the stats to only the specified mount points.\n # mount_points = [\"/\"]\n ## Ignore mount points by filesystem type.\n ignore_fs = [\"tmpfs\", \"devtmpfs\", \"devfs\", \"overlay\", \"aufs\", \"squashfs\"]\n [[inputs.diskio]]\n [[inputs.mem]]\n [[inputs.net]]\n [[inputs.processes]]\n [[inputs.swap]]\n [[inputs.system]]" "config": "# Configuration for telegraf agent\n [agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n [[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:8086\n urls = [\"http://localhost:8086\"]\n\n ## Token for authentication.\n token = \"$INFLUX_TOKEN\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"rg\"\n\n ## Destination bucket to write into.\n bucket = \"rucket_3\"\n [[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n [[inputs.disk]]\n ## By default stats will be gathered for all mount points.\n ## Set mount_points will restrict the stats to only the specified mount points.\n # mount_points = [\"/\"]\n ## Ignore mount points by filesystem type.\n ignore_fs = [\"tmpfs\", \"devtmpfs\", \"devfs\", \"overlay\", \"aufs\", \"squashfs\"]\n [[inputs.diskio]]\n [[inputs.mem]]\n [[inputs.net]]\n [[inputs.processes]]\n [[inputs.swap]]\n [[inputs.system]]"
} }
} }
] ]

View File

@ -80,8 +80,8 @@ spec:
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://localhost:9999"] urls = ["http://localhost:8086"]
## Token for authentication. ## Token for authentication.
token = "$INFLUX_TOKEN" token = "$INFLUX_TOKEN"

View File

@ -32,7 +32,7 @@ func (i *InfluxDBV2) TOML() string {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = [%s] urls = [%s]
## Token for authentication. ## Token for authentication.

View File

@ -40,7 +40,7 @@ func TestTOML(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = [] urls = []
## Token for authentication. ## Token for authentication.
@ -68,8 +68,8 @@ func TestTOML(t *testing.T) {
`, `,
&InfluxDBV2{ &InfluxDBV2{
URLs: []string{ URLs: []string{
"http://192.168.1.10:9999", "http://192.168.1.10:8086",
"http://192.168.1.11:9999", "http://192.168.1.11:8086",
}, },
Token: "tok1", Token: "tok1",
Organization: "org1", Organization: "org1",
@ -79,8 +79,8 @@ func TestTOML(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://192.168.1.10:9999", "http://192.168.1.11:9999"] urls = ["http://192.168.1.10:8086", "http://192.168.1.11:8086"]
## Token for authentication. ## Token for authentication.
token = "tok1" token = "tok1"
@ -161,16 +161,16 @@ func TestDecodeTOML(t *testing.T) {
name: "influxdb_v2 missing token", name: "influxdb_v2 missing token",
want: &InfluxDBV2{ want: &InfluxDBV2{
URLs: []string{ URLs: []string{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
}, },
wantErr: errors.New("token is missing for influxdb_v2 output plugin"), wantErr: errors.New("token is missing for influxdb_v2 output plugin"),
output: &InfluxDBV2{}, output: &InfluxDBV2{},
data: map[string]interface{}{ data: map[string]interface{}{
"urls": []interface{}{ "urls": []interface{}{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
}, },
}, },
@ -178,8 +178,8 @@ func TestDecodeTOML(t *testing.T) {
name: "influxdb_v2 missing org", name: "influxdb_v2 missing org",
want: &InfluxDBV2{ want: &InfluxDBV2{
URLs: []string{ URLs: []string{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
Token: "token1", Token: "token1",
}, },
@ -187,8 +187,8 @@ func TestDecodeTOML(t *testing.T) {
output: &InfluxDBV2{}, output: &InfluxDBV2{},
data: map[string]interface{}{ data: map[string]interface{}{
"urls": []interface{}{ "urls": []interface{}{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
"token": "token1", "token": "token1",
}, },
@ -197,8 +197,8 @@ func TestDecodeTOML(t *testing.T) {
name: "influxdb_v2 missing bucket", name: "influxdb_v2 missing bucket",
want: &InfluxDBV2{ want: &InfluxDBV2{
URLs: []string{ URLs: []string{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
Token: "token1", Token: "token1",
Organization: "org1", Organization: "org1",
@ -207,8 +207,8 @@ func TestDecodeTOML(t *testing.T) {
output: &InfluxDBV2{}, output: &InfluxDBV2{},
data: map[string]interface{}{ data: map[string]interface{}{
"urls": []interface{}{ "urls": []interface{}{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
"token": "token1", "token": "token1",
"organization": "org1", "organization": "org1",
@ -218,8 +218,8 @@ func TestDecodeTOML(t *testing.T) {
name: "influxdb_v2", name: "influxdb_v2",
want: &InfluxDBV2{ want: &InfluxDBV2{
URLs: []string{ URLs: []string{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
Token: "token1", Token: "token1",
Organization: "org1", Organization: "org1",
@ -228,8 +228,8 @@ func TestDecodeTOML(t *testing.T) {
output: &InfluxDBV2{}, output: &InfluxDBV2{},
data: map[string]interface{}{ data: map[string]interface{}{
"urls": []interface{}{ "urls": []interface{}{
"http://localhost:9999", "http://localhost:8086",
"http://192.168.0.1:9999", "http://192.168.0.1:8086",
}, },
"token": "token1", "token": "token1",
"organization": "org1", "organization": "org1",

View File

@ -908,7 +908,7 @@ var availableInputs = `{
"type": "input", "type": "input",
"name": "httpjson", "name": "httpjson",
"description": "Read flattened metrics from one or more JSON HTTP endpoints", "description": "Read flattened metrics from one or more JSON HTTP endpoints",
"config": "# Read flattened metrics from one or more JSON HTTP endpoints\n[[inputs.httpjson]]\n # alias=\"httpjson\"\n ## NOTE This plugin only reads numerical measurements, strings and booleans\n ## will be ignored.\n\n ## Name for the service being polled. Will be appended to the name of the\n ## measurement e.g. httpjson_webserver_stats\n ##\n ## Deprecated (1.3.0): Use name_override, name_suffix, name_prefix instead.\n name = \"webserver_stats\"\n\n ## URL of each server in the service's cluster\n servers = [\n \"http://localhost:9999/stats/\",\n \"http://localhost:9998/stats/\",\n ]\n ## Set response_timeout (default 5 seconds)\n response_timeout = \"5s\"\n\n ## HTTP method to use: GET or POST (case-sensitive)\n method = \"GET\"\n\n ## List of tag names to extract from top-level of JSON server response\n # tag_keys = [\n # \"my_tag_1\",\n # \"my_tag_2\"\n # ]\n\n ## Optional TLS Config\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain \u0026 host verification\n # insecure_skip_verify = false\n\n ## HTTP parameters (all values must be strings). For \"GET\" requests, data\n ## will be included in the query. For \"POST\" requests, data will be included\n ## in the request body as \"x-www-form-urlencoded\".\n # [inputs.httpjson.parameters]\n # event_type = \"cpu_spike\"\n # threshold = \"0.75\"\n\n ## HTTP Headers (all values must be strings)\n # [inputs.httpjson.headers]\n # X-Auth-Token = \"my-xauth-token\"\n # apiVersion = \"v1\"\n\n" "config": "# Read flattened metrics from one or more JSON HTTP endpoints\n[[inputs.httpjson]]\n # alias=\"httpjson\"\n ## NOTE This plugin only reads numerical measurements, strings and booleans\n ## will be ignored.\n\n ## Name for the service being polled. Will be appended to the name of the\n ## measurement e.g. httpjson_webserver_stats\n ##\n ## Deprecated (1.3.0): Use name_override, name_suffix, name_prefix instead.\n name = \"webserver_stats\"\n\n ## URL of each server in the service's cluster\n servers = [\n \"http://localhost:8086/stats/\",\n \"http://localhost:9998/stats/\",\n ]\n ## Set response_timeout (default 5 seconds)\n response_timeout = \"5s\"\n\n ## HTTP method to use: GET or POST (case-sensitive)\n method = \"GET\"\n\n ## List of tag names to extract from top-level of JSON server response\n # tag_keys = [\n # \"my_tag_1\",\n # \"my_tag_2\"\n # ]\n\n ## Optional TLS Config\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain \u0026 host verification\n # insecure_skip_verify = false\n\n ## HTTP parameters (all values must be strings). For \"GET\" requests, data\n ## will be included in the query. For \"POST\" requests, data will be included\n ## in the request body as \"x-www-form-urlencoded\".\n # [inputs.httpjson.parameters]\n # event_type = \"cpu_spike\"\n # threshold = \"0.75\"\n\n ## HTTP Headers (all values must be strings)\n # [inputs.httpjson.headers]\n # X-Auth-Token = \"my-xauth-token\"\n # apiVersion = \"v1\"\n\n"
}, },
{ {
"type": "input", "type": "input",
@ -1040,7 +1040,7 @@ var availableInputs = `{
"type": "input", "type": "input",
"name": "jolokia2_proxy", "name": "jolokia2_proxy",
"description": "Read JMX metrics from a Jolokia REST proxy endpoint", "description": "Read JMX metrics from a Jolokia REST proxy endpoint",
"config": "# Read JMX metrics from a Jolokia REST proxy endpoint\n[[inputs.jolokia2_proxy]]\n # alias=\"jolokia2_proxy\"\n # default_tag_prefix = \"\"\n # default_field_prefix = \"\"\n # default_field_separator = \".\"\n\n ## Proxy agent\n url = \"http://localhost:8080/jolokia\"\n # username = \"\"\n # password = \"\"\n # response_timeout = \"5s\"\n\n ## Optional TLS config\n # tls_ca = \"/var/private/ca.pem\"\n # tls_cert = \"/var/private/client.pem\"\n # tls_key = \"/var/private/client-key.pem\"\n # insecure_skip_verify = false\n\n ## Add proxy targets to query\n # default_target_username = \"\"\n # default_target_password = \"\"\n [[inputs.jolokia2_proxy.target]]\n url = \"service:jmx:rmi:///jndi/rmi://targethost:9999/jmxrmi\"\n # username = \"\"\n # password = \"\"\n\n ## Add metrics to read\n [[inputs.jolokia2_proxy.metric]]\n name = \"java_runtime\"\n mbean = \"java.lang:type=Runtime\"\n paths = [\"Uptime\"]\n\n" "config": "# Read JMX metrics from a Jolokia REST proxy endpoint\n[[inputs.jolokia2_proxy]]\n # alias=\"jolokia2_proxy\"\n # default_tag_prefix = \"\"\n # default_field_prefix = \"\"\n # default_field_separator = \".\"\n\n ## Proxy agent\n url = \"http://localhost:8080/jolokia\"\n # username = \"\"\n # password = \"\"\n # response_timeout = \"5s\"\n\n ## Optional TLS config\n # tls_ca = \"/var/private/ca.pem\"\n # tls_cert = \"/var/private/client.pem\"\n # tls_key = \"/var/private/client-key.pem\"\n # insecure_skip_verify = false\n\n ## Add proxy targets to query\n # default_target_username = \"\"\n # default_target_password = \"\"\n [[inputs.jolokia2_proxy.target]]\n url = \"service:jmx:rmi:///jndi/rmi://targethost:8086/jmxrmi\"\n # username = \"\"\n # password = \"\"\n\n ## Add metrics to read\n [[inputs.jolokia2_proxy.metric]]\n name = \"java_runtime\"\n mbean = \"java.lang:type=Runtime\"\n paths = [\"Uptime\"]\n\n"
}, },
{ {
"type": "input", "type": "input",
@ -1437,7 +1437,7 @@ var availableOutputs = `{
"type": "output", "type": "output",
"name": "influxdb_v2", "name": "influxdb_v2",
"description": "Configuration for sending metrics to InfluxDB", "description": "Configuration for sending metrics to InfluxDB",
"config": "# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:9999\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain \u0026 host verification\n # insecure_skip_verify = false\n\n" "config": "# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:8086\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain \u0026 host verification\n # insecure_skip_verify = false\n\n"
}, },
{ {
"type": "output", "type": "output",

View File

@ -63,7 +63,7 @@ func TestTelegrafConfigJSONDecodeWithoutID(t *testing.T) {
"comment": "3", "comment": "3",
"config": { "config": {
"urls": [ "urls": [
"http://127.0.0.1:9999" "http://127.0.0.1:8086"
], ],
"token": "token1", "token": "token1",
"organization": "org", "organization": "org",
@ -147,8 +147,8 @@ func TestTelegrafConfigJSONDecodeWithoutID(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "token1" token = "token1"
@ -175,7 +175,7 @@ func TestTelegrafConfigJSONDecodeWithoutID(t *testing.T) {
func TestTelegrafConfigJSONDecodeTOML(t *testing.T) { func TestTelegrafConfigJSONDecodeTOML(t *testing.T) {
s := `{ s := `{
"name": "config 2", "name": "config 2",
"config": "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n[[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n[[inputs.kernel]]\n[[inputs.kubernetes]]\n ## URL for the kubelet\n ## exp: http://1.1.1.1:10255\n url = \"http://1.1.1.1:12\"\n[[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:9999\n urls = [\"http://127.0.0.1:9999\"]\n\n ## Token for authentication.\n token = \"token1\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"org\"\n\n ## Destination bucket to write into.\n bucket = \"bucket\"\n" "config": "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n[[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n[[inputs.kernel]]\n[[inputs.kubernetes]]\n ## URL for the kubelet\n ## exp: http://1.1.1.1:10255\n url = \"http://1.1.1.1:12\"\n[[outputs.influxdb_v2]]\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:8086\n urls = [\"http://127.0.0.1:8086\"]\n\n ## Token for authentication.\n token = \"token1\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"org\"\n\n ## Destination bucket to write into.\n bucket = \"bucket\"\n"
}` }`
want := &TelegrafConfig{ want := &TelegrafConfig{
@ -253,8 +253,8 @@ func TestTelegrafConfigJSONDecodeTOML(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["http://127.0.0.1:9999"] urls = ["http://127.0.0.1:8086"]
## Token for authentication. ## Token for authentication.
token = "token1" token = "token1"
@ -294,7 +294,7 @@ func TestTelegrafConfigJSONCompatibleMode(t *testing.T) {
cfg: &TelegrafConfig{ cfg: &TelegrafConfig{
ID: *id1, ID: *id1,
OrgID: *id2, OrgID: *id2,
Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:9999\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n", Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:8086\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n",
}, },
expMeta: map[string]interface{}{"buckets": []string{}}, expMeta: map[string]interface{}{"buckets": []string{}},
}, },
@ -304,7 +304,7 @@ func TestTelegrafConfigJSONCompatibleMode(t *testing.T) {
cfg: &TelegrafConfig{ cfg: &TelegrafConfig{
ID: *id1, ID: *id1,
OrgID: *id2, OrgID: *id2,
Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:9999\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n", Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:8086\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n",
}, },
expMeta: map[string]interface{}{"buckets": []string{}}, expMeta: map[string]interface{}{"buckets": []string{}},
}, },
@ -314,7 +314,7 @@ func TestTelegrafConfigJSONCompatibleMode(t *testing.T) {
cfg: &TelegrafConfig{ cfg: &TelegrafConfig{
ID: *id1, ID: *id1,
OrgID: *id3, OrgID: *id3,
Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:9999\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n", Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n# Configuration for sending metrics to InfluxDB\n[[outputs.influxdb_v2]]\n # alias=\"influxdb_v2\"\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## ex: urls = [\"https://us-west-2-1.aws.cloud2.influxdata.com\"]\n urls = [\"http://127.0.0.1:8086\"]\n\n ## Token for authentication.\n token = \"\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n\n ## The value of this tag will be used to determine the bucket. If this\n ## tag is not set the 'bucket' option is used as the default.\n # bucket_tag = \"\"\n\n ## If true, the bucket tag will not be added to the metric.\n # exclude_bucket_tag = false\n\n ## Timeout for HTTP messages.\n # timeout = \"5s\"\n\n ## Additional HTTP headers\n # http_headers = {\"X-Special-Header\" = \"Special-Value\"}\n\n ## HTTP Proxy override, if unset values the standard proxy environment\n ## variables are consulted to determine which proxy, if any, should be used.\n # http_proxy = \"http://corporate.proxy:3128\"\n\n ## HTTP User-Agent\n # user_agent = \"telegraf\"\n\n ## Content-Encoding for write request body, can be set to \"gzip\" to\n ## compress body or \"identity\" to apply no encoding.\n # content_encoding = \"gzip\"\n\n ## Enable or disable uint support for writing uints influxdb 2.0.\n # influx_uint_support = false\n\n ## Optional TLS Config for use on HTTP connections.\n # tls_ca = \"/etc/telegraf/ca.pem\"\n # tls_cert = \"/etc/telegraf/cert.pem\"\n # tls_key = \"/etc/telegraf/key.pem\"\n ## Use TLS but skip chain & host verification\n # insecure_skip_verify = false\n\n",
}, },
expMeta: map[string]interface{}{"buckets": []string{}}, expMeta: map[string]interface{}{"buckets": []string{}},
}, },
@ -396,7 +396,7 @@ func TestTelegrafConfigJSON(t *testing.T) {
ID: *id1, ID: *id1,
OrgID: *id2, OrgID: *id2,
Name: "n1", Name: "n1",
Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n[[inputs.file]]\t\n ## Files to parse each interval.\n ## These accept standard unix glob matching rules, but with the addition of\n ## ** as a \"super asterisk\". ie:\n ## /var/log/**.log -> recursively find all .log files in /var/log\n ## /var/log/*/*.log -> find all .log files with a parent dir in /var/log\n ## /var/log/apache.log -> only read the apache log file\n files = [\"f1\", \"f2\"]\n\n ## The dataformat to be read from files\n ## Each data format has its own unique set of configuration options, read\n ## more about them here:\n ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md\n data_format = \"influx\"\n[[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n[[outputs.file]]\n ## Files to write to, \"stdout\" is a specially handled file.\n files = [\"stdout\"]\n[[outputs.influxdb_v2]]\t\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:9999\n urls = [\"url1\", \"url2\"]\n\n ## Token for authentication.\n token = \"tok1\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n", Config: "# Configuration for telegraf agent\n[agent]\n ## Default data collection interval for all inputs\n interval = \"10s\"\n ## Rounds collection interval to 'interval'\n ## ie, if interval=\"10s\" then always collect on :00, :10, :20, etc.\n round_interval = true\n\n ## Telegraf will send metrics to outputs in batches of at most\n ## metric_batch_size metrics.\n ## This controls the size of writes that Telegraf sends to output plugins.\n metric_batch_size = 1000\n\n ## For failed writes, telegraf will cache metric_buffer_limit metrics for each\n ## output, and will flush this buffer on a successful write. Oldest metrics\n ## are dropped first when this buffer fills.\n ## This buffer only fills when writes fail to output plugin(s).\n metric_buffer_limit = 10000\n\n ## Collection jitter is used to jitter the collection by a random amount.\n ## Each plugin will sleep for a random time within jitter before collecting.\n ## This can be used to avoid many plugins querying things like sysfs at the\n ## same time, which can have a measurable effect on the system.\n collection_jitter = \"0s\"\n\n ## Default flushing interval for all outputs. Maximum flush_interval will be\n ## flush_interval + flush_jitter\n flush_interval = \"10s\"\n ## Jitter the flush interval by a random amount. This is primarily to avoid\n ## large write spikes for users running a large number of telegraf instances.\n ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s\n flush_jitter = \"0s\"\n\n ## By default or when set to \"0s\", precision will be set to the same\n ## timestamp order as the collection interval, with the maximum being 1s.\n ## ie, when interval = \"10s\", precision will be \"1s\"\n ## when interval = \"250ms\", precision will be \"1ms\"\n ## Precision will NOT be used for service inputs. It is up to each individual\n ## service input to set the timestamp at the appropriate precision.\n ## Valid time units are \"ns\", \"us\" (or \"µs\"), \"ms\", \"s\".\n precision = \"\"\n\n ## Logging configuration:\n ## Run telegraf with debug log messages.\n debug = false\n ## Run telegraf in quiet mode (error log messages only).\n quiet = false\n ## Specify the log file name. The empty string means to log to stderr.\n logfile = \"\"\n\n ## Override default hostname, if empty use os.Hostname()\n hostname = \"\"\n ## If set to true, do no set the \"host\" tag in the telegraf agent.\n omit_hostname = false\n[[inputs.file]]\t\n ## Files to parse each interval.\n ## These accept standard unix glob matching rules, but with the addition of\n ## ** as a \"super asterisk\". ie:\n ## /var/log/**.log -> recursively find all .log files in /var/log\n ## /var/log/*/*.log -> find all .log files with a parent dir in /var/log\n ## /var/log/apache.log -> only read the apache log file\n files = [\"f1\", \"f2\"]\n\n ## The dataformat to be read from files\n ## Each data format has its own unique set of configuration options, read\n ## more about them here:\n ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md\n data_format = \"influx\"\n[[inputs.cpu]]\n ## Whether to report per-cpu stats or not\n percpu = true\n ## Whether to report total system cpu stats or not\n totalcpu = true\n ## If true, collect raw CPU time metrics.\n collect_cpu_time = false\n ## If true, compute and report the sum of all non-idle CPU states.\n report_active = false\n[[outputs.file]]\n ## Files to write to, \"stdout\" is a specially handled file.\n files = [\"stdout\"]\n[[outputs.influxdb_v2]]\t\n ## The URLs of the InfluxDB cluster nodes.\n ##\n ## Multiple URLs can be specified for a single cluster, only ONE of the\n ## urls will be written to each interval.\n ## urls exp: http://127.0.0.1:8086\n urls = [\"url1\", \"url2\"]\n\n ## Token for authentication.\n token = \"tok1\"\n\n ## Organization is the name of the organization you wish to write to; must exist.\n organization = \"\"\n\n ## Destination bucket to write into.\n bucket = \"\"\n",
Metadata: map[string]interface{}{"buckets": []string{}}, Metadata: map[string]interface{}{"buckets": []string{}},
}, },
}, },
@ -591,7 +591,7 @@ func TestLegacyStruct(t *testing.T) {
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["url1", "url2"] urls = ["url1", "url2"]
## Token for authentication. ## Token for authentication.

View File

@ -282,10 +282,10 @@ describe('Collectors', () => {
}) })
it('can add and delete urls', () => { it('can add and delete urls', () => {
cy.getByTestID('input-field').type('http://localhost:9999') cy.getByTestID('input-field').type('http://localhost:8086')
cy.contains('Add').click() cy.contains('Add').click()
cy.contains('http://localhost:9999').should('exist', () => { cy.contains('http://localhost:8086').should('exist', () => {
cy.getByTestID('input-field').type('http://example.com') cy.getByTestID('input-field').type('http://example.com')
cy.contains('Add').click() cy.contains('Add').click()

View File

@ -418,7 +418,7 @@ export const influxDB2Plugin = {
type: 'output', type: 'output',
comment: 'write to influxdb v2', comment: 'write to influxdb v2',
config: { config: {
urls: ['http://127.0.0.1:9999'], urls: ['http://127.0.0.1:8086'],
token, token,
organization: 'default', organization: 'default',
bucket: 'defbuck', bucket: 'defbuck',
@ -636,7 +636,7 @@ export const setSetupParamsResponse = {
status: 201, status: 201,
statusText: 'Created', statusText: 'Created',
headers: { headers: {
'access-control-allow-origin': 'http://localhost:9999', 'access-control-allow-origin': 'http://localhost:8086',
date: 'Fri, 11 Jan 2019 22:49:33 GMT', date: 'Fri, 11 Jan 2019 22:49:33 GMT',
'access-control-allow-headers': 'access-control-allow-headers':
'Accept, Content-Type, Content-Length, Accept-Encoding, Authorization', 'Accept, Content-Type, Content-Length, Accept-Encoding, Authorization',
@ -685,7 +685,7 @@ export const scraperTargets = [
orgID: '03636a0aabb51000', orgID: '03636a0aabb51000',
organization: 'a', organization: 'a',
type: 'prometheus', type: 'prometheus',
url: 'http://localhost:9999/metrics', url: 'http://localhost:8086/metrics',
}, },
{ {
bucket: 'a', bucket: 'a',
@ -695,7 +695,7 @@ export const scraperTargets = [
orgID: '03636a0aabb51000', orgID: '03636a0aabb51000',
organization: 'a', organization: 'a',
type: 'prometheus', type: 'prometheus',
url: 'http://localhost:9999/metrics', url: 'http://localhost:8086/metrics',
}, },
] ]

View File

@ -29,7 +29,7 @@
"test:watch": "jest --watch --verbose false", "test:watch": "jest --watch --verbose false",
"test:update": "jest --updateSnapshot", "test:update": "jest --updateSnapshot",
"test:debug": "node --inspect-brk $(npm bin)/jest --runInBand --watch --verbose false", "test:debug": "node --inspect-brk $(npm bin)/jest --runInBand --watch --verbose false",
"test:e2e": "CYPRESS_baseUrl=http://localhost:9999 cypress run --browser chrome --reporter junit --reporter-options 'mochaFile=junit-results/test-output-[hash].xml'", "test:e2e": "CYPRESS_baseUrl=http://localhost:8086 cypress run --browser chrome --reporter junit --reporter-options 'mochaFile=junit-results/test-output-[hash].xml'",
"test:e2e:report": "junit-viewer --results=junit-results --save-file=cypress/site/junit-report.html", "test:e2e:report": "junit-viewer --results=junit-results --save-file=cypress/site/junit-report.html",
"test:e2e:clean": "rm junit-results/*.xml", "test:e2e:clean": "rm junit-results/*.xml",
"test:e2e:all": "yarn test:e2e:clean && yarn test:e2e; yarn test:e2e:report;", "test:e2e:all": "yarn test:e2e:clean && yarn test:e2e; yarn test:e2e:report;",
@ -43,7 +43,7 @@
"prettier:fix": "prettier --config .prettierrc.json --write '{src,cypress}/**/*.{ts,tsx}'", "prettier:fix": "prettier --config .prettierrc.json --write '{src,cypress}/**/*.{ts,tsx}'",
"tsc": "tsc -p ./tsconfig.json --noEmit --pretty --skipLibCheck", "tsc": "tsc -p ./tsconfig.json --noEmit --pretty --skipLibCheck",
"tsc:cypress": "tsc -p ./cypress/tsconfig.json --noEmit --pretty --skipLibCheck", "tsc:cypress": "tsc -p ./cypress/tsconfig.json --noEmit --pretty --skipLibCheck",
"cy": "CYPRESS_baseUrl=http://localhost:9999 cypress open", "cy": "CYPRESS_baseUrl=http://localhost:8086 cypress open",
"cy:dev": "CYPRESS_baseUrl=http://localhost:8080 cypress open", "cy:dev": "CYPRESS_baseUrl=http://localhost:8080 cypress open",
"generate": "oats ../http/swagger.yml > ./src/client/generatedRoutes.ts" "generate": "oats ../http/swagger.yml > ./src/client/generatedRoutes.ts"
}, },

View File

@ -37,7 +37,7 @@ const TELEGRAF_OUTPUT = ` [[outputs.influxdb_v2]]
## ##
## Multiple URLs can be specified for a single cluster, only ONE of the ## Multiple URLs can be specified for a single cluster, only ONE of the
## urls will be written to each interval. ## urls will be written to each interval.
## urls exp: http://127.0.0.1:9999 ## urls exp: http://127.0.0.1:8086
urls = ["<%= server %>"] urls = ["<%= server %>"]
## Token for authentication. ## Token for authentication.
@ -51,7 +51,7 @@ const TELEGRAF_OUTPUT = ` [[outputs.influxdb_v2]]
` `
const OUTPUT_DEFAULTS = { const OUTPUT_DEFAULTS = {
server: 'http://127.0.0.1:9999', server: 'http://127.0.0.1:8086',
token: '$INFLUX_TOKEN', token: '$INFLUX_TOKEN',
org: 'orgID', org: 'orgID',
bucket: 'bucketID', bucket: 'bucketID',

View File

@ -19,7 +19,7 @@ Deprecated (1.6): use the [http](../http) input.
## URL of each server in the service's cluster ## URL of each server in the service's cluster
servers = [ servers = [
"http://localhost:9999/stats/", "http://localhost:8086/stats/",
"http://localhost:9998/stats/", "http://localhost:9998/stats/",
] ]
## Set response_timeout (default 5 seconds) ## Set response_timeout (default 5 seconds)
@ -89,7 +89,7 @@ Given the following response body:
``` ```
The following metric is produced: The following metric is produced:
`httpjson,server=http://localhost:9999/stats/ b_d=0.1,a=0.5,b_e=5,response_time=0.001` `httpjson,server=http://localhost:8086/stats/ b_d=0.1,a=0.5,b_e=5,response_time=0.001`
Note that only numerical values are extracted and the type is float. Note that only numerical values are extracted and the type is float.
@ -102,7 +102,7 @@ If `tag_keys` is included in the configuration:
Then the `service` tag will also be added: Then the `service` tag will also be added:
`httpjson,server=http://localhost:9999/stats/,service=service01 b_d=0.1,a=0.5,b_e=5,response_time=0.001` `httpjson,server=http://localhost:8086/stats/,service=service01 b_d=0.1,a=0.5,b_e=5,response_time=0.001`
**Array Output:** **Array Output:**
@ -131,5 +131,5 @@ If the service returns an array of objects, one metric is be created for each ob
] ]
``` ```
`httpjson,server=http://localhost:9999/stats/,service=service01 a=0.5,b_d=0.1,b_e=5,response_time=0.003` `httpjson,server=http://localhost:8086/stats/,service=service01 a=0.5,b_d=0.1,b_e=5,response_time=0.003`
`httpjson,server=http://localhost:9999/stats/,service=service02 a=0.6,b_d=0.2,b_e=6,response_time=0.003` `httpjson,server=http://localhost:8086/stats/,service=service02 a=0.6,b_d=0.2,b_e=6,response_time=0.003`

View File

@ -27,9 +27,9 @@ module.exports = merge(common, {
}, },
compress: true, compress: true,
proxy: { proxy: {
'/api/v2': 'http://localhost:9999', '/api/v2': 'http://localhost:8086',
'/debug/flush': 'http://localhost:9999', '/debug/flush': 'http://localhost:8086',
'/oauth': 'http://localhost:9999', '/oauth': 'http://localhost:8086',
}, },
disableHostCheck: true, disableHostCheck: true,
host: '0.0.0.0', host: '0.0.0.0',

View File

@ -37,7 +37,7 @@ Once the vault and influxdb servers have been started and initialized, you may t
```sh ```sh
curl --request GET \ curl --request GET \
--url http://localhost:9999/api/v2/orgs/<org id>/secrets \ --url http://localhost:8086/api/v2/orgs/<org id>/secrets \
--header 'authorization: Token <authorization token> --header 'authorization: Token <authorization token>
# should return # should return
@ -53,7 +53,7 @@ curl --request GET \
```sh ```sh
curl --request PATCH \ curl --request PATCH \
--url http://localhost:9999/api/v2/orgs/<org id>/secrets \ --url http://localhost:8086/api/v2/orgs/<org id>/secrets \
--header 'authorization: Token <authorization token> \ --header 'authorization: Token <authorization token> \
--header 'content-type: application/json' \ --header 'content-type: application/json' \
--data '{ --data '{
@ -66,7 +66,7 @@ curl --request PATCH \
```sh ```sh
curl --request GET \ curl --request GET \
--url http://localhost:9999/api/v2/orgs/<org id>/secrets \ --url http://localhost:8086/api/v2/orgs/<org id>/secrets \
--header 'authorization: Token <authorization token> --header 'authorization: Token <authorization token>
# should return # should return