* fix(graph-time-query): updated graphs so that graph x-axis reflects the input timeRange
* fix(dashboard-timerange): dashboard timerange is now outputting the correct timerange
* fix(ui/dashboard): connected the timeMachine to the component to make data consistently available based on timeMachine context
* fix(query): updated implementation to reduce mucking the timeMachine reducer logic
* fix(edit-veo): configuring a timerange in the dashboard should be accessible when editing cells
* fix(wip): wip
* fix(graph-time-query): updated graphs so that graph x-axis reflects the input timeRange
* fix(dashboard-timerange): dashboard timerange is now outputting the correct timerange
* fix(ui/dashboard): connected the timeMachine to the component to make data consistently available based on timeMachine context
* fix(query): updated implementation to reduce mucking the timeMachine reducer logic
* fix(edit-veo): configuring a timerange in the dashboard should be accessible when editing cells
* feat(ui): Type start and end time functions
* fix(ui): Set VEO timeMachine timeRange from parent dashboard
* feat(ui): Regularize getStartTime and getEndTime
* feat(ui): Fix prettier errors
* fix(ui): added unit tests for critical parts of the function
* fix(ui): removed unnecessary console that was failing jstest
* fix(ui): added getStartTime and getEndTime tests:
* fix(ui): updated telegraf description to be more explicit
* fix(ui): updated title to display same styling as empty text
* fix(ui): added defaultProps and checks for collectors
* fix(endpoint): when looking up a endpoint we should allow org only lookup
In the current system the api always adds "UserID" to the filter. This only
allows the system to look up endpoints that user created. The behavior should be
that we filter based on user input and use authorizor to hide things they shouldn't see.
This adds an lru cache for the columns that are produced as tags. When
producing the columns that are part of the group key, it will generate
the column and then keep it in an lru cache to reuse for future tables.
The start and stop column are effectively cached for every table because
they are special and will be the same for all of the tables.
For the tags, it retains the most recently used since they may be used
by a future table. That way most of the columns will get shared with
each other.
When the size differs, a slice is used so the underlying data is still
shared, but the size is different.
* feat(tracing): dont trace spans with full URL path names in ExtractFromHTTPRequest
* chore(multiple): replace all occurrences of julienschmidt/httprouter with influxdata/httprouter
This removes the duplicate filter that is used by the reader. The
storage engine shouldn't be sending us duplicate tables anyway and this
code hurts performance in high cardinality queries because of the memory
it uses to keep track of all of the keys that have been seen.
A KV store may optionally implement a key predicate
function to filter on keys or values, which may reduce memory and
CPU usage.
Expected FindByTaskID resource mapping improvements for a single user:
Before:
206966085 ns/op 37672164 B/op 445060 allocs/op
After:
1514118 ns/op 11184 B/op 131 allocs/op
the errors changed during the pkger http server error improvements and this
fixes it to be similar to what it once was, only now its in a flatter fashion
very similar to the pkger http server apply response.
also renames a the commands to start with cmd followed by the args. this
is an attempt to order the commands in a slightly more autocomplete friendly
way.