GoogleCloudMonitoring: Refactor types (#58644)
This commit is contained in:
committed by
GitHub
parent
16af756d50
commit
64143ea7d0
@@ -50,31 +50,60 @@ func TestNewInstanceSettings(t *testing.T) {
|
||||
func TestCloudMonitoring(t *testing.T) {
|
||||
service := &Service{}
|
||||
|
||||
t.Run("parses a time series list query", func(t *testing.T) {
|
||||
req := baseTimeSeriesList()
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
require.Len(t, queries, 1)
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].params.Encode())
|
||||
assert.Equal(t, 7, len(queries[0].params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].params["view"][0])
|
||||
assert.Equal(t, "testalias", queries[0].aliasBy)
|
||||
})
|
||||
|
||||
t.Run("parses a time series query", func(t *testing.T) {
|
||||
req := baseTimeSeriesQuery()
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueryFromInterface(t, qes)
|
||||
|
||||
require.Len(t, queries, 1)
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "foo", queries[0].parameters.Query)
|
||||
assert.Equal(t, "testalias", queries[0].aliasBy)
|
||||
})
|
||||
|
||||
t.Run("Parse migrated queries from frontend and build Google Cloud Monitoring API queries", func(t *testing.T) {
|
||||
t.Run("and query has no aggregation set", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
err := migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
require.Len(t, queries, 1)
|
||||
assert.Equal(t, "A", queries[0].RefID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].Target)
|
||||
assert.Equal(t, 7, len(queries[0].Params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].Params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].Params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].Params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].Params["view"][0])
|
||||
assert.Equal(t, "testalias", queries[0].AliasBy)
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].params.Encode())
|
||||
assert.Equal(t, 7, len(queries[0].params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].params["view"][0])
|
||||
assert.Equal(t, "testalias", queries[0].aliasBy)
|
||||
|
||||
t.Run("and generated deep link has correct parameters", func(t *testing.T) {
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -91,7 +120,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and query has filters", func(t *testing.T) {
|
||||
query := baseReq()
|
||||
query := deprecatedReq()
|
||||
query.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"filters": ["key", "=", "value", "AND", "key2", "=", "value2", "AND", "resource.type", "=", "another/resource/type"]
|
||||
@@ -101,14 +130,14 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, query)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, `metric.type="a/metric/type" key="value" key2="value2" resource.type="another/resource/type"`, queries[0].Params["filter"][0])
|
||||
assert.Equal(t, `metric.type="a/metric/type" key="value" key2="value2" resource.type="another/resource/type"`, queries[0].params["filter"][0])
|
||||
|
||||
// assign a resource type to query parameters
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
// the deep link should not contain this resource type since another resource type is included in the query filters
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -124,7 +153,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
t.Run("and alignmentPeriod is set to grafana-auto", func(t *testing.T) {
|
||||
t.Run("and IntervalMS is larger than 60000", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].Interval = 1000000 * time.Millisecond
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"alignmentPeriod": "grafana-auto",
|
||||
@@ -135,12 +164,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+1000s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+1000s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -154,7 +183,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
verifyDeepLink(t, dl, expectedTimeSelection, expectedTimeSeriesFilter)
|
||||
})
|
||||
t.Run("and IntervalMS is less than 60000", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].Interval = 30000 * time.Millisecond
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"alignmentPeriod": "grafana-auto",
|
||||
@@ -165,12 +194,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -189,7 +218,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
now := time.Now().UTC()
|
||||
|
||||
t.Run("and range is two hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 2))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -201,12 +230,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
})
|
||||
|
||||
t.Run("and range is 22 hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 22))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -218,12 +247,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
})
|
||||
|
||||
t.Run("and range is 23 hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 23))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -235,12 +264,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+300s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+300s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
})
|
||||
|
||||
t.Run("and range is 7 days", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now
|
||||
req.Queries[0].TimeRange.To = now.AddDate(0, 0, 7)
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -252,8 +281,8 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+3600s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+3600s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
})
|
||||
})
|
||||
|
||||
@@ -261,7 +290,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
now := time.Now().UTC()
|
||||
|
||||
t.Run("and range is two hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 2))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -273,12 +302,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -293,7 +322,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and range is 22 hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 22))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -305,12 +334,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -325,7 +354,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and range is 23 hours", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.Add(-(time.Hour * 23))
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -337,12 +366,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+300s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+300s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -357,7 +386,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and range is 7 days", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].TimeRange.From = now.AddDate(0, 0, -7)
|
||||
req.Queries[0].TimeRange.To = now
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
@@ -369,12 +398,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+3600s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+3600s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -391,7 +420,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
t.Run("and alignmentPeriod is set in frontend", func(t *testing.T) {
|
||||
t.Run("and alignment period is within accepted range", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].Interval = 1000
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"alignmentPeriod": "+600s"
|
||||
@@ -401,12 +430,12 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `+600s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `+600s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -422,7 +451,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and query has aggregation mean set", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_SUM",
|
||||
@@ -433,23 +462,23 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "A", queries[0].RefID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_SUM&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].Target)
|
||||
assert.Equal(t, 7, len(queries[0].Params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].Params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].Params["interval.endTime"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].Params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].Params["view"][0])
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_SUM&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].params.Encode())
|
||||
assert.Equal(t, 7, len(queries[0].params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].params["view"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -467,7 +496,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and query has group bys", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_NONE",
|
||||
@@ -479,23 +508,23 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "A", queries[0].RefID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.groupByFields=metric.label.group1&aggregation.groupByFields=metric.label.group2&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].Target)
|
||||
assert.Equal(t, 8, len(queries[0].Params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].Params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].Params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.label.group1", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "metric.label.group2", queries[0].Params["aggregation.groupByFields"][1])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].Params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].Params["view"][0])
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.groupByFields=metric.label.group1&aggregation.groupByFields=metric.label.group2&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].params.Encode())
|
||||
assert.Equal(t, 8, len(queries[0].params))
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.label.group1", queries[0].params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "metric.label.group2", queries[0].params["aggregation.groupByFields"][1])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].params["view"][0])
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -536,28 +565,31 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
},
|
||||
},
|
||||
}
|
||||
err := migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
|
||||
t.Run("and query type is metrics", func(t *testing.T) {
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "A", queries[0].RefID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.groupByFields=metric.label.group1&aggregation.groupByFields=metric.label.group2&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].Target)
|
||||
assert.Equal(t, 8, len(queries[0].Params))
|
||||
assert.Equal(t, "metric.label.group1", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "metric.label.group2", queries[0].Params["aggregation.groupByFields"][1])
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].Params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].Params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].Params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].Params["view"][0])
|
||||
assert.Equal(t, "testalias", queries[0].AliasBy)
|
||||
assert.Equal(t, []string{"metric.label.group1", "metric.label.group2"}, queries[0].GroupBys)
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "aggregation.alignmentPeriod=%2B60s&aggregation.crossSeriesReducer=REDUCE_NONE&aggregation.groupByFields=metric.label.group1&aggregation.groupByFields=metric.label.group2&aggregation.perSeriesAligner=ALIGN_MEAN&filter=metric.type%3D%22a%2Fmetric%2Ftype%22&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z&view=FULL", queries[0].params.Encode())
|
||||
assert.Equal(t, 8, len(queries[0].params))
|
||||
assert.Equal(t, "metric.label.group1", queries[0].params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "metric.label.group2", queries[0].params["aggregation.groupByFields"][1])
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "metric.type=\"a/metric/type\"", queries[0].params["filter"][0])
|
||||
assert.Equal(t, "FULL", queries[0].params["view"][0])
|
||||
assert.Equal(t, "testalias", queries[0].aliasBy)
|
||||
assert.Equal(t, []string{"metric.label.group1", "metric.label.group2"}, queries[0].parameters.GroupBys)
|
||||
|
||||
// assign resource type to query parameters to be included in the deep link filter
|
||||
// in the actual workflow this information comes from the response of the Monitoring API
|
||||
queries[0].Params.Set("resourceType", "a/resource/type")
|
||||
queries[0].params.Set("resourceType", "a/resource/type")
|
||||
dl := queries[0].buildDeepLink()
|
||||
|
||||
expectedTimeSelection := map[string]string{
|
||||
@@ -583,21 +615,18 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
},
|
||||
"sloQuery": {}
|
||||
}`)
|
||||
err = migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
|
||||
qes, err = service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
tqueries := make([]*cloudMonitoringTimeSeriesQuery, 0)
|
||||
for _, qi := range qes {
|
||||
q, ok := qi.(*cloudMonitoringTimeSeriesQuery)
|
||||
assert.True(t, ok)
|
||||
tqueries = append(tqueries, q)
|
||||
}
|
||||
|
||||
tqueries := getCloudMonitoringQueryFromInterface(t, qes)
|
||||
assert.Equal(t, 1, len(tqueries))
|
||||
assert.Equal(t, "A", tqueries[0].RefID)
|
||||
assert.Equal(t, "test-proj", tqueries[0].ProjectName)
|
||||
assert.Equal(t, "test-query", tqueries[0].Query)
|
||||
assert.Equal(t, "test-alias", tqueries[0].AliasBy)
|
||||
assert.Equal(t, "A", tqueries[0].refID)
|
||||
assert.Equal(t, "test-proj", tqueries[0].parameters.ProjectName)
|
||||
assert.Equal(t, "test-query", tqueries[0].parameters.Query)
|
||||
assert.Equal(t, "test-alias", tqueries[0].aliasBy)
|
||||
})
|
||||
|
||||
t.Run("and query type is SLOs", func(t *testing.T) {
|
||||
@@ -614,20 +643,22 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
},
|
||||
"metricQuery": {}
|
||||
}`)
|
||||
err := migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "A", queries[0].RefID)
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].Params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].Params["interval.endTime"][0])
|
||||
assert.Equal(t, `+60s`, queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "", queries[0].AliasBy)
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, `aggregation.alignmentPeriod=%2B60s&aggregation.perSeriesAligner=ALIGN_MEAN&filter=select_slo_health%28%22projects%2Ftest-proj%2Fservices%2Ftest-service%2FserviceLevelObjectives%2Ftest-slo%22%29&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z`, queries[0].Target)
|
||||
assert.Equal(t, 5, len(queries[0].Params))
|
||||
assert.Equal(t, "A", queries[0].refID)
|
||||
assert.Equal(t, "2018-03-15T13:00:00Z", queries[0].params["interval.startTime"][0])
|
||||
assert.Equal(t, "2018-03-15T13:34:00Z", queries[0].params["interval.endTime"][0])
|
||||
assert.Equal(t, `+60s`, queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "", queries[0].aliasBy)
|
||||
assert.Equal(t, "ALIGN_MEAN", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, `aggregation.alignmentPeriod=%2B60s&aggregation.perSeriesAligner=ALIGN_MEAN&filter=select_slo_health%28%22projects%2Ftest-proj%2Fservices%2Ftest-service%2FserviceLevelObjectives%2Ftest-slo%22%29&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z`, queries[0].params.Encode())
|
||||
assert.Equal(t, 5, len(queries[0].params))
|
||||
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"queryType": "slo",
|
||||
@@ -642,11 +673,13 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
},
|
||||
"metricQuery": {}
|
||||
}`)
|
||||
err = migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
|
||||
qes, err = service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
qqueries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, "ALIGN_NEXT_OLDER", qqueries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
qqueries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, "ALIGN_NEXT_OLDER", qqueries[0].params["aggregation.perSeriesAligner"][0])
|
||||
|
||||
dl := qqueries[0].buildDeepLink()
|
||||
assert.Empty(t, dl)
|
||||
@@ -665,11 +698,13 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
},
|
||||
"metricQuery": {}
|
||||
}`)
|
||||
err = migrateRequest(req)
|
||||
require.NoError(t, err)
|
||||
|
||||
qes, err = service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
qqqueries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
assert.Equal(t, `aggregation.alignmentPeriod=%2B60s&aggregation.perSeriesAligner=ALIGN_NEXT_OLDER&filter=select_slo_burn_rate%28%22projects%2Ftest-proj%2Fservices%2Ftest-service%2FserviceLevelObjectives%2Ftest-slo%22%2C+%221h%22%29&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z`, qqqueries[0].Target)
|
||||
qqqueries := getCloudMonitoringListFromInterface(t, qes)
|
||||
assert.Equal(t, `aggregation.alignmentPeriod=%2B60s&aggregation.perSeriesAligner=ALIGN_NEXT_OLDER&filter=select_slo_burn_rate%28%22projects%2Ftest-proj%2Fservices%2Ftest-service%2FserviceLevelObjectives%2Ftest-slo%22%2C+%221h%22%29&interval.endTime=2018-03-15T13%3A34%3A00Z&interval.startTime=2018-03-15T13%3A00%3A00Z`, qqqueries[0].params.Encode())
|
||||
})
|
||||
})
|
||||
|
||||
@@ -761,7 +796,7 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is not defined", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_MIN",
|
||||
@@ -775,22 +810,22 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["aggregation.groupByFields"][0])
|
||||
|
||||
assert.NotContains(t, queries[0].Params, "secondaryAggregation.crossSeriesReducer")
|
||||
assert.NotContains(t, "REDUCE_SUM", queries[0].Params, "secondaryAggregation.perSeriesAligner")
|
||||
assert.NotContains(t, "+60s", queries[0].Params, "secondaryAggregation.alignmentPeriod")
|
||||
assert.NotContains(t, "labelname", queries[0].Params, "secondaryAggregation.groupByFields")
|
||||
assert.NotContains(t, queries[0].params, "secondaryAggregation.crossSeriesReducer")
|
||||
assert.NotContains(t, "REDUCE_SUM", queries[0].params, "secondaryAggregation.perSeriesAligner")
|
||||
assert.NotContains(t, "+60s", queries[0].params, "secondaryAggregation.alignmentPeriod")
|
||||
assert.NotContains(t, "labelname", queries[0].params, "secondaryAggregation.groupByFields")
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is set to none", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_MIN",
|
||||
@@ -805,22 +840,22 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["aggregation.groupByFields"][0])
|
||||
|
||||
assert.NotContains(t, queries[0].Params, "secondaryAggregation.crossSeriesReducer")
|
||||
assert.NotContains(t, "REDUCE_SUM", queries[0].Params, "secondaryAggregation.perSeriesAligner")
|
||||
assert.NotContains(t, "+60s", queries[0].Params, "secondaryAggregation.alignmentPeriod")
|
||||
assert.NotContains(t, "labelname", queries[0].Params, "secondaryAggregation.groupByFields")
|
||||
assert.NotContains(t, queries[0].params, "secondaryAggregation.crossSeriesReducer")
|
||||
assert.NotContains(t, "REDUCE_SUM", queries[0].params, "secondaryAggregation.perSeriesAligner")
|
||||
assert.NotContains(t, "+60s", queries[0].params, "secondaryAggregation.alignmentPeriod")
|
||||
assert.NotContains(t, "labelname", queries[0].params, "secondaryAggregation.groupByFields")
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is set to rate and there's no group bys", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_SUM",
|
||||
@@ -835,20 +870,20 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_NONE", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_RATE", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "REDUCE_NONE", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_RATE", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["secondaryAggregation.alignmentPeriod"][0])
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is set to rate and group bys exist", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_SUM",
|
||||
@@ -863,22 +898,22 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_RATE", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_RATE", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["aggregation.groupByFields"][0])
|
||||
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["secondaryAggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["secondaryAggregation.groupByFields"][0])
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is set to delta and there's no group bys", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_MIN",
|
||||
@@ -893,20 +928,20 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_NONE", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_DELTA", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "REDUCE_NONE", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_DELTA", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["secondaryAggregation.alignmentPeriod"][0])
|
||||
})
|
||||
|
||||
t.Run("and query preprocessor is set to delta and group bys exist", func(t *testing.T) {
|
||||
req := baseReq()
|
||||
req := deprecatedReq()
|
||||
req.Queries[0].JSON = json.RawMessage(`{
|
||||
"metricType": "a/metric/type",
|
||||
"crossSeriesReducer": "REDUCE_MIN",
|
||||
@@ -921,27 +956,39 @@ func TestCloudMonitoring(t *testing.T) {
|
||||
|
||||
qes, err := service.buildQueryExecutors(slog, req)
|
||||
require.NoError(t, err)
|
||||
queries := getCloudMonitoringQueriesFromInterface(t, qes)
|
||||
queries := getCloudMonitoringListFromInterface(t, qes)
|
||||
|
||||
assert.Equal(t, 1, len(queries))
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_DELTA", queries[0].Params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["aggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["aggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "ALIGN_DELTA", queries[0].params["aggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["aggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["aggregation.groupByFields"][0])
|
||||
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].Params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].Params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].Params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].Params["secondaryAggregation.groupByFields"][0])
|
||||
assert.Equal(t, "REDUCE_MIN", queries[0].params["secondaryAggregation.crossSeriesReducer"][0])
|
||||
assert.Equal(t, "REDUCE_SUM", queries[0].params["secondaryAggregation.perSeriesAligner"][0])
|
||||
assert.Equal(t, "+60s", queries[0].params["secondaryAggregation.alignmentPeriod"][0])
|
||||
assert.Equal(t, "labelname", queries[0].params["secondaryAggregation.groupByFields"][0])
|
||||
})
|
||||
}
|
||||
|
||||
func getCloudMonitoringQueriesFromInterface(t *testing.T, qes []cloudMonitoringQueryExecutor) []*cloudMonitoringTimeSeriesFilter {
|
||||
func getCloudMonitoringListFromInterface(t *testing.T, qes []cloudMonitoringQueryExecutor) []*cloudMonitoringTimeSeriesList {
|
||||
t.Helper()
|
||||
|
||||
queries := make([]*cloudMonitoringTimeSeriesFilter, 0)
|
||||
queries := make([]*cloudMonitoringTimeSeriesList, 0)
|
||||
for _, qi := range qes {
|
||||
q, ok := qi.(*cloudMonitoringTimeSeriesFilter)
|
||||
q, ok := qi.(*cloudMonitoringTimeSeriesList)
|
||||
require.Truef(t, ok, "Received wrong type %T", qi)
|
||||
queries = append(queries, q)
|
||||
}
|
||||
return queries
|
||||
}
|
||||
|
||||
func getCloudMonitoringQueryFromInterface(t *testing.T, qes []cloudMonitoringQueryExecutor) []*cloudMonitoringTimeSeriesQuery {
|
||||
t.Helper()
|
||||
|
||||
queries := make([]*cloudMonitoringTimeSeriesQuery, 0)
|
||||
for _, qi := range qes {
|
||||
q, ok := qi.(*cloudMonitoringTimeSeriesQuery)
|
||||
require.Truef(t, ok, "Received wrong type %T", qi)
|
||||
queries = append(queries, q)
|
||||
}
|
||||
@@ -1009,7 +1056,7 @@ func verifyDeepLink(t *testing.T, dl string, expectedTimeSelection map[string]st
|
||||
}
|
||||
}
|
||||
|
||||
func baseReq() *backend.QueryDataRequest {
|
||||
func deprecatedReq() *backend.QueryDataRequest {
|
||||
fromStart := time.Date(2018, 3, 15, 13, 0, 0, 0, time.UTC).In(time.Local)
|
||||
query := &backend.QueryDataRequest{
|
||||
Queries: []backend.DataQuery{
|
||||
@@ -1031,6 +1078,54 @@ func baseReq() *backend.QueryDataRequest {
|
||||
return query
|
||||
}
|
||||
|
||||
func baseTimeSeriesList() *backend.QueryDataRequest {
|
||||
fromStart := time.Date(2018, 3, 15, 13, 0, 0, 0, time.UTC).In(time.Local)
|
||||
query := &backend.QueryDataRequest{
|
||||
Queries: []backend.DataQuery{
|
||||
{
|
||||
RefID: "A",
|
||||
TimeRange: backend.TimeRange{
|
||||
From: fromStart,
|
||||
To: fromStart.Add(34 * time.Minute),
|
||||
},
|
||||
QueryType: "metrics",
|
||||
JSON: json.RawMessage(`{
|
||||
"timeSeriesList": {
|
||||
"metricType": "a/metric/type",
|
||||
"view": "FULL"
|
||||
},
|
||||
"aliasBy": "testalias"
|
||||
}`),
|
||||
},
|
||||
},
|
||||
}
|
||||
return query
|
||||
}
|
||||
|
||||
func baseTimeSeriesQuery() *backend.QueryDataRequest {
|
||||
fromStart := time.Date(2018, 3, 15, 13, 0, 0, 0, time.UTC).In(time.Local)
|
||||
query := &backend.QueryDataRequest{
|
||||
Queries: []backend.DataQuery{
|
||||
{
|
||||
RefID: "A",
|
||||
TimeRange: backend.TimeRange{
|
||||
From: fromStart,
|
||||
To: fromStart.Add(34 * time.Minute),
|
||||
},
|
||||
QueryType: "metrics",
|
||||
JSON: json.RawMessage(`{
|
||||
"queryType": "metrics",
|
||||
"timeSeriesQuery": {
|
||||
"query": "foo"
|
||||
},
|
||||
"aliasBy": "testalias"
|
||||
}`),
|
||||
},
|
||||
},
|
||||
}
|
||||
return query
|
||||
}
|
||||
|
||||
func TestCheckHealth(t *testing.T) {
|
||||
t.Run("and using GCE authentation should return proper error", func(t *testing.T) {
|
||||
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
||||
|
||||
Reference in New Issue
Block a user