254577ba56
* Lattice: Point to private prerelease of aws-sdk-go (#515) * point to private prerelease of aws-sdk-go * fix build issue * Lattice: Adding a feature toggle (#549) * Adding a feature toggle for lattice * Change name of feature toggle * Lattice: List accounts (#543) * Separate layers * Introduce testify/mock library Co-authored-by: Shirley Leu <4163034+fridgepoet@users.noreply.github.com> * point to version that includes metric api changes (#574) * add accounts component (#575) * Test refactor: remove unneeded clientFactoryMock (#581) * Lattice: Add monitoring badge (#576) * add monitoring badge * fix tests * solve conflict * Lattice: Add dynamic label for account display name (#579) * Build: Automatically sync lattice-main with OSS * Lattice: Point to private prerelease of aws-sdk-go (#515) * point to private prerelease of aws-sdk-go * fix build issue * Lattice: Adding a feature toggle (#549) * Adding a feature toggle for lattice * Change name of feature toggle * Lattice: List accounts (#543) * Separate layers * Introduce testify/mock library Co-authored-by: Shirley Leu <4163034+fridgepoet@users.noreply.github.com> * point to version that includes metric api changes (#574) * add accounts component (#575) * Test refactor: remove unneeded clientFactoryMock (#581) * Lattice: Add monitoring badge (#576) * add monitoring badge * fix tests * solve conflict * add account label Co-authored-by: Shirley Leu <4163034+fridgepoet@users.noreply.github.com> Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * fix import * solve merge related problem * add account info (#608) * add back namespaces handler * Lattice: Parse account id and return it to frontend (#609) * parse account id and return to frontend * fix route test * only show badge when feature toggle is enabled (#615) * Lattice: Refactor resource response type and return account (#613) * refactor resource response type * remove not used file. * go lint * fix tests * remove commented code * Lattice: Use account as input when listing metric names and dimensions (#611) * use account in resource requests * add account to response * revert accountInfo to accountId * PR feedback * unit test account in list metrics response * remove not used asserts * don't assert on response that is not relevant to the test * removed dupe test * pr feedback * rename request package (#626) * Lattice: Move account component and add tooltip (#630) * move accounts component to the top of metric stat editor * add tooltip * CloudWatch: add account to GetMetricData queries (#627) * Add AccountId to metric stat query * Lattice: Account variable support (#625) * add variable support in accounts component * add account variable query type * update variables * interpolate variable before its sent to backend * handle variable change in hooks * remove not used import * Update public/app/plugins/datasource/cloudwatch/components/Account.tsx Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * Update public/app/plugins/datasource/cloudwatch/hooks.ts Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * add one more unit test Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * cleanup (#629) * Set account Id according to crossAccountQuerying feature flag in backend (#632) * CloudWatch: Change spelling of feature-toggle (#634) * Lattice Logs (#631) * Lattice Logs * Fixes after CR * Lattice: Bug: fix dimension keys request (#644) * fix dimension keys * fix lint * more lint * CloudWatch: Add tests for QueryData with AccountId (#637) * Update from breaking change (#645) * Update from breaking change * Remove extra interface and methods Co-authored-by: Shirley Leu <4163034+fridgepoet@users.noreply.github.com> * CloudWatch: Add business logic layer for getting log groups (#642) Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * Lattice: Fix - unset account id in region change handler (#646) * move reset of account to region change handler * fix broken test * Lattice: Add account id to metric stat query deep link (#656) add account id to metric stat link * CloudWatch: Add new log groups handler for cross-account querying (#643) * Lattice: Add feature tracking (#660) * add tracking for account id prescense in metrics query * also check feature toggle * fix broken test * CloudWatch: Add route for DescribeLogGroups for cross-account querying (#647) Co-authored-by: Erik Sundell <erik.sundell87@gmail.com> * Lattice: Handle account id default value (#662) * make sure right type is returned * set right default values * Suggestions to lattice changes (#663) * Change ListMetricsWithPageLimit response to slice of non-pointers * Change GetAccountsForCurrentUserOrRole response to be not pointer * Clean test Cleanup calls in test * Remove CloudWatchAPI as part of mock * Resolve conflicts * Add Latest SDK (#672) * add tooltip (#674) * Docs: Add documentation for CloudWatch cross account querying (#676) * wip docs * change wordings * add sections about metrics and logs * change from monitoring to observability * Update docs/sources/datasources/aws-cloudwatch/_index.md Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * Update docs/sources/datasources/aws-cloudwatch/query-editor/index.md Co-authored-by: Fiona Artiaga <89225282+GrafanaWriter@users.noreply.github.com> * Update docs/sources/datasources/aws-cloudwatch/query-editor/index.md Co-authored-by: Fiona Artiaga <89225282+GrafanaWriter@users.noreply.github.com> * Update docs/sources/datasources/aws-cloudwatch/query-editor/index.md Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> * Update docs/sources/datasources/aws-cloudwatch/query-editor/index.md Co-authored-by: Fiona Artiaga <89225282+GrafanaWriter@users.noreply.github.com> * apply pr feedback * fix file name * more pr feedback * pr feedback Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> Co-authored-by: Fiona Artiaga <89225282+GrafanaWriter@users.noreply.github.com> * use latest version of the aws-sdk-go * Fix tests' mock response type * Remove change in Azure Monitor Co-authored-by: Sarah Zinger <sarah.zinger@grafana.com> Co-authored-by: Shirley Leu <4163034+fridgepoet@users.noreply.github.com> Co-authored-by: Fiona Artiaga <89225282+GrafanaWriter@users.noreply.github.com>
657 lines
19 KiB
Go
657 lines
19 KiB
Go
package cloudwatch
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/aws/aws-sdk-go/aws"
|
|
"github.com/aws/aws-sdk-go/aws/session"
|
|
"github.com/aws/aws-sdk-go/service/cloudwatchlogs"
|
|
"github.com/aws/aws-sdk-go/service/cloudwatchlogs/cloudwatchlogsiface"
|
|
"github.com/grafana/grafana-plugin-sdk-go/backend"
|
|
"github.com/grafana/grafana-plugin-sdk-go/backend/datasource"
|
|
"github.com/grafana/grafana-plugin-sdk-go/backend/instancemgmt"
|
|
"github.com/grafana/grafana-plugin-sdk-go/data"
|
|
"github.com/grafana/grafana/pkg/services/featuremgmt"
|
|
"github.com/grafana/grafana/pkg/tsdb/cloudwatch/models"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestQuery_GetLogEvents(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
const refID = "A"
|
|
|
|
testCases := map[string]struct {
|
|
query string
|
|
expectedInput []*cloudwatchlogs.GetLogEventsInput
|
|
}{
|
|
"Nil startTime": {
|
|
query: `{
|
|
"type": "logAction",
|
|
"subtype": "GetLogEvents",
|
|
"logGroupName": "foo",
|
|
"logStreamName": "bar",
|
|
"endTime": 1,
|
|
"startFromHead": false
|
|
}`,
|
|
expectedInput: []*cloudwatchlogs.GetLogEventsInput{
|
|
{
|
|
EndTime: aws.Int64(1),
|
|
Limit: aws.Int64(10),
|
|
LogGroupName: aws.String("foo"),
|
|
LogStreamName: aws.String("bar"),
|
|
StartFromHead: aws.Bool(false),
|
|
},
|
|
},
|
|
},
|
|
"Nil endTime": {
|
|
query: `{
|
|
"type": "logAction",
|
|
"subtype": "GetLogEvents",
|
|
"logGroupName": "foo",
|
|
"logStreamName": "bar",
|
|
"startTime": 1,
|
|
"startFromHead": true
|
|
}`,
|
|
expectedInput: []*cloudwatchlogs.GetLogEventsInput{
|
|
{
|
|
StartTime: aws.Int64(1),
|
|
Limit: aws.Int64(10),
|
|
LogGroupName: aws.String("foo"),
|
|
LogStreamName: aws.String("bar"),
|
|
StartFromHead: aws.Bool(true),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
for name, test := range testCases {
|
|
t.Run(name, func(t *testing.T) {
|
|
cli = fakeCWLogsClient{}
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: refID,
|
|
TimeRange: backend.TimeRange{From: time.Unix(0, 0), To: time.Unix(1, 0)},
|
|
JSON: json.RawMessage(test.query),
|
|
},
|
|
},
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
require.Len(t, cli.calls.getEventsWithContext, 1)
|
|
assert.Equal(t, test.expectedInput, cli.calls.getEventsWithContext)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestQuery_GetLogGroupFields(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
|
|
cli = fakeCWLogsClient{
|
|
logGroupFields: cloudwatchlogs.GetLogGroupFieldsOutput{
|
|
LogGroupFields: []*cloudwatchlogs.LogGroupField{
|
|
{
|
|
Name: aws.String("field_a"),
|
|
Percent: aws.Int64(100),
|
|
},
|
|
{
|
|
Name: aws.String("field_b"),
|
|
Percent: aws.Int64(30),
|
|
},
|
|
{
|
|
Name: aws.String("field_c"),
|
|
Percent: aws.Int64(55),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
const refID = "A"
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
resp, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: refID,
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "GetLogGroupFields",
|
|
"logGroupName": "group_a",
|
|
"limit": 50
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
require.NoError(t, err)
|
|
require.NotNil(t, resp)
|
|
|
|
expFrame := &data.Frame{
|
|
Name: refID,
|
|
Fields: []*data.Field{
|
|
data.NewField("name", nil, []*string{
|
|
aws.String("field_a"), aws.String("field_b"), aws.String("field_c"),
|
|
}),
|
|
data.NewField("percent", nil, []*int64{
|
|
aws.Int64(100), aws.Int64(30), aws.Int64(55),
|
|
}),
|
|
},
|
|
}
|
|
expFrame.RefID = refID
|
|
assert.Equal(t, &backend.QueryDataResponse{Responses: backend.Responses{
|
|
refID: backend.DataResponse{
|
|
Frames: data.Frames{expFrame},
|
|
},
|
|
},
|
|
}, resp)
|
|
}
|
|
|
|
func TestQuery_StartQuery(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
|
|
t.Run("invalid time range", func(t *testing.T) {
|
|
cli = fakeCWLogsClient{
|
|
logGroupFields: cloudwatchlogs.GetLogGroupFieldsOutput{
|
|
LogGroupFields: []*cloudwatchlogs.LogGroupField{
|
|
{
|
|
Name: aws.String("field_a"),
|
|
Percent: aws.Int64(100),
|
|
},
|
|
{
|
|
Name: aws.String("field_b"),
|
|
Percent: aws.Int64(30),
|
|
},
|
|
{
|
|
Name: aws.String("field_c"),
|
|
Percent: aws.Int64(55),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
timeRange := backend.TimeRange{
|
|
From: time.Unix(1584873443, 0),
|
|
To: time.Unix(1584700643, 0),
|
|
}
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
TimeRange: timeRange,
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery",
|
|
"limit": 50,
|
|
"region": "default",
|
|
"queryString": "fields @message"
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
require.Error(t, err)
|
|
|
|
assert.Contains(t, err.Error(), "invalid time range: start time must be before end time")
|
|
})
|
|
|
|
t.Run("valid time range", func(t *testing.T) {
|
|
const refID = "A"
|
|
cli = fakeCWLogsClient{
|
|
logGroupFields: cloudwatchlogs.GetLogGroupFieldsOutput{
|
|
LogGroupFields: []*cloudwatchlogs.LogGroupField{
|
|
{
|
|
Name: aws.String("field_a"),
|
|
Percent: aws.Int64(100),
|
|
},
|
|
{
|
|
Name: aws.String("field_b"),
|
|
Percent: aws.Int64(30),
|
|
},
|
|
{
|
|
Name: aws.String("field_c"),
|
|
Percent: aws.Int64(55),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
timeRange := backend.TimeRange{
|
|
From: time.Unix(1584700643000, 0),
|
|
To: time.Unix(1584873443000, 0),
|
|
}
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
resp, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: refID,
|
|
TimeRange: timeRange,
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery",
|
|
"limit": 50,
|
|
"region": "default",
|
|
"queryString": "fields @message"
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
expFrame := data.NewFrame(
|
|
refID,
|
|
data.NewField("queryId", nil, []string{"abcd-efgh-ijkl-mnop"}),
|
|
)
|
|
expFrame.RefID = refID
|
|
expFrame.Meta = &data.FrameMeta{
|
|
Custom: map[string]interface{}{
|
|
"Region": "default",
|
|
},
|
|
}
|
|
assert.Equal(t, &backend.QueryDataResponse{Responses: backend.Responses{
|
|
refID: {
|
|
Frames: data.Frames{expFrame},
|
|
},
|
|
},
|
|
}, resp)
|
|
})
|
|
}
|
|
|
|
func Test_executeStartQuery(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
|
|
t.Run("successfully parses information from JSON to StartQueryWithContext", func(t *testing.T) {
|
|
cli = fakeCWLogsClient{}
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{}},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: "A",
|
|
TimeRange: backend.TimeRange{From: time.Unix(0, 0), To: time.Unix(1, 0)},
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery",
|
|
"limit": 12,
|
|
"queryString":"fields @message",
|
|
"logGroupNames":["some name","another name"]
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, []*cloudwatchlogs.StartQueryInput{
|
|
{
|
|
StartTime: aws.Int64(0),
|
|
EndTime: aws.Int64(1),
|
|
Limit: aws.Int64(12),
|
|
QueryString: aws.String("fields @timestamp,ltrim(@log) as __log__grafana_internal__,ltrim(@logStream) as __logstream__grafana_internal__|fields @message"),
|
|
LogGroupNames: []*string{aws.String("some name"), aws.String("another name")},
|
|
},
|
|
}, cli.calls.startQueryWithContext)
|
|
})
|
|
|
|
t.Run("does not populate StartQueryInput.limit when no limit provided", func(t *testing.T) {
|
|
cli = fakeCWLogsClient{}
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{}},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: "A",
|
|
TimeRange: backend.TimeRange{From: time.Unix(0, 0), To: time.Unix(1, 0)},
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery"
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
require.Len(t, cli.calls.startQueryWithContext, 1)
|
|
assert.Nil(t, cli.calls.startQueryWithContext[0].Limit)
|
|
})
|
|
|
|
t.Run("attaches logGroupIdentifiers if the crossAccount feature is enabled", func(t *testing.T) {
|
|
cli = fakeCWLogsClient{}
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures(featuremgmt.FlagCloudWatchCrossAccountQuerying))
|
|
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{}},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: "A",
|
|
TimeRange: backend.TimeRange{From: time.Unix(0, 0), To: time.Unix(1, 0)},
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery",
|
|
"limit": 12,
|
|
"queryString":"fields @message",
|
|
"logGroups":[{"value": "fakeARN"}]
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, []*cloudwatchlogs.StartQueryInput{
|
|
{
|
|
StartTime: aws.Int64(0),
|
|
EndTime: aws.Int64(1),
|
|
Limit: aws.Int64(12),
|
|
QueryString: aws.String("fields @timestamp,ltrim(@log) as __log__grafana_internal__,ltrim(@logStream) as __logstream__grafana_internal__|fields @message"),
|
|
LogGroupIdentifiers: []*string{aws.String("fakeARN")},
|
|
},
|
|
}, cli.calls.startQueryWithContext)
|
|
})
|
|
|
|
t.Run("attaches logGroupIdentifiers if the crossAccount feature is enabled and strips out trailing *", func(t *testing.T) {
|
|
cli = fakeCWLogsClient{}
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures(featuremgmt.FlagCloudWatchCrossAccountQuerying))
|
|
|
|
_, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{}},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: "A",
|
|
TimeRange: backend.TimeRange{From: time.Unix(0, 0), To: time.Unix(1, 0)},
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StartQuery",
|
|
"limit": 12,
|
|
"queryString":"fields @message",
|
|
"logGroups":[{"value": "*fake**ARN*"}]
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, []*cloudwatchlogs.StartQueryInput{
|
|
{
|
|
StartTime: aws.Int64(0),
|
|
EndTime: aws.Int64(1),
|
|
Limit: aws.Int64(12),
|
|
QueryString: aws.String("fields @timestamp,ltrim(@log) as __log__grafana_internal__,ltrim(@logStream) as __logstream__grafana_internal__|fields @message"),
|
|
LogGroupIdentifiers: []*string{aws.String("*fake**ARN")},
|
|
},
|
|
}, cli.calls.startQueryWithContext)
|
|
})
|
|
}
|
|
|
|
func TestQuery_StopQuery(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
|
|
cli = fakeCWLogsClient{
|
|
logGroupFields: cloudwatchlogs.GetLogGroupFieldsOutput{
|
|
LogGroupFields: []*cloudwatchlogs.LogGroupField{
|
|
{
|
|
Name: aws.String("field_a"),
|
|
Percent: aws.Int64(100),
|
|
},
|
|
{
|
|
Name: aws.String("field_b"),
|
|
Percent: aws.Int64(30),
|
|
},
|
|
{
|
|
Name: aws.String("field_c"),
|
|
Percent: aws.Int64(55),
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
timeRange := backend.TimeRange{
|
|
From: time.Unix(1584873443, 0),
|
|
To: time.Unix(1584700643, 0),
|
|
}
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
resp, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
TimeRange: timeRange,
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "StopQuery",
|
|
"queryId": "abcd-efgh-ijkl-mnop"
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
expFrame := &data.Frame{
|
|
Name: "StopQueryResponse",
|
|
Fields: []*data.Field{
|
|
data.NewField("success", nil, []bool{true}),
|
|
},
|
|
}
|
|
assert.Equal(t, &backend.QueryDataResponse{Responses: backend.Responses{
|
|
"": {
|
|
Frames: data.Frames{expFrame},
|
|
},
|
|
},
|
|
}, resp)
|
|
}
|
|
|
|
func TestQuery_GetQueryResults(t *testing.T) {
|
|
origNewCWLogsClient := NewCWLogsClient
|
|
t.Cleanup(func() {
|
|
NewCWLogsClient = origNewCWLogsClient
|
|
})
|
|
|
|
var cli fakeCWLogsClient
|
|
|
|
NewCWLogsClient = func(sess *session.Session) cloudwatchlogsiface.CloudWatchLogsAPI {
|
|
return &cli
|
|
}
|
|
|
|
const refID = "A"
|
|
cli = fakeCWLogsClient{
|
|
queryResults: cloudwatchlogs.GetQueryResultsOutput{
|
|
Results: [][]*cloudwatchlogs.ResultField{
|
|
{
|
|
{
|
|
Field: aws.String("@timestamp"),
|
|
Value: aws.String("2020-03-20 10:37:23.000"),
|
|
},
|
|
{
|
|
Field: aws.String("field_b"),
|
|
Value: aws.String("b_1"),
|
|
},
|
|
{
|
|
Field: aws.String("@ptr"),
|
|
Value: aws.String("abcdefg"),
|
|
},
|
|
},
|
|
{
|
|
{
|
|
Field: aws.String("@timestamp"),
|
|
Value: aws.String("2020-03-20 10:40:43.000"),
|
|
},
|
|
{
|
|
Field: aws.String("field_b"),
|
|
Value: aws.String("b_2"),
|
|
},
|
|
{
|
|
Field: aws.String("@ptr"),
|
|
Value: aws.String("hijklmnop"),
|
|
},
|
|
},
|
|
},
|
|
Statistics: &cloudwatchlogs.QueryStatistics{
|
|
BytesScanned: aws.Float64(512),
|
|
RecordsMatched: aws.Float64(256),
|
|
RecordsScanned: aws.Float64(1024),
|
|
},
|
|
Status: aws.String("Complete"),
|
|
},
|
|
}
|
|
|
|
im := datasource.NewInstanceManager(func(s backend.DataSourceInstanceSettings) (instancemgmt.Instance, error) {
|
|
return DataSource{Settings: models.CloudWatchSettings{}}, nil
|
|
})
|
|
|
|
executor := newExecutor(im, newTestConfig(), &fakeSessionCache{}, featuremgmt.WithFeatures())
|
|
resp, err := executor.QueryData(context.Background(), &backend.QueryDataRequest{
|
|
PluginContext: backend.PluginContext{
|
|
DataSourceInstanceSettings: &backend.DataSourceInstanceSettings{},
|
|
},
|
|
Queries: []backend.DataQuery{
|
|
{
|
|
RefID: refID,
|
|
JSON: json.RawMessage(`{
|
|
"type": "logAction",
|
|
"subtype": "GetQueryResults",
|
|
"queryId": "abcd-efgh-ijkl-mnop"
|
|
}`),
|
|
},
|
|
},
|
|
})
|
|
require.NoError(t, err)
|
|
|
|
time1, err := time.Parse("2006-01-02 15:04:05.000", "2020-03-20 10:37:23.000")
|
|
require.NoError(t, err)
|
|
time2, err := time.Parse("2006-01-02 15:04:05.000", "2020-03-20 10:40:43.000")
|
|
require.NoError(t, err)
|
|
expField1 := data.NewField("@timestamp", nil, []*time.Time{
|
|
aws.Time(time1), aws.Time(time2),
|
|
})
|
|
expField1.SetConfig(&data.FieldConfig{DisplayName: "Time"})
|
|
expField2 := data.NewField("field_b", nil, []*string{
|
|
aws.String("b_1"), aws.String("b_2"),
|
|
})
|
|
expFrame := data.NewFrame(refID, expField1, expField2)
|
|
expFrame.RefID = refID
|
|
expFrame.Meta = &data.FrameMeta{
|
|
Custom: map[string]interface{}{
|
|
"Status": "Complete",
|
|
},
|
|
Stats: []data.QueryStat{
|
|
{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Bytes scanned"},
|
|
Value: 512,
|
|
},
|
|
{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Records scanned"},
|
|
Value: 1024,
|
|
},
|
|
{
|
|
FieldConfig: data.FieldConfig{DisplayName: "Records matched"},
|
|
Value: 256,
|
|
},
|
|
},
|
|
PreferredVisualization: "logs",
|
|
}
|
|
|
|
assert.Equal(t, &backend.QueryDataResponse{Responses: backend.Responses{
|
|
refID: {
|
|
Frames: data.Frames{expFrame},
|
|
},
|
|
},
|
|
}, resp)
|
|
}
|