Files
grafana/pkg/services/caching/service.go
Denis Vodopianov 0e460a267e chore : Deprecating FeatureToggles.IsEnabled (#113062)
* Deprecating features.IsEnabled

* add one more nolint

* add one more nolint

* Give better hints to devs in the deprecation message of IsEnabledGlobally

* adding more doc strings

* fix linter after rebase

* Extend deprecation message
2025-11-21 18:43:42 +01:00

288 lines
10 KiB
Go

package caching
import (
"bytes"
"context"
"crypto/sha256"
"encoding/hex"
"encoding/json"
"io"
"strconv"
"strings"
"time"
"github.com/grafana/grafana-aws-sdk/pkg/awsds"
"github.com/grafana/grafana/pkg/infra/log"
"github.com/grafana/grafana-plugin-sdk-go/backend"
"github.com/grafana/grafana/pkg/services/contexthandler"
"github.com/grafana/grafana/pkg/services/featuremgmt"
"github.com/prometheus/client_golang/prometheus"
)
type CacheStatus string
const (
XCacheHeader = "X-Cache"
StatusHit CacheStatus = "HIT"
StatusMiss CacheStatus = "MISS"
StatusBypass CacheStatus = "BYPASS"
StatusError CacheStatus = "ERROR"
StatusDisabled CacheStatus = "DISABLED"
)
// needed to mock the function for testing
var ShouldCacheQuery = awsds.ShouldCacheQuery
type CacheQueryResponseFn func(context.Context, *backend.QueryDataResponse)
type CacheResourceResponseFn func(context.Context, *backend.CallResourceResponse)
type CachedQueryDataResponse struct {
// The cached data response associated with a query, or nil if no cached data is found
Response *backend.QueryDataResponse
// A function that should be used to cache a QueryDataResponse for a given query.
// It can be set to nil by the method implementation (if there is an error, for example), so it should be checked before being called.
UpdateCacheFn CacheQueryResponseFn
}
type CachedResourceDataResponse struct {
// The cached response associated with a resource request, or nil if no cached data is found
Response *backend.CallResourceResponse
// A function that should be used to cache a CallResourceResponse for a given resource request.
// It can be set to nil by the method implementation (if there is an error, for example), so it should be checked before being called.
// Because plugins can send multiple responses asynchronously, the implementation should be able to handle multiple calls to this function for one request.
UpdateCacheFn CacheResourceResponseFn
}
func ProvideCachingService() *OSSCachingService {
return &OSSCachingService{}
}
type CachingService interface {
// HandleQueryRequest uses a QueryDataRequest to check the cache for any existing results for that query.
// If none are found, it should return false and a CachedQueryDataResponse with an UpdateCacheFn which can be used to update the results cache after the fact.
// This function may populate any response headers (accessible through the context) with the cache status using the X-Cache header.
HandleQueryRequest(ctx context.Context, req *backend.QueryDataRequest) (bool, CachedQueryDataResponse, CacheStatus)
// HandleResourceRequest uses a CallResourceRequest to check the cache for any existing results for that request. If none are found, it should return false.
// This function may populate any response headers (accessible through the context) with the cache status using the X-Cache header.
HandleResourceRequest(ctx context.Context, req *backend.CallResourceRequest) (bool, CachedResourceDataResponse, CacheStatus)
}
// Implementation of interface - does nothing
type OSSCachingService struct {
}
func (s *OSSCachingService) HandleQueryRequest(ctx context.Context, req *backend.QueryDataRequest) (bool, CachedQueryDataResponse, CacheStatus) {
return false, CachedQueryDataResponse{}, ""
}
func (s *OSSCachingService) HandleResourceRequest(ctx context.Context, req *backend.CallResourceRequest) (bool, CachedResourceDataResponse, CacheStatus) {
return false, CachedResourceDataResponse{}, ""
}
var _ CachingService = &OSSCachingService{}
// GetKey creates a prefixed cache key and uses the internal `encoder` to encode the query into a string
func GetKey(namespace, prefix string, query interface{}) (string, error) {
keybuf := bytes.NewBuffer(nil)
encoder := &JSONEncoder{}
if err := encoder.Encode(keybuf, query); err != nil {
return "", err
}
key, err := SHA256KeyFunc(keybuf)
if err != nil {
return "", err
}
// The namespace is empty only when this function is used by the legacy caching module.
// This case can be removed when the legacy caching module is not being used anymore.
if namespace != "" {
return strings.Join([]string{namespace, prefix, key}, ":"), nil
}
return strings.Join([]string{prefix, key}, ":"), nil
}
// SHA256KeyFunc copies the data from `r` into a sha256.Hash, and returns the encoded Sum.
func SHA256KeyFunc(r io.Reader) (string, error) {
hash := sha256.New()
// Read all data from the provided reader
if _, err := io.Copy(hash, r); err != nil {
return "", err
}
// Encode the written values to SHA256
return hex.EncodeToString(hash.Sum(nil)), nil
}
// JSONEncoder encodes and decodes struct data to/from JSON
type JSONEncoder struct{}
// NewJSONEncoder creates a pointer to a new JSONEncoder, which implements the `Encoder` interface
func NewJSONEncoder() *JSONEncoder {
return &JSONEncoder{}
}
func (e *JSONEncoder) EncodeBytes(w io.Writer, b []byte) error {
_, err := w.Write(b)
if err != nil {
return err
}
return nil
}
func (e *JSONEncoder) DecodeBytes(r io.Reader) ([]byte, error) {
encBytes, err := io.ReadAll(r)
if err != nil {
return []byte{}, err
}
return encBytes, err
}
// Encode encodes the `v` interface into `w` using a json.Encoder
func (e *JSONEncoder) Encode(w io.Writer, v interface{}) error {
return json.NewEncoder(w).Encode(v)
}
// Decode encodes the io.Reader `r` into the interface `v` using a json.Decoder
func (e *JSONEncoder) Decode(r io.Reader, v interface{}) error {
return json.NewDecoder(r).Decode(v)
}
// A service that provides methods to cache requests.
// It can be used to cache requests using `caching.CachingService` without reimplementing
// the caching logic at every call site.
type CachingServiceClient struct {
cachingService CachingService
features featuremgmt.FeatureToggles
}
func ProvideCachingServiceClient(cachingService CachingService, features featuremgmt.FeatureToggles) *CachingServiceClient {
log := log.New("caching_service_client")
if err := prometheus.Register(QueryCachingRequestHistogram); err != nil {
log.Error("Error registering prometheus collector 'QueryRequestHistogram'", "error", err)
}
if err := prometheus.Register(ResourceCachingRequestHistogram); err != nil {
log.Error("Error registering prometheus collector 'ResourceRequestHistogram'", "error", err)
}
return &CachingServiceClient{cachingService: cachingService, features: features}
}
// WithQueryDataCaching calls `f` and caches the returned value if `req` has not been cached already.
// Returns the cached value otherwise.
func (c *CachingServiceClient) WithQueryDataCaching(ctx context.Context, req *backend.QueryDataRequest, f func() (*backend.QueryDataResponse, error)) (*backend.QueryDataResponse, error) {
if c == nil || req == nil {
return f()
}
reqCtx := contexthandler.FromContext(ctx)
// time how long this request takes
start := time.Now()
// First look in the query cache if enabled
hit, cr, status := c.cachingService.HandleQueryRequest(ctx, req)
// record request duration if caching was used
if reqCtx != nil {
reqCtx.Resp.Header().Set(XCacheHeader, string(status))
defer func() {
QueryCachingRequestHistogram.With(prometheus.Labels{
"datasource_type": getDatasourceType(req.PluginContext),
"cache": string(status),
"query_type": getQueryType(reqCtx),
}).Observe(time.Since(start).Seconds())
}()
}
// Cache hit; return the response
if hit {
return cr.Response, nil
}
// Cache miss; do the actual queries
resp, err := f()
// Update the query cache with the result for this metrics request
if err == nil && cr.UpdateCacheFn != nil {
// If AWS async caching is not enabled, use the old code path
//nolint:staticcheck // not yet migrated to OpenFeature
if c.features == nil || !c.features.IsEnabled(ctx, featuremgmt.FlagAwsAsyncQueryCaching) {
cr.UpdateCacheFn(ctx, resp)
} else if reqCtx != nil {
// time how long shouldCacheQuery takes
startShouldCacheQuery := time.Now()
shouldCache := ShouldCacheQuery(resp)
ShouldCacheQueryHistogram.With(prometheus.Labels{
"datasource_type": req.PluginContext.DataSourceInstanceSettings.Type,
"cache": string(status),
"shouldCache": strconv.FormatBool(shouldCache),
"query_type": getQueryType(reqCtx),
}).Observe(time.Since(startShouldCacheQuery).Seconds())
// If AWS async caching is enabled and resp is for a running async query, don't cache it
if shouldCache {
cr.UpdateCacheFn(ctx, resp)
}
}
}
return resp, err
}
// WithCallResourceCaching calls `f` and caches the returned value if `req` has not been cached already.
// Returns the cached value otherwise.
func (c *CachingServiceClient) WithCallResourceCaching(ctx context.Context, req *backend.CallResourceRequest, sender backend.CallResourceResponseSender, f func(backend.CallResourceResponseSender) error) error {
if c == nil || req == nil {
return f(sender)
}
reqCtx := contexthandler.FromContext(ctx)
// time how long this request takes
start := time.Now()
// First look in the resource cache if enabled
hit, cr, status := c.cachingService.HandleResourceRequest(ctx, req)
if reqCtx != nil {
reqCtx.Resp.Header().Set(XCacheHeader, string(status))
}
// record request duration if caching was used
defer func() {
ResourceCachingRequestHistogram.With(prometheus.Labels{
"plugin_id": req.PluginContext.PluginID,
"cache": string(status),
}).Observe(time.Since(start).Seconds())
}()
// Cache hit; send the response and return
if hit {
return sender.Send(cr.Response)
}
// Cache miss; do the actual request
// If there is no update cache func, just pass in the original sender
if cr.UpdateCacheFn == nil {
return f(sender)
}
// Otherwise, intercept the responses in a wrapped sender so we can cache them first
cacheSender := backend.CallResourceResponseSenderFunc(func(res *backend.CallResourceResponse) error {
cr.UpdateCacheFn(ctx, res)
return sender.Send(res)
})
return f(cacheSender)
}
func getDatasourceType(pluginCtx backend.PluginContext) string {
if pluginCtx.DataSourceInstanceSettings == nil {
return "unknown"
}
return pluginCtx.DataSourceInstanceSettings.Name
}