mirror of
https://github.com/SigNoz/signoz.git
synced 2025-12-17 07:26:20 +00:00
* chore: added custom distpatcher * feat(notification-grouping): added notification grouping * feat(notification-grouping): addded integration test dependency * feat(notification-grouping): linting and test cases * feat(notification-grouping): linting and test cases * feat(notification-grouping): linting and test cases * feat(notification-grouping): addded integration test dependency * feat(notification-grouping): debug log lines * feat(notification-grouping): debug log lines * feat(notification-grouping): debug log lines * feat(notification-grouping): addded integration test dependency * feat(notification-grouping): addded integration test dependency * feat(notification-grouping): addded integration test dependency * feat(notification-grouping): added structure changes * feat(notification-grouping): added structure changes * feat(notification-routing): added notification routing * chore(notification-grouping): added notificaiton grouping * Update pkg/alertmanager/nfmanager/rulebasednotification/provider.go Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore(notification-grouping): added renotification interval * fix(notification-grouping): added fix for renotification * chore(notificaiton-grouping): added no data renotify * chore(notificaiton-grouping): added no data renotify * chore(notificaiton-grouping): added no data renotify * chore(notification-grouping): added no data renotify interval * chore(notification-grouping): removed errors package from dispatcher * chore(notification-grouping): removed errors package from dispatcher * chore(notification-grouping): removed unwanted tests * chore(notification-grouping): removed unwanted pkg name * chore(notification-grouping): added delete notification setting * chore(notification-grouping): added delete notification setting * Update pkg/alertmanager/nfmanager/nfmanagertest/provider.go Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore(notification-grouping): removed nfmanager config| notification settings in postable rule * chore(notification-grouping): removed nfmanager config| notification settings in postable rule * chore(notification-grouping): added test for dispatcher * chore(notification-grouping): added test for dispatcher * chore(notification-grouping): go linting errors * chore(notification-grouping): added test cases for aggGroupPerRoute * chore(notification-grouping): added test cases for aggGroupPerRoute * chore(notification-grouping): corrected get notification config logic * Update pkg/alertmanager/nfmanager/rulebasednotification/provider_test.go Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore(notification-routing): added notification routing policies * feat(notification-routing): added test cases for dispatcher * chore(notification-routing): added notification routing policies * chore(notification-routing): added notification routing policies * Apply suggestions from code review Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore(notification-routing): added notification routing policies * chore(notification-routing): added notification routing policies * Update pkg/alertmanager/alertmanagerserver/distpatcher_test.go Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore(notification-routing): sorted imports * chore(notification-routing): minor edit |pr resolve comments * chore(notification-grouping): corrected dispatcher test cases * chore(notification-routing): added notification routing policies * chore(notification-routing): corrected race condition in test * chore: resolved pr comments * chore: passing threshold value to tempalte * chore: completed delete rule functionality * chore: added grouping disabled functionality * chore: added grouping disabled functionality * chore(notification-routing): resolved pr comments * chore(notification-routing): resolved pr comments * chore(notification-routing): resolved pr comments * chore(notification-routing): sorted imports * chore(notification-routing): fix linting errors * chore(notification-routing): removed enabled flags * fix: test rule multiple threhsold (#9224) * chore: corrected linting errors * chore: corrected linting errors * chore: corrected linting errors * chore: corrected linting errors * chore: corrected migration errors * chore: corrected migration errors * chore: corrected migration errors * chore: corrected migration errors * Update pkg/sqlmigration/049_add_route_policy.go Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com> * chore: added org_is as foreign key * chore: resolved pr comments * chore: removed route store unused --------- Co-authored-by: Srikanth Chekuri <srikanth.chekuri92@gmail.com> Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
511 lines
15 KiB
Go
511 lines
15 KiB
Go
package rules
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"log/slog"
|
|
"net/url"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/SigNoz/signoz/pkg/errors"
|
|
"github.com/SigNoz/signoz/pkg/query-service/interfaces"
|
|
"github.com/SigNoz/signoz/pkg/query-service/model"
|
|
v3 "github.com/SigNoz/signoz/pkg/query-service/model/v3"
|
|
qslabels "github.com/SigNoz/signoz/pkg/query-service/utils/labels"
|
|
"github.com/SigNoz/signoz/pkg/sqlstore"
|
|
ruletypes "github.com/SigNoz/signoz/pkg/types/ruletypes"
|
|
"github.com/SigNoz/signoz/pkg/valuer"
|
|
"go.uber.org/zap"
|
|
)
|
|
|
|
// BaseRule contains common fields and methods for all rule types
|
|
type BaseRule struct {
|
|
id string
|
|
name string
|
|
orgID valuer.UUID
|
|
source string
|
|
handledRestart bool
|
|
|
|
// Type of the rule
|
|
typ ruletypes.AlertType
|
|
|
|
ruleCondition *ruletypes.RuleCondition
|
|
|
|
Threshold ruletypes.RuleThreshold
|
|
// evalWindow is the time window used for evaluating the rule
|
|
// i.e each time we lookback from the current time, we look at data for the last
|
|
// evalWindow duration
|
|
evalWindow time.Duration
|
|
// holdDuration is the duration for which the alert waits before firing
|
|
holdDuration time.Duration
|
|
|
|
// evalDelay is the delay in evaluation of the rule
|
|
// this is useful in cases where the data is not available immediately
|
|
evalDelay time.Duration
|
|
|
|
// holds the static set of labels and annotations for the rule
|
|
// these are the same for all alerts created for this rule
|
|
labels qslabels.BaseLabels
|
|
annotations qslabels.BaseLabels
|
|
// preferredChannels is the list of channels to send the alert to
|
|
// if the rule is triggered
|
|
preferredChannels []string
|
|
mtx sync.Mutex
|
|
// the time it took to evaluate the rule (most recent evaluation)
|
|
evaluationDuration time.Duration
|
|
// the timestamp of the last evaluation
|
|
evaluationTimestamp time.Time
|
|
|
|
health ruletypes.RuleHealth
|
|
lastError error
|
|
Active map[uint64]*ruletypes.Alert
|
|
|
|
// lastTimestampWithDatapoints is the timestamp of the last datapoint we observed
|
|
// for this rule
|
|
// this is used for missing data alerts
|
|
lastTimestampWithDatapoints time.Time
|
|
|
|
reader interfaces.Reader
|
|
|
|
logger *slog.Logger
|
|
|
|
// sendUnmatched sends observed metric values
|
|
// even if they dont match the rule condition. this is
|
|
// useful in testing the rule
|
|
sendUnmatched bool
|
|
|
|
// sendAlways will send alert irresepective of resendDelay
|
|
// or other params
|
|
sendAlways bool
|
|
|
|
// TemporalityMap is a map of metric name to temporality
|
|
// to avoid fetching temporality for the same metric multiple times
|
|
// querying the v4 table on low cardinal temporality column
|
|
// should be fast but we can still avoid the query if we have the data in memory
|
|
TemporalityMap map[string]map[v3.Temporality]bool
|
|
|
|
sqlstore sqlstore.SQLStore
|
|
|
|
evaluation ruletypes.Evaluation
|
|
}
|
|
|
|
type RuleOption func(*BaseRule)
|
|
|
|
func WithSendAlways() RuleOption {
|
|
return func(r *BaseRule) {
|
|
r.sendAlways = true
|
|
}
|
|
}
|
|
|
|
func WithSendUnmatched() RuleOption {
|
|
return func(r *BaseRule) {
|
|
r.sendUnmatched = true
|
|
}
|
|
}
|
|
|
|
func WithEvalDelay(dur time.Duration) RuleOption {
|
|
return func(r *BaseRule) {
|
|
r.evalDelay = dur
|
|
}
|
|
}
|
|
|
|
func WithLogger(logger *slog.Logger) RuleOption {
|
|
return func(r *BaseRule) {
|
|
r.logger = logger
|
|
}
|
|
}
|
|
|
|
func WithSQLStore(sqlstore sqlstore.SQLStore) RuleOption {
|
|
return func(r *BaseRule) {
|
|
r.sqlstore = sqlstore
|
|
}
|
|
}
|
|
|
|
func NewBaseRule(id string, orgID valuer.UUID, p *ruletypes.PostableRule, reader interfaces.Reader, opts ...RuleOption) (*BaseRule, error) {
|
|
if p.RuleCondition == nil || !p.RuleCondition.IsValid() {
|
|
return nil, fmt.Errorf("invalid rule condition")
|
|
}
|
|
threshold, err := p.RuleCondition.Thresholds.GetRuleThreshold()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
evaluation, err := p.Evaluation.GetEvaluation()
|
|
if err != nil {
|
|
return nil, errors.NewInvalidInputf(errors.CodeInvalidInput, "failed to get evaluation: %v", err)
|
|
}
|
|
|
|
baseRule := &BaseRule{
|
|
id: id,
|
|
orgID: orgID,
|
|
name: p.AlertName,
|
|
source: p.Source,
|
|
typ: p.AlertType,
|
|
ruleCondition: p.RuleCondition,
|
|
evalWindow: time.Duration(p.EvalWindow),
|
|
labels: qslabels.FromMap(p.Labels),
|
|
annotations: qslabels.FromMap(p.Annotations),
|
|
preferredChannels: p.PreferredChannels,
|
|
health: ruletypes.HealthUnknown,
|
|
Active: map[uint64]*ruletypes.Alert{},
|
|
reader: reader,
|
|
TemporalityMap: make(map[string]map[v3.Temporality]bool),
|
|
Threshold: threshold,
|
|
evaluation: evaluation,
|
|
}
|
|
|
|
if baseRule.evalWindow == 0 {
|
|
baseRule.evalWindow = 5 * time.Minute
|
|
}
|
|
|
|
for _, opt := range opts {
|
|
opt(baseRule)
|
|
}
|
|
|
|
return baseRule, nil
|
|
}
|
|
|
|
func (r *BaseRule) matchType() ruletypes.MatchType {
|
|
if r.ruleCondition == nil {
|
|
return ruletypes.AtleastOnce
|
|
}
|
|
return r.ruleCondition.MatchType
|
|
}
|
|
|
|
func (r *BaseRule) compareOp() ruletypes.CompareOp {
|
|
if r.ruleCondition == nil {
|
|
return ruletypes.ValueIsEq
|
|
}
|
|
return r.ruleCondition.CompareOp
|
|
}
|
|
|
|
func (r *BaseRule) currentAlerts() []*ruletypes.Alert {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
|
|
alerts := make([]*ruletypes.Alert, 0, len(r.Active))
|
|
for _, a := range r.Active {
|
|
anew := *a
|
|
alerts = append(alerts, &anew)
|
|
}
|
|
return alerts
|
|
}
|
|
|
|
func (r *BaseRule) EvalDelay() time.Duration {
|
|
return r.evalDelay
|
|
}
|
|
|
|
func (r *BaseRule) EvalWindow() time.Duration {
|
|
return r.evalWindow
|
|
}
|
|
|
|
func (r *BaseRule) HoldDuration() time.Duration {
|
|
return r.holdDuration
|
|
}
|
|
|
|
func (r *ThresholdRule) hostFromSource() string {
|
|
parsedUrl, err := url.Parse(r.source)
|
|
if err != nil {
|
|
return ""
|
|
}
|
|
if parsedUrl.Port() != "" {
|
|
return fmt.Sprintf("%s://%s:%s", parsedUrl.Scheme, parsedUrl.Hostname(), parsedUrl.Port())
|
|
}
|
|
return fmt.Sprintf("%s://%s", parsedUrl.Scheme, parsedUrl.Hostname())
|
|
}
|
|
|
|
func (r *BaseRule) ID() string { return r.id }
|
|
func (r *BaseRule) OrgID() valuer.UUID { return r.orgID }
|
|
func (r *BaseRule) Name() string { return r.name }
|
|
func (r *BaseRule) Condition() *ruletypes.RuleCondition { return r.ruleCondition }
|
|
func (r *BaseRule) Labels() qslabels.BaseLabels { return r.labels }
|
|
func (r *BaseRule) Annotations() qslabels.BaseLabels { return r.annotations }
|
|
func (r *BaseRule) PreferredChannels() []string { return r.preferredChannels }
|
|
|
|
func (r *BaseRule) GeneratorURL() string {
|
|
return ruletypes.PrepareRuleGeneratorURL(r.ID(), r.source)
|
|
}
|
|
|
|
func (r *BaseRule) Unit() string {
|
|
if r.ruleCondition != nil && r.ruleCondition.CompositeQuery != nil {
|
|
return r.ruleCondition.CompositeQuery.Unit
|
|
}
|
|
return ""
|
|
}
|
|
|
|
func (r *BaseRule) Timestamps(ts time.Time) (time.Time, time.Time) {
|
|
|
|
st, en := r.evaluation.NextWindowFor(ts)
|
|
start := st.UnixMilli()
|
|
end := en.UnixMilli()
|
|
|
|
if r.evalDelay > 0 {
|
|
start = start - int64(r.evalDelay.Milliseconds())
|
|
end = end - int64(r.evalDelay.Milliseconds())
|
|
}
|
|
// round to minute otherwise we could potentially miss data
|
|
start = start - (start % (60 * 1000))
|
|
end = end - (end % (60 * 1000))
|
|
|
|
return time.UnixMilli(start), time.UnixMilli(end)
|
|
}
|
|
|
|
func (r *BaseRule) SetLastError(err error) {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
r.lastError = err
|
|
}
|
|
|
|
func (r *BaseRule) LastError() error {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
return r.lastError
|
|
}
|
|
|
|
func (r *BaseRule) SetHealth(health ruletypes.RuleHealth) {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
r.health = health
|
|
}
|
|
|
|
func (r *BaseRule) Health() ruletypes.RuleHealth {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
return r.health
|
|
}
|
|
|
|
func (r *BaseRule) SetEvaluationDuration(dur time.Duration) {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
r.evaluationDuration = dur
|
|
}
|
|
|
|
func (r *BaseRule) GetEvaluationDuration() time.Duration {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
return r.evaluationDuration
|
|
}
|
|
|
|
func (r *BaseRule) SetEvaluationTimestamp(ts time.Time) {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
r.evaluationTimestamp = ts
|
|
}
|
|
|
|
func (r *BaseRule) GetEvaluationTimestamp() time.Time {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
return r.evaluationTimestamp
|
|
}
|
|
|
|
func (r *BaseRule) State() model.AlertState {
|
|
maxState := model.StateInactive
|
|
for _, a := range r.Active {
|
|
if a.State > maxState {
|
|
maxState = a.State
|
|
}
|
|
}
|
|
return maxState
|
|
}
|
|
|
|
func (r *BaseRule) ActiveAlerts() []*ruletypes.Alert {
|
|
var res []*ruletypes.Alert
|
|
for _, a := range r.currentAlerts() {
|
|
if a.ResolvedAt.IsZero() {
|
|
res = append(res, a)
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
func (r *BaseRule) SendAlerts(ctx context.Context, ts time.Time, resendDelay time.Duration, interval time.Duration, notifyFunc NotifyFunc) {
|
|
var orgID string
|
|
err := r.
|
|
sqlstore.
|
|
BunDB().
|
|
NewSelect().
|
|
Table("organizations").
|
|
ColumnExpr("id").
|
|
Limit(1).
|
|
Scan(ctx, &orgID)
|
|
if err != nil {
|
|
r.logger.ErrorContext(ctx, "failed to get org ids", "error", err)
|
|
return
|
|
}
|
|
|
|
alerts := []*ruletypes.Alert{}
|
|
r.ForEachActiveAlert(func(alert *ruletypes.Alert) {
|
|
if alert.NeedsSending(ts, resendDelay) {
|
|
alert.LastSentAt = ts
|
|
delta := resendDelay
|
|
if interval > resendDelay {
|
|
delta = interval
|
|
}
|
|
alert.ValidUntil = ts.Add(4 * delta)
|
|
anew := *alert
|
|
alerts = append(alerts, &anew)
|
|
}
|
|
})
|
|
notifyFunc(ctx, orgID, "", alerts...)
|
|
}
|
|
|
|
func (r *BaseRule) ForEachActiveAlert(f func(*ruletypes.Alert)) {
|
|
r.mtx.Lock()
|
|
defer r.mtx.Unlock()
|
|
|
|
for _, a := range r.Active {
|
|
f(a)
|
|
}
|
|
}
|
|
|
|
func (r *BaseRule) RecordRuleStateHistory(ctx context.Context, prevState, currentState model.AlertState, itemsToAdd []model.RuleStateHistory) error {
|
|
zap.L().Debug("recording rule state history", zap.String("ruleid", r.ID()), zap.Any("prevState", prevState), zap.Any("currentState", currentState), zap.Any("itemsToAdd", itemsToAdd))
|
|
revisedItemsToAdd := map[uint64]model.RuleStateHistory{}
|
|
|
|
lastSavedState, err := r.reader.GetLastSavedRuleStateHistory(ctx, r.ID())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
// if the query-service has been restarted, or the rule has been modified (which re-initializes the rule),
|
|
// the state would reset so we need to add the corresponding state changes to previously saved states
|
|
if !r.handledRestart && len(lastSavedState) > 0 {
|
|
zap.L().Debug("handling restart", zap.String("ruleid", r.ID()), zap.Any("lastSavedState", lastSavedState))
|
|
l := map[uint64]model.RuleStateHistory{}
|
|
for _, item := range itemsToAdd {
|
|
l[item.Fingerprint] = item
|
|
}
|
|
|
|
shouldSkip := map[uint64]bool{}
|
|
|
|
for _, item := range lastSavedState {
|
|
// for the last saved item with fingerprint, check if there is a corresponding entry in the current state
|
|
currentState, ok := l[item.Fingerprint]
|
|
if !ok {
|
|
// there was a state change in the past, but not in the current state
|
|
// if the state was firing, then we should add a resolved state change
|
|
if item.State == model.StateFiring || item.State == model.StateNoData {
|
|
item.State = model.StateInactive
|
|
item.StateChanged = true
|
|
item.UnixMilli = time.Now().UnixMilli()
|
|
revisedItemsToAdd[item.Fingerprint] = item
|
|
}
|
|
// there is nothing to do if the prev state was normal
|
|
} else {
|
|
if item.State != currentState.State {
|
|
item.State = currentState.State
|
|
item.StateChanged = true
|
|
item.UnixMilli = time.Now().UnixMilli()
|
|
revisedItemsToAdd[item.Fingerprint] = item
|
|
}
|
|
}
|
|
// do not add this item to revisedItemsToAdd as it is already processed
|
|
shouldSkip[item.Fingerprint] = true
|
|
}
|
|
zap.L().Debug("after lastSavedState loop", zap.String("ruleid", r.ID()), zap.Any("revisedItemsToAdd", revisedItemsToAdd))
|
|
|
|
// if there are any new state changes that were not saved, add them to the revised items
|
|
for _, item := range itemsToAdd {
|
|
if _, ok := revisedItemsToAdd[item.Fingerprint]; !ok && !shouldSkip[item.Fingerprint] {
|
|
revisedItemsToAdd[item.Fingerprint] = item
|
|
}
|
|
}
|
|
zap.L().Debug("after itemsToAdd loop", zap.String("ruleid", r.ID()), zap.Any("revisedItemsToAdd", revisedItemsToAdd))
|
|
|
|
newState := model.StateInactive
|
|
for _, item := range revisedItemsToAdd {
|
|
if item.State == model.StateFiring || item.State == model.StateNoData {
|
|
newState = model.StateFiring
|
|
break
|
|
}
|
|
}
|
|
zap.L().Debug("newState", zap.String("ruleid", r.ID()), zap.Any("newState", newState))
|
|
|
|
// if there is a change in the overall state, update the overall state
|
|
if lastSavedState[0].OverallState != newState {
|
|
for fingerprint, item := range revisedItemsToAdd {
|
|
item.OverallState = newState
|
|
item.OverallStateChanged = true
|
|
revisedItemsToAdd[fingerprint] = item
|
|
}
|
|
}
|
|
zap.L().Debug("revisedItemsToAdd after newState", zap.String("ruleid", r.ID()), zap.Any("revisedItemsToAdd", revisedItemsToAdd))
|
|
|
|
} else {
|
|
for _, item := range itemsToAdd {
|
|
revisedItemsToAdd[item.Fingerprint] = item
|
|
}
|
|
}
|
|
|
|
if len(revisedItemsToAdd) > 0 && r.reader != nil {
|
|
zap.L().Debug("writing rule state history", zap.String("ruleid", r.ID()), zap.Any("revisedItemsToAdd", revisedItemsToAdd))
|
|
|
|
entries := make([]model.RuleStateHistory, 0, len(revisedItemsToAdd))
|
|
for _, item := range revisedItemsToAdd {
|
|
entries = append(entries, item)
|
|
}
|
|
err := r.reader.AddRuleStateHistory(ctx, entries)
|
|
if err != nil {
|
|
zap.L().Error("error while inserting rule state history", zap.Error(err), zap.Any("itemsToAdd", itemsToAdd))
|
|
}
|
|
}
|
|
r.handledRestart = true
|
|
|
|
return nil
|
|
}
|
|
|
|
func (r *BaseRule) PopulateTemporality(ctx context.Context, orgID valuer.UUID, qp *v3.QueryRangeParamsV3) error {
|
|
|
|
missingTemporality := make([]string, 0)
|
|
metricNameToTemporality := make(map[string]map[v3.Temporality]bool)
|
|
if qp.CompositeQuery != nil && len(qp.CompositeQuery.BuilderQueries) > 0 {
|
|
for _, query := range qp.CompositeQuery.BuilderQueries {
|
|
// if there is no temporality specified in the query but we have it in the map
|
|
// then use the value from the map
|
|
if query.Temporality == "" && r.TemporalityMap[query.AggregateAttribute.Key] != nil {
|
|
// We prefer delta if it is available
|
|
if r.TemporalityMap[query.AggregateAttribute.Key][v3.Delta] {
|
|
query.Temporality = v3.Delta
|
|
} else if r.TemporalityMap[query.AggregateAttribute.Key][v3.Cumulative] {
|
|
query.Temporality = v3.Cumulative
|
|
} else {
|
|
query.Temporality = v3.Unspecified
|
|
}
|
|
}
|
|
// we don't have temporality for this metric
|
|
if query.DataSource == v3.DataSourceMetrics && query.Temporality == "" {
|
|
missingTemporality = append(missingTemporality, query.AggregateAttribute.Key)
|
|
}
|
|
if _, ok := metricNameToTemporality[query.AggregateAttribute.Key]; !ok {
|
|
metricNameToTemporality[query.AggregateAttribute.Key] = make(map[v3.Temporality]bool)
|
|
}
|
|
}
|
|
}
|
|
|
|
var nameToTemporality map[string]map[v3.Temporality]bool
|
|
var err error
|
|
|
|
if len(missingTemporality) > 0 {
|
|
nameToTemporality, err = r.reader.FetchTemporality(ctx, orgID, missingTemporality)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
if qp.CompositeQuery != nil && len(qp.CompositeQuery.BuilderQueries) > 0 {
|
|
for name := range qp.CompositeQuery.BuilderQueries {
|
|
query := qp.CompositeQuery.BuilderQueries[name]
|
|
if query.DataSource == v3.DataSourceMetrics && query.Temporality == "" {
|
|
if nameToTemporality[query.AggregateAttribute.Key][v3.Delta] {
|
|
query.Temporality = v3.Delta
|
|
} else if nameToTemporality[query.AggregateAttribute.Key][v3.Cumulative] {
|
|
query.Temporality = v3.Cumulative
|
|
} else {
|
|
query.Temporality = v3.Unspecified
|
|
}
|
|
r.TemporalityMap[query.AggregateAttribute.Key] = nameToTemporality[query.AggregateAttribute.Key]
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|