forked from influxdata/kapacitor
-
Notifications
You must be signed in to change notification settings - Fork 0
/
delete.go
127 lines (117 loc) · 3.01 KB
/
delete.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
package kapacitor
import (
"log"
"github.com/influxdata/kapacitor/expvar"
"github.com/influxdata/kapacitor/models"
"github.com/influxdata/kapacitor/pipeline"
)
const (
statsFieldsDeleted = "fields_deleted"
statsTagsDeleted = "tags_deleted"
)
type DeleteNode struct {
node
d *pipeline.DeleteNode
fieldsDeleted *expvar.Int
tagsDeleted *expvar.Int
tags map[string]bool
}
// Create a new DeleteNode which applies a transformation func to each point in a stream and returns a single point.
func newDeleteNode(et *ExecutingTask, n *pipeline.DeleteNode, l *log.Logger) (*DeleteNode, error) {
dn := &DeleteNode{
node: node{Node: n, et: et, logger: l},
d: n,
tags: make(map[string]bool),
}
for _, tag := range n.Tags {
dn.tags[tag] = true
}
dn.node.runF = dn.runDelete
return dn, nil
}
func (e *DeleteNode) runDelete(snapshot []byte) error {
e.fieldsDeleted = &expvar.Int{}
e.tagsDeleted = &expvar.Int{}
e.statMap.Set(statsFieldsDeleted, e.fieldsDeleted)
e.statMap.Set(statsTagsDeleted, e.tagsDeleted)
switch e.Provides() {
case pipeline.StreamEdge:
for p, ok := e.ins[0].NextPoint(); ok; p, ok = e.ins[0].NextPoint() {
e.timer.Start()
p.Fields, p.Tags = e.doDeletes(p.Fields, p.Tags)
// Check if we deleted a group by dimension
updateDims := false
for _, dim := range p.Dimensions.TagNames {
if !e.tags[dim] {
updateDims = true
break
}
}
if updateDims {
newDims := make([]string, 0, len(p.Dimensions.TagNames))
for _, dim := range p.Dimensions.TagNames {
if !e.tags[dim] {
newDims = append(newDims, dim)
}
}
p.Dimensions.TagNames = newDims
p.Group = models.ToGroupID(p.Name, p.Tags, p.Dimensions)
}
e.timer.Stop()
for _, child := range e.outs {
err := child.CollectPoint(p)
if err != nil {
return err
}
}
}
case pipeline.BatchEdge:
for b, ok := e.ins[0].NextBatch(); ok; b, ok = e.ins[0].NextBatch() {
e.timer.Start()
b.Points = b.ShallowCopyPoints()
for i := range b.Points {
b.Points[i].Fields, b.Points[i].Tags = e.doDeletes(b.Points[i].Fields, b.Points[i].Tags)
}
_, newTags := e.doDeletes(nil, b.Tags)
if len(newTags) != len(b.Tags) {
b.Tags = newTags
b.Group = models.ToGroupID(b.Name, b.Tags, b.PointDimensions())
}
e.timer.Stop()
for _, child := range e.outs {
err := child.CollectBatch(b)
if err != nil {
return err
}
}
}
}
return nil
}
func (d *DeleteNode) doDeletes(fields models.Fields, tags models.Tags) (models.Fields, models.Tags) {
newFields := fields
fieldsCopied := false
for _, field := range d.d.Fields {
if _, ok := fields[field]; ok {
if !fieldsCopied {
newFields = newFields.Copy()
fieldsCopied = true
}
d.fieldsDeleted.Add(1)
delete(newFields, field)
}
}
newTags := tags
tagsCopied := false
for _, tag := range d.d.Tags {
if _, ok := tags[tag]; ok {
if !tagsCopied {
newTags = newTags.Copy()
tagsCopied = true
}
d.tagsDeleted.Add(1)
delete(newTags, tag)
}
}
return newFields, newTags
}