mirror of
https://github.com/pomerium/pomerium.git
synced 2025-04-30 02:46:30 +02:00
zero/telemetry: calculate DAU and MAU (#4810)
This commit is contained in:
parent
e0ac870442
commit
c4dd965f2d
9 changed files with 485 additions and 0 deletions
|
@ -31,3 +31,12 @@ func (s *Hash[T]) Has(element T) bool {
|
||||||
func (s *Hash[T]) Size() int {
|
func (s *Hash[T]) Size() int {
|
||||||
return len(s.m)
|
return len(s.m)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Items returns the set's elements as a slice.
|
||||||
|
func (s *Hash[T]) Items() []T {
|
||||||
|
items := make([]T, 0, len(s.m))
|
||||||
|
for item := range s.m {
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
return items
|
||||||
|
}
|
||||||
|
|
80
internal/zero/analytics/activeusers.go
Normal file
80
internal/zero/analytics/activeusers.go
Normal file
|
@ -0,0 +1,80 @@
|
||||||
|
package analytics
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/pkg/counter"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// activeUsersCap is the number of active users we would be able to track.
|
||||||
|
// for counter to work within the 1% error limit, actual number should be 80% of the cap.
|
||||||
|
activeUsersCap = 10_000
|
||||||
|
)
|
||||||
|
|
||||||
|
// IntervalResetFunc is a function that determines if a counter should be reset
|
||||||
|
type IntervalResetFunc func(lastReset time.Time, now time.Time) bool
|
||||||
|
|
||||||
|
// ResetMonthlyUTC resets the counter on a monthly interval
|
||||||
|
func ResetMonthlyUTC(lastReset time.Time, now time.Time) bool {
|
||||||
|
lastResetUTC := lastReset.UTC()
|
||||||
|
nowUTC := now.UTC()
|
||||||
|
return lastResetUTC.Year() != nowUTC.Year() ||
|
||||||
|
lastResetUTC.Month() != nowUTC.Month()
|
||||||
|
}
|
||||||
|
|
||||||
|
// ResetDailyUTC resets the counter on a daily interval
|
||||||
|
func ResetDailyUTC(lastReset time.Time, now time.Time) bool {
|
||||||
|
lastResetUTC := lastReset.UTC()
|
||||||
|
nowUTC := now.UTC()
|
||||||
|
return lastResetUTC.Year() != nowUTC.Year() ||
|
||||||
|
lastResetUTC.Month() != nowUTC.Month() ||
|
||||||
|
lastResetUTC.Day() != nowUTC.Day()
|
||||||
|
}
|
||||||
|
|
||||||
|
// ActiveUsersCounter is a counter that resets on a given interval
|
||||||
|
type ActiveUsersCounter struct {
|
||||||
|
*counter.Counter
|
||||||
|
lastReset time.Time
|
||||||
|
needsReset IntervalResetFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewActiveUsersCounter creates a new active users counter
|
||||||
|
func NewActiveUsersCounter(needsReset IntervalResetFunc, now time.Time) *ActiveUsersCounter {
|
||||||
|
return &ActiveUsersCounter{
|
||||||
|
Counter: counter.New(activeUsersCap),
|
||||||
|
lastReset: now,
|
||||||
|
needsReset: needsReset,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// LoadActiveUsersCounter loads an active users counter from a binary state
|
||||||
|
func LoadActiveUsersCounter(state []byte, lastReset time.Time, resetFn IntervalResetFunc) (*ActiveUsersCounter, error) {
|
||||||
|
c, err := counter.FromBinary(state)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return &ActiveUsersCounter{
|
||||||
|
Counter: c,
|
||||||
|
lastReset: lastReset,
|
||||||
|
needsReset: resetFn,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update updates the counter with the current users
|
||||||
|
func (c *ActiveUsersCounter) Update(users []string, now time.Time) (current uint, wasReset bool) {
|
||||||
|
if c.needsReset(c.lastReset, now) {
|
||||||
|
c.Counter.Reset()
|
||||||
|
c.lastReset = now
|
||||||
|
wasReset = true
|
||||||
|
}
|
||||||
|
for _, user := range users {
|
||||||
|
c.Mark(user)
|
||||||
|
}
|
||||||
|
return c.Count(), wasReset
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetLastReset returns the last time the counter was reset
|
||||||
|
func (c *ActiveUsersCounter) GetLastReset() time.Time {
|
||||||
|
return c.lastReset
|
||||||
|
}
|
37
internal/zero/analytics/activeusers_test.go
Normal file
37
internal/zero/analytics/activeusers_test.go
Normal file
|
@ -0,0 +1,37 @@
|
||||||
|
package analytics_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/internal/zero/analytics"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestActiveUsers(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
startTime := time.Now().UTC()
|
||||||
|
|
||||||
|
// Create a new counter that resets on a daily interval
|
||||||
|
c := analytics.NewActiveUsersCounter(analytics.ResetDailyUTC, startTime)
|
||||||
|
|
||||||
|
count, wasReset := c.Update([]string{"user1", "user2"}, startTime.Add(time.Minute))
|
||||||
|
assert.False(t, wasReset)
|
||||||
|
assert.EqualValues(t, 2, count)
|
||||||
|
|
||||||
|
count, wasReset = c.Update([]string{"user1", "user2", "user3"}, startTime.Add(time.Minute*2))
|
||||||
|
assert.False(t, wasReset)
|
||||||
|
assert.EqualValues(t, 3, count)
|
||||||
|
|
||||||
|
// Update the counter with a new user after lapse
|
||||||
|
count, wasReset = c.Update([]string{"user1", "user2", "user3", "user4"}, startTime.Add(time.Hour*25))
|
||||||
|
assert.True(t, wasReset)
|
||||||
|
assert.EqualValues(t, 4, count)
|
||||||
|
|
||||||
|
// Update the counter with a new user after lapse
|
||||||
|
count, wasReset = c.Update([]string{"user4"}, startTime.Add(time.Hour*25*2))
|
||||||
|
assert.True(t, wasReset)
|
||||||
|
assert.EqualValues(t, 1, count)
|
||||||
|
}
|
131
internal/zero/analytics/collector.go
Normal file
131
internal/zero/analytics/collector.go
Normal file
|
@ -0,0 +1,131 @@
|
||||||
|
// Package analytics collects active user metrics and reports them to the cloud dashboard
|
||||||
|
package analytics
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/internal/log"
|
||||||
|
"github.com/pomerium/pomerium/pkg/grpc/databroker"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Collect collects metrics and reports them to the cloud
|
||||||
|
func Collect(
|
||||||
|
ctx context.Context,
|
||||||
|
client databroker.DataBrokerServiceClient,
|
||||||
|
updateInterval time.Duration,
|
||||||
|
) error {
|
||||||
|
c := &collector{
|
||||||
|
client: client,
|
||||||
|
counters: make(map[string]*ActiveUsersCounter),
|
||||||
|
updateInterval: updateInterval,
|
||||||
|
}
|
||||||
|
|
||||||
|
leaser := databroker.NewLeaser("pomerium-zero-analytics", c.leaseTTL(), c)
|
||||||
|
return leaser.Run(ctx)
|
||||||
|
}
|
||||||
|
|
||||||
|
type collector struct {
|
||||||
|
client databroker.DataBrokerServiceClient
|
||||||
|
counters map[string]*ActiveUsersCounter
|
||||||
|
updateInterval time.Duration
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) RunLeased(ctx context.Context) error {
|
||||||
|
err := c.loadCounters(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to load counters: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = c.runPeriodicUpdate(ctx)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to run periodic update: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) GetDataBrokerServiceClient() databroker.DataBrokerServiceClient {
|
||||||
|
return c.client
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) loadCounters(ctx context.Context) error {
|
||||||
|
now := time.Now()
|
||||||
|
for key, resetFn := range map[string]IntervalResetFunc{
|
||||||
|
"mau": ResetMonthlyUTC,
|
||||||
|
"dau": ResetDailyUTC,
|
||||||
|
} {
|
||||||
|
state, err := LoadMetricState(ctx, c.client, key)
|
||||||
|
if err != nil && !databroker.IsNotFound(err) {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if state == nil {
|
||||||
|
c.counters[key] = NewActiveUsersCounter(resetFn, now)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
counter, err := LoadActiveUsersCounter(state.Data, state.LastReset, resetFn)
|
||||||
|
if err != nil {
|
||||||
|
log.Ctx(ctx).Error().Err(err).Str("metric", key).Msg("failed to load metric state, resetting")
|
||||||
|
counter = NewActiveUsersCounter(resetFn, now)
|
||||||
|
}
|
||||||
|
c.counters[key] = counter
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) runPeriodicUpdate(ctx context.Context) error {
|
||||||
|
ticker := time.NewTicker(c.updateInterval)
|
||||||
|
defer ticker.Stop()
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return nil
|
||||||
|
case <-ticker.C:
|
||||||
|
if err := c.update(ctx); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) update(ctx context.Context) error {
|
||||||
|
users, err := CurrentUsers(ctx, c.client)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to get current users: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
now := time.Now()
|
||||||
|
for key, counter := range c.counters {
|
||||||
|
before := counter.Count()
|
||||||
|
after, _ := counter.Update(users, now)
|
||||||
|
if before == after {
|
||||||
|
log.Ctx(ctx).Debug().Msgf("metric %s not changed: %d", key, counter.Count())
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
log.Ctx(ctx).Debug().Msgf("metric %s updated: %d", key, counter.Count())
|
||||||
|
|
||||||
|
data, err := counter.ToBinary()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to marshal metric %s: %w", key, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = SaveMetricState(ctx, c.client, key, data, after, counter.GetLastReset())
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to save metric %s: %w", key, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *collector) leaseTTL() time.Duration {
|
||||||
|
const defaultTTL = time.Minute * 5
|
||||||
|
if defaultTTL < c.updateInterval {
|
||||||
|
return defaultTTL
|
||||||
|
}
|
||||||
|
return c.updateInterval
|
||||||
|
}
|
53
internal/zero/analytics/sessions.go
Normal file
53
internal/zero/analytics/sessions.go
Normal file
|
@ -0,0 +1,53 @@
|
||||||
|
package analytics
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/internal/sets"
|
||||||
|
"github.com/pomerium/pomerium/pkg/grpc/databroker"
|
||||||
|
"github.com/pomerium/pomerium/pkg/grpc/session"
|
||||||
|
"github.com/pomerium/pomerium/pkg/protoutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
sessionTypeURL = protoutil.GetTypeURL(new(session.Session))
|
||||||
|
)
|
||||||
|
|
||||||
|
// CurrentUsers returns a list of users active within the current UTC day
|
||||||
|
func CurrentUsers(
|
||||||
|
ctx context.Context,
|
||||||
|
client databroker.DataBrokerServiceClient,
|
||||||
|
) ([]string, error) {
|
||||||
|
records, _, _, err := databroker.InitialSync(ctx, client, &databroker.SyncLatestRequest{
|
||||||
|
Type: sessionTypeURL,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("fetching sessions: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
users := sets.NewHash[string]()
|
||||||
|
utcNow := time.Now().UTC()
|
||||||
|
threshold := time.Date(utcNow.Year(), utcNow.Month(), utcNow.Day(), 0, 0, 0, 0, time.UTC)
|
||||||
|
|
||||||
|
for _, record := range records {
|
||||||
|
var s session.Session
|
||||||
|
err := record.GetData().UnmarshalTo(&s)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("unmarshaling session: %w", err)
|
||||||
|
}
|
||||||
|
if s.UserId == "" { // session creation is in progress
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if s.AccessedAt == nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if s.AccessedAt.AsTime().Before(threshold) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
users.Add(s.UserId)
|
||||||
|
}
|
||||||
|
|
||||||
|
return users.Items(), nil
|
||||||
|
}
|
126
internal/zero/analytics/storage.go
Normal file
126
internal/zero/analytics/storage.go
Normal file
|
@ -0,0 +1,126 @@
|
||||||
|
package analytics
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"encoding/base64"
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"google.golang.org/protobuf/types/known/anypb"
|
||||||
|
"google.golang.org/protobuf/types/known/structpb"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/pkg/grpc/databroker"
|
||||||
|
"github.com/pomerium/pomerium/pkg/protoutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
metricStateTypeURL = "pomerium.io/ActiveUsersMetricState"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SaveMetricState saves the state of a metric to the databroker
|
||||||
|
func SaveMetricState(
|
||||||
|
ctx context.Context,
|
||||||
|
client databroker.DataBrokerServiceClient,
|
||||||
|
id string,
|
||||||
|
data []byte,
|
||||||
|
value uint,
|
||||||
|
lastReset time.Time,
|
||||||
|
) error {
|
||||||
|
_, err := client.Put(ctx, &databroker.PutRequest{
|
||||||
|
Records: []*databroker.Record{{
|
||||||
|
Type: metricStateTypeURL,
|
||||||
|
Id: id,
|
||||||
|
Data: (&MetricState{
|
||||||
|
Data: data,
|
||||||
|
LastReset: lastReset,
|
||||||
|
Count: value,
|
||||||
|
}).ToAny(),
|
||||||
|
}},
|
||||||
|
})
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// LoadMetricState loads the state of a metric from the databroker
|
||||||
|
func LoadMetricState(
|
||||||
|
ctx context.Context, client databroker.DataBrokerServiceClient, id string,
|
||||||
|
) (*MetricState, error) {
|
||||||
|
resp, err := client.Get(ctx, &databroker.GetRequest{
|
||||||
|
Type: metricStateTypeURL,
|
||||||
|
Id: id,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("load metric state: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
var state MetricState
|
||||||
|
err = state.FromAny(resp.GetRecord().GetData())
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("load metric state: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return &state, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// MetricState is the persistent state of a metric
|
||||||
|
type MetricState struct {
|
||||||
|
Data []byte
|
||||||
|
LastReset time.Time
|
||||||
|
Count uint
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
countKey = "count"
|
||||||
|
dataKey = "data"
|
||||||
|
lastResetKey = "last_reset"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ToAny marshals a MetricState into an anypb.Any
|
||||||
|
func (r *MetricState) ToAny() *anypb.Any {
|
||||||
|
return protoutil.NewAny(&structpb.Struct{
|
||||||
|
Fields: map[string]*structpb.Value{
|
||||||
|
countKey: structpb.NewNumberValue(float64(r.Count)),
|
||||||
|
dataKey: structpb.NewStringValue(base64.StdEncoding.EncodeToString(r.Data)),
|
||||||
|
lastResetKey: structpb.NewStringValue(r.LastReset.Format(time.RFC3339)),
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// FromAny unmarshals an anypb.Any into a MetricState
|
||||||
|
func (r *MetricState) FromAny(any *anypb.Any) error {
|
||||||
|
var s structpb.Struct
|
||||||
|
err := any.UnmarshalTo(&s)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("unmarshal struct: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
vData, ok := s.GetFields()[dataKey]
|
||||||
|
if !ok {
|
||||||
|
return fmt.Errorf("missing %s field", dataKey)
|
||||||
|
}
|
||||||
|
data, err := base64.StdEncoding.DecodeString(vData.GetStringValue())
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("decode state: %w", err)
|
||||||
|
}
|
||||||
|
if len(data) == 0 {
|
||||||
|
return fmt.Errorf("empty data")
|
||||||
|
}
|
||||||
|
|
||||||
|
vLastReset, ok := s.GetFields()[lastResetKey]
|
||||||
|
if !ok {
|
||||||
|
return fmt.Errorf("missing %s field", lastResetKey)
|
||||||
|
}
|
||||||
|
lastReset, err := time.Parse(time.RFC3339, vLastReset.GetStringValue())
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("parse last reset: %w", err)
|
||||||
|
}
|
||||||
|
vCount, ok := s.GetFields()[countKey]
|
||||||
|
if !ok {
|
||||||
|
return fmt.Errorf("missing %s field", countKey)
|
||||||
|
}
|
||||||
|
|
||||||
|
r.Data = data
|
||||||
|
r.LastReset = lastReset
|
||||||
|
r.Count = uint(vCount.GetNumberValue())
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
29
internal/zero/analytics/storage_test.go
Normal file
29
internal/zero/analytics/storage_test.go
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
package analytics_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
|
||||||
|
"github.com/pomerium/pomerium/internal/zero/analytics"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestStorage(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
now := time.Date(2020, 1, 2, 3, 4, 5, 6, time.UTC)
|
||||||
|
state := &analytics.MetricState{
|
||||||
|
Data: []byte("data"),
|
||||||
|
LastReset: now,
|
||||||
|
}
|
||||||
|
|
||||||
|
pbany := state.ToAny()
|
||||||
|
assert.NotNil(t, pbany)
|
||||||
|
|
||||||
|
var newState analytics.MetricState
|
||||||
|
err := newState.FromAny(pbany)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.EqualValues(t, state.Data, newState.Data)
|
||||||
|
assert.EqualValues(t, state.LastReset.Truncate(time.Second), newState.LastReset.Truncate(time.Second))
|
||||||
|
}
|
|
@ -5,11 +5,13 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/rs/zerolog"
|
"github.com/rs/zerolog"
|
||||||
"golang.org/x/sync/errgroup"
|
"golang.org/x/sync/errgroup"
|
||||||
|
|
||||||
"github.com/pomerium/pomerium/internal/log"
|
"github.com/pomerium/pomerium/internal/log"
|
||||||
|
"github.com/pomerium/pomerium/internal/zero/analytics"
|
||||||
"github.com/pomerium/pomerium/internal/zero/bootstrap"
|
"github.com/pomerium/pomerium/internal/zero/bootstrap"
|
||||||
"github.com/pomerium/pomerium/internal/zero/reconciler"
|
"github.com/pomerium/pomerium/internal/zero/reconciler"
|
||||||
"github.com/pomerium/pomerium/pkg/cmd/pomerium"
|
"github.com/pomerium/pomerium/pkg/cmd/pomerium"
|
||||||
|
@ -43,6 +45,7 @@ func Run(ctx context.Context, opts ...Option) error {
|
||||||
eg.Go(func() error { return run(ctx, "pomerium-core", c.runPomeriumCore, src.WaitReady) })
|
eg.Go(func() error { return run(ctx, "pomerium-core", c.runPomeriumCore, src.WaitReady) })
|
||||||
eg.Go(func() error { return run(ctx, "zero-reconciler", c.runReconciler, src.WaitReady) })
|
eg.Go(func() error { return run(ctx, "zero-reconciler", c.runReconciler, src.WaitReady) })
|
||||||
eg.Go(func() error { return run(ctx, "connect-log", c.RunConnectLog, nil) })
|
eg.Go(func() error { return run(ctx, "connect-log", c.RunConnectLog, nil) })
|
||||||
|
eg.Go(func() error { return run(ctx, "zero-analytics", c.runAnalytics, src.WaitReady) })
|
||||||
return eg.Wait()
|
return eg.Wait()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -117,3 +120,13 @@ func (c *controller) runReconciler(ctx context.Context) error {
|
||||||
reconciler.WithDataBrokerClient(c.GetDataBrokerServiceClient()),
|
reconciler.WithDataBrokerClient(c.GetDataBrokerServiceClient()),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *controller) runAnalytics(ctx context.Context) error {
|
||||||
|
err := analytics.Collect(ctx, c.GetDataBrokerServiceClient(), time.Second*30)
|
||||||
|
if err != nil && ctx.Err() == nil {
|
||||||
|
log.Ctx(ctx).Error().Err(err).Msg("error collecting analytics, disabling")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
|
@ -8,6 +8,8 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
|
|
||||||
|
"google.golang.org/grpc/codes"
|
||||||
|
status "google.golang.org/grpc/status"
|
||||||
"google.golang.org/protobuf/encoding/protojson"
|
"google.golang.org/protobuf/encoding/protojson"
|
||||||
"google.golang.org/protobuf/proto"
|
"google.golang.org/protobuf/proto"
|
||||||
structpb "google.golang.org/protobuf/types/known/structpb"
|
structpb "google.golang.org/protobuf/types/known/structpb"
|
||||||
|
@ -33,6 +35,11 @@ func NewRecord(object recordObject) *Record {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsNotFound returns true if the error is a not found error.
|
||||||
|
func IsNotFound(err error) bool {
|
||||||
|
return status.Code(err) == codes.NotFound
|
||||||
|
}
|
||||||
|
|
||||||
// Get gets a record from the databroker and unmarshals it into the object.
|
// Get gets a record from the databroker and unmarshals it into the object.
|
||||||
func Get(ctx context.Context, client DataBrokerServiceClient, object recordObject) error {
|
func Get(ctx context.Context, client DataBrokerServiceClient, object recordObject) error {
|
||||||
res, err := client.Get(ctx, &GetRequest{
|
res, err := client.Get(ctx, &GetRequest{
|
||||||
|
|
Loading…
Add table
Reference in a new issue