satellite/metabase/rangedloop: measure observer duration (#5350)

Track duration of all segment loop observers. Factor out functions to
reduce size.

Still need to send the measurements out via monkit.

Part of https://github.com/storj/storj/issues/5223

Change-Id: Iae0260e250f8ea33affed95c6592a1f42df384eb
This commit is contained in:
Erik van Velzen 2022-12-21 21:58:08 +01:00 committed by GitHub
parent cda1d67465
commit 37b4981cc0
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 213 additions and 66 deletions

View File

@ -0,0 +1,45 @@
// Copyright (C) 2022 Storj Labs, Inc.
// See LICENSE for copying information.
package rangedlooptest
import (
"context"
"time"
"storj.io/storj/satellite/metabase/rangedloop"
"storj.io/storj/satellite/metabase/segmentloop"
)
// SleepObserver is a subscriber to the segment loop which sleeps for every batch.
type SleepObserver struct {
Duration time.Duration
}
// Start is the callback for segment loop start.
func (c *SleepObserver) Start(ctx context.Context, time time.Time) error {
return nil
}
// Fork splits the observer to process a segment range.
func (c *SleepObserver) Fork(ctx context.Context) (rangedloop.Partial, error) {
return c, nil
}
// Join is a noop.
func (c *SleepObserver) Join(ctx context.Context, partial rangedloop.Partial) error {
// Range done
return nil
}
// Finish is the callback for segment loop end.
func (c *SleepObserver) Finish(ctx context.Context) error {
return nil
}
// Process sleeps for every batch of segments to simulate execution time.
func (c *SleepObserver) Process(ctx context.Context, segments []segmentloop.Segment) error {
sleepTime := time.Duration(c.Duration.Nanoseconds() * int64(len(segments)))
time.Sleep(sleepTime)
return nil
}

View File

@ -50,7 +50,18 @@ func NewService(log *zap.Logger, config Config, provider RangeSplitter, observer
// Improvement: track duration.
type observerState struct {
observer Observer
rangeObservers []Partial
rangeObservers []*rangeObserverState
}
type rangeObserverState struct {
rangeObserver Partial
duration time.Duration
}
// ObserverDuration reports back on how long it took the observer to process all the segments.
type ObserverDuration struct {
Observer Observer
Duration time.Duration
}
// Run starts the looping service.
@ -58,7 +69,8 @@ func (service *Service) Run(ctx context.Context) (err error) {
defer mon.Task()(&ctx)(&err)
for {
if err := service.RunOnce(ctx); err != nil {
_, err := service.RunOnce(ctx)
if err != nil {
service.log.Error("ranged loop failure", zap.Error(err))
if errs2.IsCanceled(err) {
@ -74,81 +86,115 @@ func (service *Service) Run(ctx context.Context) (err error) {
}
// RunOnce goes through one time and sends information to observers.
func (service *Service) RunOnce(ctx context.Context) (err error) {
func (service *Service) RunOnce(ctx context.Context) (observerDurations []ObserverDuration, err error) {
defer mon.Task()(&ctx)(&err)
startTime := time.Now()
observerStates := []observerState{}
for _, obs := range service.observers {
err := obs.Start(ctx, startTime)
observerStates, err := startObservers(ctx, service.observers)
if err != nil {
return err
}
observerStates = append(observerStates, observerState{
observer: obs,
})
return nil, err
}
rangeProviders, err := service.provider.CreateRanges(service.config.Parallelism, service.config.BatchSize)
if err != nil {
return err
return nil, err
}
group := errs2.Group{}
for rangeIndex, rangeProvider := range rangeProviders {
rangeObservers := []Partial{}
for _, rangeProvider := range rangeProviders {
rangeObservers := []*rangeObserverState{}
for i, observerState := range observerStates {
rangeObserver, err := observerState.observer.Fork(ctx)
if err != nil {
return err
return nil, err
}
rangeObservers = append(rangeObservers, rangeObserver)
observerStates[i].rangeObservers = append(observerStates[i].rangeObservers, rangeObserver)
rangeState := &rangeObserverState{
rangeObserver: rangeObserver,
}
rangeObservers = append(rangeObservers, rangeState)
observerStates[i].rangeObservers = append(observerStates[i].rangeObservers, rangeState)
}
// Create closure to capture loop variables.
createClosure := func(ctx context.Context, rangeIndex int, rangeProvider SegmentProvider, rangeObservers []Partial) func() error {
return func() (err error) {
defer mon.Task()(&ctx)(&err)
return rangeProvider.Iterate(ctx, func(segments []segmentloop.Segment) error {
for _, rangeObserver := range rangeObservers {
err := rangeObserver.Process(ctx, segments)
if err != nil {
return err
}
}
return nil
})
}
}
group.Go(createClosure(ctx, rangeIndex, rangeProvider, rangeObservers))
group.Go(createGoroutineClosure(ctx, rangeProvider, rangeObservers))
}
// Improvement: stop all ranges when one has an error.
errList := group.Wait()
if errList != nil {
return errs.Combine(errList...)
return nil, errs.Combine(errList...)
}
// Segment loop has ended.
// This is the reduce step.
for _, state := range observerStates {
for _, rangeObserver := range state.rangeObservers {
err := state.observer.Join(ctx, rangeObserver)
return finishObservers(ctx, observerStates)
}
func createGoroutineClosure(ctx context.Context, rangeProvider SegmentProvider, states []*rangeObserverState) func() error {
return func() (err error) {
defer mon.Task()(&ctx)(&err)
return rangeProvider.Iterate(ctx, func(segments []segmentloop.Segment) error {
for _, state := range states {
start := time.Now()
err := state.rangeObserver.Process(ctx, segments)
if err != nil {
return err
}
state.duration += time.Since(start)
}
err := state.observer.Finish(ctx)
if err != nil {
return err
}
}
return nil
})
}
}
func startObservers(ctx context.Context, observers []Observer) (observerStates []observerState, err error) {
startTime := time.Now()
for _, obs := range observers {
state, err := startObserver(ctx, startTime, obs)
if err != nil {
return nil, err
}
observerStates = append(observerStates, state)
}
return observerStates, nil
}
func startObserver(ctx context.Context, startTime time.Time, observer Observer) (observerState, error) {
err := observer.Start(ctx, startTime)
return observerState{
observer: observer,
}, err
}
func finishObservers(ctx context.Context, observerStates []observerState) (observerDurations []ObserverDuration, err error) {
for _, state := range observerStates {
observerDuration, err := finishObserver(ctx, state)
if err != nil {
return nil, err
}
observerDurations = append(observerDurations, observerDuration)
}
return observerDurations, nil
}
// Iterating over the segments is done.
// This is the reduce step.
func finishObserver(ctx context.Context, state observerState) (ObserverDuration, error) {
var duration time.Duration
for _, rangeObserver := range state.rangeObservers {
err := state.observer.Join(ctx, rangeObserver.rangeObserver)
if err != nil {
return ObserverDuration{}, err
}
duration += rangeObserver.duration
}
return ObserverDuration{
Duration: duration,
Observer: state.observer,
}, state.observer.Finish(ctx)
}

View File

@ -4,26 +4,28 @@
package rangedloop_test
import (
"context"
"fmt"
"testing"
"time"
"github.com/stretchr/testify/require"
"go.uber.org/zap/zaptest"
"storj.io/common/testcontext"
"storj.io/common/uuid"
"storj.io/storj/satellite/metabase/rangedloop"
"storj.io/storj/satellite/metabase/rangedloop/rangedlooptest"
"storj.io/storj/satellite/metabase/segmentloop"
)
func TestLoop(t *testing.T) {
func TestLoopCount(t *testing.T) {
for _, parallelism := range []int{1, 2, 3} {
for _, nSegments := range []int{0, 1, 2, 11} {
for _, nObservers := range []int{0, 1, 2} {
t.Run(
fmt.Sprintf("par%d_seg%d_obs%d", parallelism, nSegments, nObservers),
func(t *testing.T) {
RunTest(t, parallelism, nSegments, nObservers)
runCountTest(t, parallelism, nSegments, nObservers)
},
)
}
@ -31,9 +33,9 @@ func TestLoop(t *testing.T) {
}
}
func RunTest(t *testing.T, parallelism int, nSegments int, nObservers int) {
func runCountTest(t *testing.T, parallelism int, nSegments int, nObservers int) {
batchSize := 2
ctx := context.Background()
ctx := testcontext.New(t)
observers := []rangedloop.Observer{}
for i := 0; i < nObservers; i++ {
@ -45,7 +47,6 @@ func RunTest(t *testing.T, parallelism int, nSegments int, nObservers int) {
rangedloop.Config{
BatchSize: batchSize,
Parallelism: parallelism,
AsOfSystemInterval: 0,
},
&rangedlooptest.RangeSplitter{
Segments: make([]segmentloop.Segment, nSegments),
@ -53,11 +54,66 @@ func RunTest(t *testing.T, parallelism int, nSegments int, nObservers int) {
observers,
)
err := loopService.RunOnce(ctx)
observerDurations, err := loopService.RunOnce(ctx)
require.NoError(t, err)
require.Len(t, observerDurations, nObservers)
for _, observer := range observers {
countObserver := observer.(*rangedlooptest.CountObserver)
require.Equal(t, nSegments, countObserver.NumSegments)
}
}
func TestLoopDuration(t *testing.T) {
t.Skip("Flaky test because it validates concurrency by measuring time")
nSegments := 8
nObservers := 2
parallelism := 4
batchSize := 2
sleepIncrement := time.Millisecond * 10
ctx := testcontext.New(t)
observers := []rangedloop.Observer{}
for i := 0; i < nObservers; i++ {
observers = append(observers, &rangedlooptest.SleepObserver{
Duration: sleepIncrement,
})
}
segments := []segmentloop.Segment{}
for i := 0; i < nSegments; i++ {
streamId, err := uuid.FromBytes([]byte{byte(i), 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
require.NoError(t, err)
segments = append(segments, segmentloop.Segment{
StreamID: streamId,
})
}
loopService := rangedloop.NewService(
zaptest.NewLogger(t),
rangedloop.Config{
BatchSize: batchSize,
Parallelism: parallelism,
},
&rangedlooptest.RangeSplitter{
Segments: segments,
},
observers,
)
start := time.Now()
observerDurations, err := loopService.RunOnce(ctx)
require.NoError(t, err)
duration := time.Since(start)
expectedDuration := time.Duration(int64(nSegments) * int64(sleepIncrement) * int64(nObservers) / int64(parallelism))
require.Equal(t, expectedDuration, duration.Truncate(sleepIncrement))
require.Len(t, observerDurations, nObservers)
for _, observerDuration := range observerDurations {
expectedSleep := time.Duration(int64(nSegments) * int64(sleepIncrement))
require.Equal(t, expectedSleep, observerDuration.Duration.Round(sleepIncrement))
}
}

View File

@ -43,7 +43,7 @@ func TestObserver(t *testing.T) {
rangedloop.Config{BatchSize: 2, Parallelism: 2},
&rangedlooptest.RangeSplitter{Segments: combineSegments(streams...)},
[]rangedloop.Observer{obs})
err := service.RunOnce(ctx)
_, err := service.RunOnce(ctx)
require.NoError(tb, err)
return obs.TestingMetrics()
}