2022-12-08 14:27:21 +00:00
|
|
|
// Copyright (C) 2022 Storj Labs, Inc.
|
|
|
|
// See LICENSE for copying information.
|
|
|
|
|
|
|
|
package rangedloop_test
|
|
|
|
|
|
|
|
import (
|
2023-01-05 15:32:30 +00:00
|
|
|
"context"
|
2023-01-11 21:23:17 +00:00
|
|
|
"errors"
|
2022-12-08 14:27:21 +00:00
|
|
|
"fmt"
|
2023-01-11 21:23:17 +00:00
|
|
|
"sync/atomic"
|
2022-12-08 14:27:21 +00:00
|
|
|
"testing"
|
2022-12-21 20:58:08 +00:00
|
|
|
"time"
|
2022-12-08 14:27:21 +00:00
|
|
|
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"go.uber.org/zap/zaptest"
|
|
|
|
|
2022-12-21 20:58:08 +00:00
|
|
|
"storj.io/common/testcontext"
|
|
|
|
"storj.io/common/uuid"
|
2022-12-08 14:27:21 +00:00
|
|
|
"storj.io/storj/satellite/metabase/rangedloop"
|
|
|
|
"storj.io/storj/satellite/metabase/rangedloop/rangedlooptest"
|
|
|
|
"storj.io/storj/satellite/metabase/segmentloop"
|
|
|
|
)
|
|
|
|
|
2022-12-21 20:58:08 +00:00
|
|
|
func TestLoopCount(t *testing.T) {
|
2022-12-08 14:27:21 +00:00
|
|
|
for _, parallelism := range []int{1, 2, 3} {
|
|
|
|
for _, nSegments := range []int{0, 1, 2, 11} {
|
|
|
|
for _, nObservers := range []int{0, 1, 2} {
|
|
|
|
t.Run(
|
|
|
|
fmt.Sprintf("par%d_seg%d_obs%d", parallelism, nSegments, nObservers),
|
|
|
|
func(t *testing.T) {
|
2022-12-21 20:58:08 +00:00
|
|
|
runCountTest(t, parallelism, nSegments, nObservers)
|
2022-12-08 14:27:21 +00:00
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-12-21 20:58:08 +00:00
|
|
|
func runCountTest(t *testing.T, parallelism int, nSegments int, nObservers int) {
|
2022-12-08 14:27:21 +00:00
|
|
|
batchSize := 2
|
2022-12-21 20:58:08 +00:00
|
|
|
ctx := testcontext.New(t)
|
2022-12-08 14:27:21 +00:00
|
|
|
|
|
|
|
observers := []rangedloop.Observer{}
|
|
|
|
for i := 0; i < nObservers; i++ {
|
|
|
|
observers = append(observers, &rangedlooptest.CountObserver{})
|
|
|
|
}
|
|
|
|
|
|
|
|
loopService := rangedloop.NewService(
|
|
|
|
zaptest.NewLogger(t),
|
|
|
|
rangedloop.Config{
|
2022-12-21 20:58:08 +00:00
|
|
|
BatchSize: batchSize,
|
|
|
|
Parallelism: parallelism,
|
2022-12-08 14:27:21 +00:00
|
|
|
},
|
2022-12-09 15:40:23 +00:00
|
|
|
&rangedlooptest.RangeSplitter{
|
2022-12-08 14:27:21 +00:00
|
|
|
Segments: make([]segmentloop.Segment, nSegments),
|
|
|
|
},
|
|
|
|
observers,
|
|
|
|
)
|
|
|
|
|
2022-12-21 20:58:08 +00:00
|
|
|
observerDurations, err := loopService.RunOnce(ctx)
|
2022-12-08 14:27:21 +00:00
|
|
|
require.NoError(t, err)
|
2022-12-21 20:58:08 +00:00
|
|
|
require.Len(t, observerDurations, nObservers)
|
2022-12-08 14:27:21 +00:00
|
|
|
|
|
|
|
for _, observer := range observers {
|
|
|
|
countObserver := observer.(*rangedlooptest.CountObserver)
|
|
|
|
require.Equal(t, nSegments, countObserver.NumSegments)
|
|
|
|
}
|
|
|
|
}
|
2022-12-21 20:58:08 +00:00
|
|
|
|
|
|
|
func TestLoopDuration(t *testing.T) {
|
|
|
|
t.Skip("Flaky test because it validates concurrency by measuring time")
|
|
|
|
|
|
|
|
nSegments := 8
|
|
|
|
nObservers := 2
|
|
|
|
parallelism := 4
|
|
|
|
batchSize := 2
|
|
|
|
sleepIncrement := time.Millisecond * 10
|
|
|
|
|
|
|
|
ctx := testcontext.New(t)
|
|
|
|
|
|
|
|
observers := []rangedloop.Observer{}
|
|
|
|
for i := 0; i < nObservers; i++ {
|
|
|
|
observers = append(observers, &rangedlooptest.SleepObserver{
|
|
|
|
Duration: sleepIncrement,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
segments := []segmentloop.Segment{}
|
|
|
|
for i := 0; i < nSegments; i++ {
|
|
|
|
streamId, err := uuid.FromBytes([]byte{byte(i), 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
|
|
|
|
require.NoError(t, err)
|
|
|
|
segments = append(segments, segmentloop.Segment{
|
|
|
|
StreamID: streamId,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
loopService := rangedloop.NewService(
|
|
|
|
zaptest.NewLogger(t),
|
|
|
|
rangedloop.Config{
|
|
|
|
BatchSize: batchSize,
|
|
|
|
Parallelism: parallelism,
|
|
|
|
},
|
|
|
|
&rangedlooptest.RangeSplitter{
|
|
|
|
Segments: segments,
|
|
|
|
},
|
|
|
|
observers,
|
|
|
|
)
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
observerDurations, err := loopService.RunOnce(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
duration := time.Since(start)
|
|
|
|
expectedDuration := time.Duration(int64(nSegments) * int64(sleepIncrement) * int64(nObservers) / int64(parallelism))
|
|
|
|
require.Equal(t, expectedDuration, duration.Truncate(sleepIncrement))
|
|
|
|
|
|
|
|
require.Len(t, observerDurations, nObservers)
|
|
|
|
for _, observerDuration := range observerDurations {
|
|
|
|
expectedSleep := time.Duration(int64(nSegments) * int64(sleepIncrement))
|
|
|
|
require.Equal(t, expectedSleep, observerDuration.Duration.Round(sleepIncrement))
|
|
|
|
}
|
|
|
|
}
|
2023-01-05 15:32:30 +00:00
|
|
|
|
|
|
|
func TestLoopCancellation(t *testing.T) {
|
|
|
|
parallelism := 2
|
|
|
|
batchSize := 1
|
|
|
|
ctx, cancel := context.WithCancel(testcontext.NewWithTimeout(t, time.Second*10))
|
|
|
|
|
|
|
|
observers := []rangedloop.Observer{
|
|
|
|
&rangedlooptest.CountObserver{},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
// cancel from inside the loop, when it is certain that the loop has started
|
|
|
|
cancel()
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
loopService := rangedloop.NewService(
|
|
|
|
zaptest.NewLogger(t),
|
|
|
|
rangedloop.Config{
|
|
|
|
BatchSize: batchSize,
|
|
|
|
Parallelism: parallelism,
|
|
|
|
},
|
|
|
|
&rangedlooptest.InfiniteSegmentProvider{},
|
|
|
|
observers,
|
|
|
|
)
|
|
|
|
|
|
|
|
_, err := loopService.RunOnce(ctx)
|
|
|
|
|
|
|
|
require.ErrorIs(t, err, context.Canceled)
|
|
|
|
}
|
2023-01-11 21:23:17 +00:00
|
|
|
|
|
|
|
func TestLoopContinuesAfterObserverError(t *testing.T) {
|
|
|
|
parallelism := 2
|
|
|
|
batchSize := 1
|
|
|
|
segments := make([]segmentloop.Segment, 2)
|
|
|
|
|
|
|
|
numOnStartCalls := 0
|
|
|
|
numOnForkCalls := 0
|
|
|
|
numOnProcessCalls := int32(0)
|
|
|
|
numOnJoinCalls := 0
|
|
|
|
numOnFinishCalls := 0
|
|
|
|
|
|
|
|
incNumOnProcessCalls := func() {
|
|
|
|
atomic.AddInt32(&numOnProcessCalls, 1)
|
|
|
|
}
|
|
|
|
|
|
|
|
// first and last observer emit no error
|
|
|
|
// other observers emit an error at different stages
|
|
|
|
observers := []rangedloop.Observer{
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
incNumOnProcessCalls()
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
numOnJoinCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
numOnFinishCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return errors.New("Test OnStart error")
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
require.Fail(t, "OnFork should not be called")
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
require.Fail(t, "OnProcess should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
require.Fail(t, "OnJoin should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
require.Fail(t, "OnFinish should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, errors.New("Test OnFork error")
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
require.Fail(t, "OnProcess should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
require.Fail(t, "OnJoin should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
require.Fail(t, "OnFinish should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
incNumOnProcessCalls()
|
|
|
|
return errors.New("Test OnProcess error")
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
require.Fail(t, "OnJoin should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
require.Fail(t, "OnFinish should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
incNumOnProcessCalls()
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
numOnJoinCalls++
|
|
|
|
return errors.New("Test OnJoin error")
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
require.Fail(t, "OnFinish should not be called")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
incNumOnProcessCalls()
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
numOnJoinCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
numOnFinishCalls++
|
|
|
|
return errors.New("Test OnFinish error")
|
|
|
|
},
|
|
|
|
},
|
|
|
|
&rangedlooptest.CallbackObserver{
|
|
|
|
OnStart: func(ctx context.Context, t time.Time) error {
|
|
|
|
numOnStartCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFork: func(ctx context.Context) (rangedloop.Partial, error) {
|
|
|
|
numOnForkCalls++
|
|
|
|
return nil, nil
|
|
|
|
},
|
|
|
|
OnProcess: func(ctx context.Context, segments []segmentloop.Segment) error {
|
|
|
|
incNumOnProcessCalls()
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnJoin: func(ctx context.Context, partial rangedloop.Partial) error {
|
|
|
|
numOnJoinCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OnFinish: func(ctx context.Context) error {
|
|
|
|
numOnFinishCalls++
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
loopService := rangedloop.NewService(
|
|
|
|
zaptest.NewLogger(t),
|
|
|
|
rangedloop.Config{
|
|
|
|
BatchSize: batchSize,
|
|
|
|
Parallelism: parallelism,
|
|
|
|
},
|
|
|
|
&rangedlooptest.RangeSplitter{
|
|
|
|
Segments: segments,
|
|
|
|
},
|
|
|
|
observers,
|
|
|
|
)
|
|
|
|
|
|
|
|
observerDurations, err := loopService.RunOnce(testcontext.New(t))
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, observerDurations, len(observers))
|
|
|
|
|
|
|
|
require.EqualValues(t, 7, numOnStartCalls)
|
|
|
|
require.EqualValues(t, 6*parallelism, numOnForkCalls)
|
|
|
|
require.EqualValues(t, 5*parallelism-1, numOnProcessCalls)
|
|
|
|
require.EqualValues(t, 4*parallelism-1, numOnJoinCalls)
|
|
|
|
require.EqualValues(t, 3, numOnFinishCalls)
|
|
|
|
|
|
|
|
// success observer should have the duration reported
|
|
|
|
require.Greater(t, observerDurations[0].Duration, time.Duration(0))
|
|
|
|
require.Greater(t, observerDurations[6].Duration, time.Duration(0))
|
|
|
|
|
|
|
|
// error observers should have sentinel duration reported
|
|
|
|
require.Equal(t, observerDurations[1].Duration, -1*time.Second)
|
|
|
|
require.Equal(t, observerDurations[2].Duration, -1*time.Second)
|
|
|
|
require.Equal(t, observerDurations[3].Duration, -1*time.Second)
|
|
|
|
require.Equal(t, observerDurations[4].Duration, -1*time.Second)
|
|
|
|
require.Equal(t, observerDurations[5].Duration, -1*time.Second)
|
|
|
|
}
|