1f5fbbd24a
It's possible that single object will have a lot of segments and at the moment we are trying to collect all pieces at once and send to storage nodes information about deletion. Such approach my lead to using extensive amount of memory. This change is handling this problem by calling DeletePieces function multiple times with only part of pieces to delete for a single call. Change-Id: Ie1e66cd9d86d130eb89a61cf6e23f38b8cb8859e
146 lines
4.0 KiB
Go
146 lines
4.0 KiB
Go
// Copyright (C) 2020 Storj Labs, Inc.
|
|
// See LICENSE for copying information.
|
|
|
|
package metabase
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"errors"
|
|
|
|
"storj.io/common/uuid"
|
|
"storj.io/private/dbutil"
|
|
"storj.io/private/tagsql"
|
|
)
|
|
|
|
const deleteBatchSizeLimit = 100
|
|
|
|
const deletePieceBatchLimit = 1000
|
|
|
|
// DeleteBucketObjects contains arguments for deleting a whole bucket.
|
|
type DeleteBucketObjects struct {
|
|
Bucket BucketLocation
|
|
BatchSize int
|
|
|
|
// DeletePiecesBatchSize maximum number of DeletedSegmentInfo entries
|
|
// passed to DeletePieces function at once.
|
|
DeletePiecesBatchSize int
|
|
// DeletePieces is called for every batch of objects.
|
|
// Slice `segments` will be reused between calls.
|
|
DeletePieces func(ctx context.Context, segments []DeletedSegmentInfo) error
|
|
}
|
|
|
|
// DeleteBucketObjects deletes all objects in the specified bucket.
|
|
func (db *DB) DeleteBucketObjects(ctx context.Context, opts DeleteBucketObjects) (deletedObjectCount int64, err error) {
|
|
defer mon.Task()(&ctx)(&err)
|
|
|
|
if err := opts.Bucket.Verify(); err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
batchSize := opts.BatchSize
|
|
if batchSize <= 0 || batchSize > deleteBatchSizeLimit {
|
|
batchSize = deleteBatchSizeLimit
|
|
}
|
|
|
|
deletePiecesBatchSize := opts.DeletePiecesBatchSize
|
|
if deletePiecesBatchSize <= 0 || deletePiecesBatchSize > deletePieceBatchLimit {
|
|
deletePiecesBatchSize = deletePieceBatchLimit
|
|
}
|
|
|
|
var query string
|
|
switch db.impl {
|
|
case dbutil.Cockroach:
|
|
query = `
|
|
WITH deleted_objects AS (
|
|
DELETE FROM objects
|
|
WHERE project_id = $1 AND bucket_name = $2 LIMIT $3
|
|
RETURNING objects.stream_id
|
|
)
|
|
DELETE FROM segments
|
|
WHERE segments.stream_id in (SELECT deleted_objects.stream_id FROM deleted_objects)
|
|
RETURNING segments.stream_id, segments.root_piece_id, segments.remote_alias_pieces
|
|
`
|
|
case dbutil.Postgres:
|
|
query = `
|
|
WITH deleted_objects AS (
|
|
DELETE FROM objects
|
|
WHERE stream_id IN (
|
|
SELECT stream_id FROM objects
|
|
WHERE project_id = $1 AND bucket_name = $2
|
|
LIMIT $3
|
|
)
|
|
RETURNING objects.stream_id
|
|
)
|
|
DELETE FROM segments
|
|
WHERE segments.stream_id in (SELECT deleted_objects.stream_id FROM deleted_objects)
|
|
RETURNING segments.stream_id, segments.root_piece_id, segments.remote_alias_pieces
|
|
`
|
|
default:
|
|
return 0, Error.New("unhandled database: %v", db.impl)
|
|
}
|
|
|
|
// TODO: fix the count for objects without segments
|
|
deletedSegmentsBatch := make([]DeletedSegmentInfo, 0, deletePiecesBatchSize)
|
|
for {
|
|
deletedSegmentsBatch = deletedSegmentsBatch[:0]
|
|
batchDeletedObjects := 0
|
|
deletedSegments := 0
|
|
err = withRows(db.db.Query(ctx, query,
|
|
opts.Bucket.ProjectID, []byte(opts.Bucket.BucketName), batchSize))(func(rows tagsql.Rows) error {
|
|
ids := map[uuid.UUID]struct{}{} // TODO: avoid map here
|
|
for rows.Next() {
|
|
var streamID uuid.UUID
|
|
var segment DeletedSegmentInfo
|
|
var aliasPieces AliasPieces
|
|
err := rows.Scan(&streamID, &segment.RootPieceID, &aliasPieces)
|
|
if err != nil {
|
|
return Error.Wrap(err)
|
|
}
|
|
segment.Pieces, err = db.aliasCache.ConvertAliasesToPieces(ctx, aliasPieces)
|
|
if err != nil {
|
|
return Error.Wrap(err)
|
|
}
|
|
|
|
ids[streamID] = struct{}{}
|
|
deletedSegmentsBatch = append(deletedSegmentsBatch, segment)
|
|
|
|
if len(deletedSegmentsBatch) == deletePiecesBatchSize {
|
|
if opts.DeletePieces != nil {
|
|
err = opts.DeletePieces(ctx, deletedSegmentsBatch)
|
|
if err != nil {
|
|
return Error.Wrap(err)
|
|
}
|
|
deletedSegmentsBatch = deletedSegmentsBatch[:0]
|
|
}
|
|
}
|
|
|
|
deletedSegments++
|
|
}
|
|
batchDeletedObjects = len(ids)
|
|
deletedObjectCount += int64(len(ids))
|
|
return nil
|
|
})
|
|
|
|
mon.Meter("object_delete").Mark(batchDeletedObjects)
|
|
mon.Meter("segment_delete").Mark(deletedSegments)
|
|
|
|
if err != nil {
|
|
if errors.Is(err, sql.ErrNoRows) {
|
|
return deletedObjectCount, nil
|
|
}
|
|
return deletedObjectCount, Error.Wrap(err)
|
|
}
|
|
if len(deletedSegmentsBatch) == 0 {
|
|
return deletedObjectCount, nil
|
|
}
|
|
|
|
if opts.DeletePieces != nil {
|
|
err = opts.DeletePieces(ctx, deletedSegmentsBatch)
|
|
if err != nil {
|
|
return deletedObjectCount, Error.Wrap(err)
|
|
}
|
|
}
|
|
}
|
|
}
|