mirror of
https://github.com/garethgeorge/backrest.git
synced 2025-12-14 17:45:36 +00:00
fix: add API test coverage and fix minor bugs
This commit is contained in:
@@ -67,6 +67,11 @@ func (s *BackrestHandler) SetConfig(ctx context.Context, req *connect.Request[v1
|
||||
if existing.Modno != req.Msg.Modno {
|
||||
return nil, errors.New("config modno mismatch, reload and try again")
|
||||
}
|
||||
|
||||
if err := config.ValidateConfig(req.Msg); err != nil {
|
||||
return nil, fmt.Errorf("validation error: %w", err)
|
||||
}
|
||||
|
||||
req.Msg.Modno += 1
|
||||
|
||||
if err := s.config.Update(req.Msg); err != nil {
|
||||
|
||||
285
internal/api/backresthandler_test.go
Normal file
285
internal/api/backresthandler_test.go
Normal file
@@ -0,0 +1,285 @@
|
||||
package api
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"slices"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"connectrpc.com/connect"
|
||||
"github.com/garethgeorge/backrest/gen/go/types"
|
||||
v1 "github.com/garethgeorge/backrest/gen/go/v1"
|
||||
"github.com/garethgeorge/backrest/internal/config"
|
||||
"github.com/garethgeorge/backrest/internal/oplog"
|
||||
"github.com/garethgeorge/backrest/internal/orchestrator"
|
||||
"github.com/garethgeorge/backrest/internal/resticinstaller"
|
||||
"github.com/garethgeorge/backrest/internal/rotatinglog"
|
||||
"google.golang.org/protobuf/proto"
|
||||
)
|
||||
|
||||
func TestUpdateConfig(t *testing.T) {
|
||||
sut := createSystemUnderTest(t, &config.MemoryStore{
|
||||
Config: &v1.Config{
|
||||
Modno: 1234,
|
||||
},
|
||||
})
|
||||
|
||||
tests := []struct {
|
||||
name string
|
||||
req *v1.Config
|
||||
wantErr bool
|
||||
res *v1.Config
|
||||
}{
|
||||
{
|
||||
name: "bad modno",
|
||||
req: &v1.Config{
|
||||
Modno: 4321,
|
||||
},
|
||||
wantErr: true,
|
||||
},
|
||||
{
|
||||
name: "good modno",
|
||||
req: &v1.Config{
|
||||
Modno: 1234,
|
||||
},
|
||||
wantErr: false,
|
||||
res: &v1.Config{
|
||||
Modno: 1235,
|
||||
},
|
||||
},
|
||||
{
|
||||
name: "reject when validation fails",
|
||||
req: &v1.Config{
|
||||
Modno: 1235,
|
||||
Repos: []*v1.Repo{
|
||||
{},
|
||||
},
|
||||
},
|
||||
wantErr: true,
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
res, err := sut.handler.SetConfig(context.Background(), connect.NewRequest(tt.req))
|
||||
if (err != nil) != tt.wantErr {
|
||||
t.Errorf("SetConfig() error = %v, wantErr %v", err, tt.wantErr)
|
||||
return
|
||||
}
|
||||
|
||||
if err == nil && !proto.Equal(res.Msg, tt.res) {
|
||||
t.Errorf("SetConfig() got = %v, want %v", res, tt.res)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestBackup(t *testing.T) {
|
||||
sut := createSystemUnderTest(t, &config.MemoryStore{
|
||||
Config: &v1.Config{
|
||||
Modno: 1234,
|
||||
Repos: []*v1.Repo{
|
||||
{
|
||||
Id: "local",
|
||||
Uri: t.TempDir(),
|
||||
Password: "test",
|
||||
},
|
||||
},
|
||||
Plans: []*v1.Plan{
|
||||
{
|
||||
Id: "test",
|
||||
Repo: "local",
|
||||
Paths: []string{
|
||||
t.TempDir(),
|
||||
},
|
||||
Cron: "0 0 1 1 *",
|
||||
Retention: &v1.RetentionPolicy{
|
||||
KeepHourly: 1,
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
go func() {
|
||||
sut.orch.Run(ctx)
|
||||
}()
|
||||
|
||||
_, err := sut.handler.Backup(context.Background(), connect.NewRequest(&types.StringValue{Value: "test"}))
|
||||
if err != nil {
|
||||
t.Fatalf("Backup() error = %v", err)
|
||||
}
|
||||
|
||||
// Check that there is a successful backup recorded in the log.
|
||||
if slices.IndexFunc(getOperations(t, sut.oplog), func(op *v1.Operation) bool {
|
||||
_, ok := op.GetOp().(*v1.Operation_OperationBackup)
|
||||
return op.Status == v1.OperationStatus_STATUS_SUCCESS && ok
|
||||
}) == -1 {
|
||||
t.Fatalf("Expected a backup operation")
|
||||
}
|
||||
|
||||
// Wait for the index snapshot operation to appear in the oplog.
|
||||
var snapshotId string
|
||||
if err := retry(t, 10, 2*time.Second, func() error {
|
||||
operations := getOperations(t, sut.oplog)
|
||||
if index := slices.IndexFunc(operations, func(op *v1.Operation) bool {
|
||||
_, ok := op.GetOp().(*v1.Operation_OperationIndexSnapshot)
|
||||
return op.Status == v1.OperationStatus_STATUS_SUCCESS && ok
|
||||
}); index != -1 {
|
||||
op := operations[index]
|
||||
snapshotId = op.SnapshotId
|
||||
return nil
|
||||
}
|
||||
return errors.New("snapshot not indexed")
|
||||
}); err != nil {
|
||||
t.Fatalf("Couldn't find snapshot in oplog")
|
||||
}
|
||||
|
||||
if snapshotId == "" {
|
||||
t.Fatalf("snapshotId must be set")
|
||||
}
|
||||
|
||||
// Wait for a forget operation to appear in the oplog.
|
||||
if err := retry(t, 10, 2*time.Second, func() error {
|
||||
operations := getOperations(t, sut.oplog)
|
||||
if index := slices.IndexFunc(operations, func(op *v1.Operation) bool {
|
||||
_, ok := op.GetOp().(*v1.Operation_OperationForget)
|
||||
return op.Status == v1.OperationStatus_STATUS_SUCCESS && ok
|
||||
}); index != -1 {
|
||||
op := operations[index]
|
||||
if op.SnapshotId != snapshotId {
|
||||
t.Fatalf("Snapshot ID mismatch on forget operation")
|
||||
}
|
||||
return nil
|
||||
}
|
||||
return errors.New("forget not indexed")
|
||||
}); err != nil {
|
||||
t.Fatalf("Couldn't find forget in oplog")
|
||||
}
|
||||
}
|
||||
|
||||
func TestMultipleBackup(t *testing.T) {
|
||||
sut := createSystemUnderTest(t, &config.MemoryStore{
|
||||
Config: &v1.Config{
|
||||
Modno: 1234,
|
||||
Repos: []*v1.Repo{
|
||||
{
|
||||
Id: "local",
|
||||
Uri: t.TempDir(),
|
||||
Password: "test",
|
||||
},
|
||||
},
|
||||
Plans: []*v1.Plan{
|
||||
{
|
||||
Id: "test",
|
||||
Repo: "local",
|
||||
Paths: []string{
|
||||
t.TempDir(),
|
||||
},
|
||||
Cron: "0 0 1 1 *",
|
||||
Retention: &v1.RetentionPolicy{
|
||||
KeepLastN: 1,
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
go func() {
|
||||
sut.orch.Run(ctx)
|
||||
}()
|
||||
|
||||
for i := 0; i < 2; i++ {
|
||||
_, err := sut.handler.Backup(context.Background(), connect.NewRequest(&types.StringValue{Value: "test"}))
|
||||
if err != nil {
|
||||
t.Fatalf("Backup() error = %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
// Wait for a forget that removed 1 snapshot to appear in the oplog
|
||||
if err := retry(t, 10, 2*time.Second, func() error {
|
||||
operations := getOperations(t, sut.oplog)
|
||||
if index := slices.IndexFunc(operations, func(op *v1.Operation) bool {
|
||||
forget, ok := op.GetOp().(*v1.Operation_OperationForget)
|
||||
return op.Status == v1.OperationStatus_STATUS_SUCCESS && ok && len(forget.OperationForget.Forget) == 1
|
||||
}); index != -1 {
|
||||
return nil
|
||||
}
|
||||
return errors.New("forget not indexed")
|
||||
}); err != nil {
|
||||
t.Fatalf("Couldn't find forget with 1 item removed in the operation log")
|
||||
}
|
||||
}
|
||||
|
||||
type systemUnderTest struct {
|
||||
handler *BackrestHandler
|
||||
oplog *oplog.OpLog
|
||||
orch *orchestrator.Orchestrator
|
||||
logStore *rotatinglog.RotatingLog
|
||||
config *v1.Config
|
||||
}
|
||||
|
||||
func createSystemUnderTest(t *testing.T, config config.ConfigStore) systemUnderTest {
|
||||
dir := t.TempDir()
|
||||
|
||||
cfg, err := config.Get()
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to get config: %v", err)
|
||||
}
|
||||
|
||||
resticBin, err := resticinstaller.FindOrInstallResticBinary()
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to find or install restic binary: %v", err)
|
||||
}
|
||||
oplog, err := oplog.NewOpLog(dir + "/oplog.boltdb")
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create oplog: %v", err)
|
||||
}
|
||||
logStore := rotatinglog.NewRotatingLog(dir+"/log", 10)
|
||||
orch, err := orchestrator.NewOrchestrator(
|
||||
resticBin, cfg, oplog, logStore,
|
||||
)
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create orchestrator: %v", err)
|
||||
}
|
||||
|
||||
h := NewBackrestHandler(config, orch, oplog, logStore)
|
||||
|
||||
return systemUnderTest{
|
||||
handler: h,
|
||||
oplog: oplog,
|
||||
orch: orch,
|
||||
logStore: logStore,
|
||||
config: cfg,
|
||||
}
|
||||
}
|
||||
|
||||
func retry(t *testing.T, times int, backoff time.Duration, f func() error) error {
|
||||
t.Helper()
|
||||
var err error
|
||||
for i := 0; i < times; i++ {
|
||||
err = f()
|
||||
if err == nil {
|
||||
return nil
|
||||
}
|
||||
time.Sleep(backoff)
|
||||
}
|
||||
return err
|
||||
}
|
||||
|
||||
func getOperations(t *testing.T, oplog *oplog.OpLog) []*v1.Operation {
|
||||
operations := []*v1.Operation{}
|
||||
if err := oplog.ForAll(func(op *v1.Operation) error {
|
||||
operations = append(operations, op)
|
||||
t.Logf("operation %t status %s", op.GetOp(), op.Status)
|
||||
return nil
|
||||
}); err != nil {
|
||||
t.Fatalf("Failed to read oplog: %v", err)
|
||||
}
|
||||
return operations
|
||||
}
|
||||
@@ -38,8 +38,8 @@ func NewHookExecutor(oplog *oplog.OpLog, bigOutputStore *rotatinglog.RotatingLog
|
||||
func (e *HookExecutor) ExecuteHooks(repo *v1.Repo, plan *v1.Plan, snapshotId string, events []v1.Hook_Condition, vars HookVars) {
|
||||
operationBase := v1.Operation{
|
||||
Status: v1.OperationStatus_STATUS_INPROGRESS,
|
||||
PlanId: plan.Id,
|
||||
RepoId: plan.Repo,
|
||||
PlanId: plan.GetId(),
|
||||
RepoId: repo.GetId(),
|
||||
SnapshotId: snapshotId,
|
||||
}
|
||||
|
||||
|
||||
@@ -165,7 +165,6 @@ func backupHelper(ctx context.Context, t Task, orchestrator *Orchestrator, plan
|
||||
if plan.Retention != nil {
|
||||
orchestrator.ScheduleTask(NewOneoffForgetTask(orchestrator, plan, op.SnapshotId, at), TaskPriorityForget)
|
||||
}
|
||||
|
||||
orchestrator.ScheduleTask(NewOneoffIndexSnapshotsTask(orchestrator, plan.Repo, at), TaskPriorityIndexSnapshots)
|
||||
orchestrator.ScheduleTask(NewOneoffStatsTask(orchestrator, plan, op.SnapshotId, at), TaskPriorityStats)
|
||||
|
||||
|
||||
@@ -16,6 +16,8 @@ import (
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"go.uber.org/zap"
|
||||
)
|
||||
|
||||
var ErrFileNotFound = errors.New("file not found")
|
||||
@@ -76,6 +78,9 @@ func (r *RotatingLog) Write(data []byte) (string, error) {
|
||||
return "", err
|
||||
}
|
||||
r.lastFile = file
|
||||
if err := r.removeExpiredFiles(); err != nil {
|
||||
zap.L().Error("failed to remove expired files for rotatinglog", zap.Error(err), zap.String("dir", r.dir))
|
||||
}
|
||||
}
|
||||
f, err := os.OpenFile(file, os.O_RDWR|os.O_CREATE, os.ModePerm)
|
||||
if err != nil {
|
||||
@@ -96,9 +101,9 @@ func (r *RotatingLog) Write(data []byte) (string, error) {
|
||||
}
|
||||
tw := tar.NewWriter(f)
|
||||
defer tw.Close()
|
||||
name := fmt.Sprintf("%s/%d", path.Base(file), pos)
|
||||
|
||||
tw.WriteHeader(&tar.Header{
|
||||
Name: name,
|
||||
Name: fmt.Sprintf("%d.gz", pos),
|
||||
Size: int64(len(data)),
|
||||
Mode: 0600,
|
||||
Typeflag: tar.TypeReg,
|
||||
@@ -110,7 +115,7 @@ func (r *RotatingLog) Write(data []byte) (string, error) {
|
||||
return "", err
|
||||
}
|
||||
|
||||
return name, nil
|
||||
return fmt.Sprintf("%s/%d", path.Base(file), pos), nil
|
||||
}
|
||||
|
||||
func (r *RotatingLog) Read(name string) ([]byte, error) {
|
||||
@@ -140,6 +145,7 @@ func (r *RotatingLog) Read(name string) ([]byte, error) {
|
||||
f.Seek(int64(offset), io.SeekStart)
|
||||
|
||||
// search for the tarball segment in the tarball and read + decompress it if found
|
||||
seekName := fmt.Sprintf("%d.gz", offset)
|
||||
tr := tar.NewReader(f)
|
||||
for {
|
||||
hdr, err := tr.Next()
|
||||
@@ -149,7 +155,7 @@ func (r *RotatingLog) Read(name string) ([]byte, error) {
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("next failed: %v", err)
|
||||
}
|
||||
if hdr.Name == name {
|
||||
if hdr.Name == seekName {
|
||||
buf := make([]byte, hdr.Size)
|
||||
_, err := io.ReadFull(tr, buf)
|
||||
if err != nil {
|
||||
|
||||
Reference in New Issue
Block a user