Steps 12 & 12b: gRPC server and directory recursion + mirror.
Step 12: GardenSync gRPC server with 5 RPC handlers — PushManifest (timestamp comparison, missing blob detection), PushBlobs (chunked streaming, manifest replacement), PullManifest, PullBlobs, Prune. Added store.List() and garden.ListBlobs()/DeleteBlob() for prune. In-process tests via bufconn. Step 12b: Add now recurses directories (walks files/symlinks, skips dir entries). Mirror up syncs filesystem → manifest (add new, remove deleted, rehash changed). Mirror down syncs manifest → filesystem (restore + delete untracked with optional confirm). 7 tests. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
225
server/server.go
Normal file
225
server/server.go
Normal file
@@ -0,0 +1,225 @@
|
||||
// Package server implements the GardenSync gRPC service.
|
||||
package server
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"io"
|
||||
"sync"
|
||||
|
||||
"github.com/kisom/sgard/garden"
|
||||
"github.com/kisom/sgard/manifest"
|
||||
"github.com/kisom/sgard/sgardpb"
|
||||
"google.golang.org/grpc"
|
||||
"google.golang.org/grpc/codes"
|
||||
"google.golang.org/grpc/status"
|
||||
"google.golang.org/protobuf/types/known/timestamppb"
|
||||
)
|
||||
|
||||
const chunkSize = 64 * 1024 // 64 KiB
|
||||
|
||||
// Server implements the sgardpb.GardenSyncServer interface.
|
||||
type Server struct {
|
||||
sgardpb.UnimplementedGardenSyncServer
|
||||
garden *garden.Garden
|
||||
mu sync.RWMutex
|
||||
pendingManifest *manifest.Manifest
|
||||
}
|
||||
|
||||
// New creates a new Server backed by the given Garden.
|
||||
func New(g *garden.Garden) *Server {
|
||||
return &Server{garden: g}
|
||||
}
|
||||
|
||||
// PushManifest compares the client manifest against the server manifest and
|
||||
// decides whether to accept, reject, or report up-to-date.
|
||||
func (s *Server) PushManifest(_ context.Context, req *sgardpb.PushManifestRequest) (*sgardpb.PushManifestResponse, error) {
|
||||
s.mu.Lock()
|
||||
defer s.mu.Unlock()
|
||||
|
||||
serverManifest := s.garden.GetManifest()
|
||||
clientManifest := ProtoToManifest(req.GetManifest())
|
||||
|
||||
resp := &sgardpb.PushManifestResponse{
|
||||
ServerUpdated: timestamppb.New(serverManifest.Updated),
|
||||
}
|
||||
|
||||
switch {
|
||||
case clientManifest.Updated.After(serverManifest.Updated):
|
||||
resp.Decision = sgardpb.PushManifestResponse_ACCEPTED
|
||||
|
||||
var missing []string
|
||||
for _, e := range clientManifest.Files {
|
||||
if e.Type == "file" && e.Hash != "" && !s.garden.BlobExists(e.Hash) {
|
||||
missing = append(missing, e.Hash)
|
||||
}
|
||||
}
|
||||
resp.MissingBlobs = missing
|
||||
s.pendingManifest = clientManifest
|
||||
|
||||
case serverManifest.Updated.After(clientManifest.Updated):
|
||||
resp.Decision = sgardpb.PushManifestResponse_REJECTED
|
||||
|
||||
default:
|
||||
resp.Decision = sgardpb.PushManifestResponse_UP_TO_DATE
|
||||
}
|
||||
|
||||
return resp, nil
|
||||
}
|
||||
|
||||
// PushBlobs receives a stream of blob chunks, reassembles them, writes each
|
||||
// blob to the store, and then applies the pending manifest.
|
||||
func (s *Server) PushBlobs(stream grpc.ClientStreamingServer[sgardpb.PushBlobsRequest, sgardpb.PushBlobsResponse]) error {
|
||||
s.mu.Lock()
|
||||
defer s.mu.Unlock()
|
||||
|
||||
var (
|
||||
currentHash string
|
||||
buf []byte
|
||||
blobCount int32
|
||||
)
|
||||
|
||||
for {
|
||||
req, err := stream.Recv()
|
||||
if errors.Is(err, io.EOF) {
|
||||
break
|
||||
}
|
||||
if err != nil {
|
||||
return status.Errorf(codes.Internal, "receiving blob chunk: %v", err)
|
||||
}
|
||||
|
||||
chunk := req.GetChunk()
|
||||
if chunk == nil {
|
||||
continue
|
||||
}
|
||||
|
||||
if chunk.GetHash() != "" {
|
||||
// New blob starting. Write out the previous one if any.
|
||||
if currentHash != "" {
|
||||
if err := s.writeAndVerify(currentHash, buf); err != nil {
|
||||
return err
|
||||
}
|
||||
blobCount++
|
||||
}
|
||||
currentHash = chunk.GetHash()
|
||||
buf = append([]byte(nil), chunk.GetData()...)
|
||||
} else {
|
||||
buf = append(buf, chunk.GetData()...)
|
||||
}
|
||||
}
|
||||
|
||||
// Write the last accumulated blob.
|
||||
if currentHash != "" {
|
||||
if err := s.writeAndVerify(currentHash, buf); err != nil {
|
||||
return err
|
||||
}
|
||||
blobCount++
|
||||
}
|
||||
|
||||
// Apply pending manifest.
|
||||
if s.pendingManifest != nil {
|
||||
if err := s.garden.ReplaceManifest(s.pendingManifest); err != nil {
|
||||
return status.Errorf(codes.Internal, "replacing manifest: %v", err)
|
||||
}
|
||||
s.pendingManifest = nil
|
||||
}
|
||||
|
||||
return stream.SendAndClose(&sgardpb.PushBlobsResponse{
|
||||
BlobsReceived: blobCount,
|
||||
})
|
||||
}
|
||||
|
||||
// writeAndVerify writes data to the blob store and verifies the hash matches.
|
||||
func (s *Server) writeAndVerify(expectedHash string, data []byte) error {
|
||||
gotHash, err := s.garden.WriteBlob(data)
|
||||
if err != nil {
|
||||
return status.Errorf(codes.Internal, "writing blob: %v", err)
|
||||
}
|
||||
if gotHash != expectedHash {
|
||||
return status.Errorf(codes.DataLoss, "blob hash mismatch: expected %s, got %s", expectedHash, gotHash)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// PullManifest returns the server's current manifest.
|
||||
func (s *Server) PullManifest(_ context.Context, _ *sgardpb.PullManifestRequest) (*sgardpb.PullManifestResponse, error) {
|
||||
s.mu.RLock()
|
||||
defer s.mu.RUnlock()
|
||||
|
||||
return &sgardpb.PullManifestResponse{
|
||||
Manifest: ManifestToProto(s.garden.GetManifest()),
|
||||
}, nil
|
||||
}
|
||||
|
||||
// PullBlobs streams the requested blobs back to the client in 64 KiB chunks.
|
||||
func (s *Server) PullBlobs(req *sgardpb.PullBlobsRequest, stream grpc.ServerStreamingServer[sgardpb.PullBlobsResponse]) error {
|
||||
s.mu.RLock()
|
||||
defer s.mu.RUnlock()
|
||||
|
||||
for _, hash := range req.GetHashes() {
|
||||
data, err := s.garden.ReadBlob(hash)
|
||||
if err != nil {
|
||||
return status.Errorf(codes.NotFound, "reading blob %s: %v", hash, err)
|
||||
}
|
||||
|
||||
for i := 0; i < len(data); i += chunkSize {
|
||||
end := i + chunkSize
|
||||
if end > len(data) {
|
||||
end = len(data)
|
||||
}
|
||||
chunk := &sgardpb.BlobChunk{
|
||||
Data: data[i:end],
|
||||
}
|
||||
if i == 0 {
|
||||
chunk.Hash = hash
|
||||
}
|
||||
if err := stream.Send(&sgardpb.PullBlobsResponse{Chunk: chunk}); err != nil {
|
||||
return status.Errorf(codes.Internal, "sending blob chunk: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
// Handle empty blobs: send a single chunk with the hash.
|
||||
if len(data) == 0 {
|
||||
if err := stream.Send(&sgardpb.PullBlobsResponse{
|
||||
Chunk: &sgardpb.BlobChunk{Hash: hash},
|
||||
}); err != nil {
|
||||
return status.Errorf(codes.Internal, "sending empty blob chunk: %v", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// Prune removes orphaned blobs that are not referenced by the current manifest.
|
||||
func (s *Server) Prune(_ context.Context, _ *sgardpb.PruneRequest) (*sgardpb.PruneResponse, error) {
|
||||
s.mu.Lock()
|
||||
defer s.mu.Unlock()
|
||||
|
||||
// Collect all referenced hashes from the manifest.
|
||||
referenced := make(map[string]bool)
|
||||
for _, e := range s.garden.GetManifest().Files {
|
||||
if e.Type == "file" && e.Hash != "" {
|
||||
referenced[e.Hash] = true
|
||||
}
|
||||
}
|
||||
|
||||
// List all blobs in the store.
|
||||
allBlobs, err := s.garden.ListBlobs()
|
||||
if err != nil {
|
||||
return nil, status.Errorf(codes.Internal, "listing blobs: %v", err)
|
||||
}
|
||||
|
||||
// Delete orphans.
|
||||
var removed int32
|
||||
for _, hash := range allBlobs {
|
||||
if !referenced[hash] {
|
||||
if err := s.garden.DeleteBlob(hash); err != nil {
|
||||
return nil, status.Errorf(codes.Internal, "deleting blob %s: %v", hash, err)
|
||||
}
|
||||
removed++
|
||||
}
|
||||
}
|
||||
|
||||
return &sgardpb.PruneResponse{BlobsRemoved: removed}, nil
|
||||
}
|
||||
336
server/server_test.go
Normal file
336
server/server_test.go
Normal file
@@ -0,0 +1,336 @@
|
||||
package server
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"io"
|
||||
"net"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/kisom/sgard/garden"
|
||||
"github.com/kisom/sgard/manifest"
|
||||
"github.com/kisom/sgard/sgardpb"
|
||||
"google.golang.org/grpc"
|
||||
"google.golang.org/grpc/credentials/insecure"
|
||||
"google.golang.org/grpc/test/bufconn"
|
||||
)
|
||||
|
||||
const bufSize = 1024 * 1024
|
||||
|
||||
// setupTest creates a client-server pair using in-process bufconn.
|
||||
// It returns a gRPC client, the server Garden, and a client Garden.
|
||||
func setupTest(t *testing.T) (sgardpb.GardenSyncClient, *garden.Garden, *garden.Garden) {
|
||||
t.Helper()
|
||||
|
||||
serverDir := t.TempDir()
|
||||
serverGarden, err := garden.Init(serverDir)
|
||||
if err != nil {
|
||||
t.Fatalf("init server garden: %v", err)
|
||||
}
|
||||
|
||||
clientDir := t.TempDir()
|
||||
clientGarden, err := garden.Init(clientDir)
|
||||
if err != nil {
|
||||
t.Fatalf("init client garden: %v", err)
|
||||
}
|
||||
|
||||
lis := bufconn.Listen(bufSize)
|
||||
srv := grpc.NewServer()
|
||||
sgardpb.RegisterGardenSyncServer(srv, New(serverGarden))
|
||||
t.Cleanup(func() { srv.Stop() })
|
||||
|
||||
go func() {
|
||||
_ = srv.Serve(lis)
|
||||
}()
|
||||
|
||||
conn, err := grpc.NewClient("passthrough:///bufconn",
|
||||
grpc.WithContextDialer(func(context.Context, string) (net.Conn, error) {
|
||||
return lis.Dial()
|
||||
}),
|
||||
grpc.WithTransportCredentials(insecure.NewCredentials()),
|
||||
)
|
||||
if err != nil {
|
||||
t.Fatalf("dial bufconn: %v", err)
|
||||
}
|
||||
t.Cleanup(func() { _ = conn.Close() })
|
||||
|
||||
client := sgardpb.NewGardenSyncClient(conn)
|
||||
return client, serverGarden, clientGarden
|
||||
}
|
||||
|
||||
func TestPushManifest_Accepted(t *testing.T) {
|
||||
client, serverGarden, _ := setupTest(t)
|
||||
ctx := context.Background()
|
||||
|
||||
// Server has an old manifest (default init time).
|
||||
// Client has a newer manifest with a file entry.
|
||||
now := time.Now().UTC()
|
||||
clientManifest := &manifest.Manifest{
|
||||
Version: 1,
|
||||
Created: now,
|
||||
Updated: now.Add(time.Hour),
|
||||
Files: []manifest.Entry{
|
||||
{
|
||||
Path: "~/.bashrc",
|
||||
Hash: "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa",
|
||||
Type: "file",
|
||||
Mode: "0644",
|
||||
Updated: now,
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
resp, err := client.PushManifest(ctx, &sgardpb.PushManifestRequest{
|
||||
Manifest: ManifestToProto(clientManifest),
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatalf("PushManifest: %v", err)
|
||||
}
|
||||
|
||||
if resp.Decision != sgardpb.PushManifestResponse_ACCEPTED {
|
||||
t.Errorf("decision: got %v, want ACCEPTED", resp.Decision)
|
||||
}
|
||||
|
||||
// The blob doesn't exist on server, so it should be in missing_blobs.
|
||||
if len(resp.MissingBlobs) != 1 {
|
||||
t.Fatalf("missing_blobs count: got %d, want 1", len(resp.MissingBlobs))
|
||||
}
|
||||
if resp.MissingBlobs[0] != "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa" {
|
||||
t.Errorf("missing_blobs[0]: got %s, want aaaa...", resp.MissingBlobs[0])
|
||||
}
|
||||
|
||||
// Write the blob to server and try again: it should not be missing.
|
||||
_, err = serverGarden.WriteBlob([]byte("test data"))
|
||||
if err != nil {
|
||||
t.Fatalf("WriteBlob: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestPushManifest_Rejected(t *testing.T) {
|
||||
client, serverGarden, _ := setupTest(t)
|
||||
ctx := context.Background()
|
||||
|
||||
// Make the server manifest newer.
|
||||
serverManifest := serverGarden.GetManifest()
|
||||
serverManifest.Updated = time.Now().UTC().Add(2 * time.Hour)
|
||||
if err := serverGarden.ReplaceManifest(serverManifest); err != nil {
|
||||
t.Fatalf("ReplaceManifest: %v", err)
|
||||
}
|
||||
|
||||
// Client manifest is at default init time (older).
|
||||
clientManifest := &manifest.Manifest{
|
||||
Version: 1,
|
||||
Created: time.Now().UTC(),
|
||||
Updated: time.Now().UTC(),
|
||||
Files: []manifest.Entry{},
|
||||
}
|
||||
|
||||
resp, err := client.PushManifest(ctx, &sgardpb.PushManifestRequest{
|
||||
Manifest: ManifestToProto(clientManifest),
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatalf("PushManifest: %v", err)
|
||||
}
|
||||
|
||||
if resp.Decision != sgardpb.PushManifestResponse_REJECTED {
|
||||
t.Errorf("decision: got %v, want REJECTED", resp.Decision)
|
||||
}
|
||||
}
|
||||
|
||||
func TestPushManifest_UpToDate(t *testing.T) {
|
||||
client, serverGarden, _ := setupTest(t)
|
||||
ctx := context.Background()
|
||||
|
||||
// Set both to the same timestamp.
|
||||
ts := time.Date(2026, 1, 15, 12, 0, 0, 0, time.UTC)
|
||||
serverManifest := serverGarden.GetManifest()
|
||||
serverManifest.Updated = ts
|
||||
if err := serverGarden.ReplaceManifest(serverManifest); err != nil {
|
||||
t.Fatalf("ReplaceManifest: %v", err)
|
||||
}
|
||||
|
||||
clientManifest := &manifest.Manifest{
|
||||
Version: 1,
|
||||
Created: ts,
|
||||
Updated: ts,
|
||||
Files: []manifest.Entry{},
|
||||
}
|
||||
|
||||
resp, err := client.PushManifest(ctx, &sgardpb.PushManifestRequest{
|
||||
Manifest: ManifestToProto(clientManifest),
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatalf("PushManifest: %v", err)
|
||||
}
|
||||
|
||||
if resp.Decision != sgardpb.PushManifestResponse_UP_TO_DATE {
|
||||
t.Errorf("decision: got %v, want UP_TO_DATE", resp.Decision)
|
||||
}
|
||||
}
|
||||
|
||||
func TestPushAndPullBlobs(t *testing.T) {
|
||||
client, serverGarden, _ := setupTest(t)
|
||||
ctx := context.Background()
|
||||
|
||||
// Write some test data as blobs directly to simulate a client garden.
|
||||
blob1Data := []byte("hello world from bashrc")
|
||||
blob2Data := []byte("vimrc content here")
|
||||
|
||||
// We need the actual hashes for our manifest entries.
|
||||
// Write to a throwaway garden to get hashes.
|
||||
tmpDir := t.TempDir()
|
||||
tmpGarden, err := garden.Init(tmpDir)
|
||||
if err != nil {
|
||||
t.Fatalf("init tmp garden: %v", err)
|
||||
}
|
||||
hash1, err := tmpGarden.WriteBlob(blob1Data)
|
||||
if err != nil {
|
||||
t.Fatalf("WriteBlob 1: %v", err)
|
||||
}
|
||||
hash2, err := tmpGarden.WriteBlob(blob2Data)
|
||||
if err != nil {
|
||||
t.Fatalf("WriteBlob 2: %v", err)
|
||||
}
|
||||
|
||||
now := time.Now().UTC().Add(time.Hour)
|
||||
clientManifest := &manifest.Manifest{
|
||||
Version: 1,
|
||||
Created: now,
|
||||
Updated: now,
|
||||
Files: []manifest.Entry{
|
||||
{Path: "~/.bashrc", Hash: hash1, Type: "file", Mode: "0644", Updated: now},
|
||||
{Path: "~/.vimrc", Hash: hash2, Type: "file", Mode: "0644", Updated: now},
|
||||
{Path: "~/.config", Type: "directory", Mode: "0755", Updated: now},
|
||||
},
|
||||
}
|
||||
|
||||
// Step 1: PushManifest.
|
||||
pushResp, err := client.PushManifest(ctx, &sgardpb.PushManifestRequest{
|
||||
Manifest: ManifestToProto(clientManifest),
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatalf("PushManifest: %v", err)
|
||||
}
|
||||
if pushResp.Decision != sgardpb.PushManifestResponse_ACCEPTED {
|
||||
t.Fatalf("decision: got %v, want ACCEPTED", pushResp.Decision)
|
||||
}
|
||||
if len(pushResp.MissingBlobs) != 2 {
|
||||
t.Fatalf("missing_blobs: got %d, want 2", len(pushResp.MissingBlobs))
|
||||
}
|
||||
|
||||
// Step 2: PushBlobs.
|
||||
stream, err := client.PushBlobs(ctx)
|
||||
if err != nil {
|
||||
t.Fatalf("PushBlobs: %v", err)
|
||||
}
|
||||
|
||||
// Send blob1.
|
||||
if err := stream.Send(&sgardpb.PushBlobsRequest{
|
||||
Chunk: &sgardpb.BlobChunk{Hash: hash1, Data: blob1Data},
|
||||
}); err != nil {
|
||||
t.Fatalf("Send blob1: %v", err)
|
||||
}
|
||||
|
||||
// Send blob2.
|
||||
if err := stream.Send(&sgardpb.PushBlobsRequest{
|
||||
Chunk: &sgardpb.BlobChunk{Hash: hash2, Data: blob2Data},
|
||||
}); err != nil {
|
||||
t.Fatalf("Send blob2: %v", err)
|
||||
}
|
||||
|
||||
blobResp, err := stream.CloseAndRecv()
|
||||
if err != nil {
|
||||
t.Fatalf("CloseAndRecv: %v", err)
|
||||
}
|
||||
if blobResp.BlobsReceived != 2 {
|
||||
t.Errorf("blobs_received: got %d, want 2", blobResp.BlobsReceived)
|
||||
}
|
||||
|
||||
// Verify blobs exist on server.
|
||||
if !serverGarden.BlobExists(hash1) {
|
||||
t.Error("blob1 not found on server")
|
||||
}
|
||||
if !serverGarden.BlobExists(hash2) {
|
||||
t.Error("blob2 not found on server")
|
||||
}
|
||||
|
||||
// Verify manifest was applied on server.
|
||||
sm := serverGarden.GetManifest()
|
||||
if len(sm.Files) != 3 {
|
||||
t.Fatalf("server manifest files: got %d, want 3", len(sm.Files))
|
||||
}
|
||||
|
||||
// Step 3: PullManifest from the server.
|
||||
pullMResp, err := client.PullManifest(ctx, &sgardpb.PullManifestRequest{})
|
||||
if err != nil {
|
||||
t.Fatalf("PullManifest: %v", err)
|
||||
}
|
||||
pulledManifest := ProtoToManifest(pullMResp.GetManifest())
|
||||
if len(pulledManifest.Files) != 3 {
|
||||
t.Fatalf("pulled manifest files: got %d, want 3", len(pulledManifest.Files))
|
||||
}
|
||||
|
||||
// Step 4: PullBlobs from the server.
|
||||
pullBResp, err := client.PullBlobs(ctx, &sgardpb.PullBlobsRequest{
|
||||
Hashes: []string{hash1, hash2},
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatalf("PullBlobs: %v", err)
|
||||
}
|
||||
|
||||
// Reassemble blobs from the stream.
|
||||
pulledBlobs := make(map[string][]byte)
|
||||
var currentHash string
|
||||
for {
|
||||
resp, err := pullBResp.Recv()
|
||||
if errors.Is(err, io.EOF) {
|
||||
break
|
||||
}
|
||||
if err != nil {
|
||||
t.Fatalf("PullBlobs Recv: %v", err)
|
||||
}
|
||||
chunk := resp.GetChunk()
|
||||
if chunk.GetHash() != "" {
|
||||
currentHash = chunk.GetHash()
|
||||
}
|
||||
pulledBlobs[currentHash] = append(pulledBlobs[currentHash], chunk.GetData()...)
|
||||
}
|
||||
|
||||
if string(pulledBlobs[hash1]) != string(blob1Data) {
|
||||
t.Errorf("blob1 data mismatch: got %q, want %q", pulledBlobs[hash1], blob1Data)
|
||||
}
|
||||
if string(pulledBlobs[hash2]) != string(blob2Data) {
|
||||
t.Errorf("blob2 data mismatch: got %q, want %q", pulledBlobs[hash2], blob2Data)
|
||||
}
|
||||
}
|
||||
|
||||
func TestPrune(t *testing.T) {
|
||||
client, serverGarden, _ := setupTest(t)
|
||||
ctx := context.Background()
|
||||
|
||||
// Write a blob to the server.
|
||||
blobData := []byte("orphan blob data")
|
||||
hash, err := serverGarden.WriteBlob(blobData)
|
||||
if err != nil {
|
||||
t.Fatalf("WriteBlob: %v", err)
|
||||
}
|
||||
|
||||
// The manifest does NOT reference this blob, so it is orphaned.
|
||||
if !serverGarden.BlobExists(hash) {
|
||||
t.Fatal("blob should exist before prune")
|
||||
}
|
||||
|
||||
resp, err := client.Prune(ctx, &sgardpb.PruneRequest{})
|
||||
if err != nil {
|
||||
t.Fatalf("Prune: %v", err)
|
||||
}
|
||||
|
||||
if resp.BlobsRemoved != 1 {
|
||||
t.Errorf("blobs_removed: got %d, want 1", resp.BlobsRemoved)
|
||||
}
|
||||
|
||||
if serverGarden.BlobExists(hash) {
|
||||
t.Error("orphan blob should be deleted after prune")
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user