mirror of
https://github.com/moby/moby.git
synced 2022-11-09 12:21:53 -05:00
Merge pull request #26564 from miminar/prevent-blob-uploads
Avoid unnecessary blob uploads
This commit is contained in:
commit
762556c705
4 changed files with 999 additions and 104 deletions
|
@ -1,52 +1,130 @@
|
||||||
package metadata
|
package metadata
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"crypto/hmac"
|
||||||
|
"crypto/sha256"
|
||||||
|
"encoding/hex"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
|
||||||
"github.com/docker/distribution/digest"
|
"github.com/docker/distribution/digest"
|
||||||
|
"github.com/docker/docker/api/types"
|
||||||
"github.com/docker/docker/layer"
|
"github.com/docker/docker/layer"
|
||||||
)
|
)
|
||||||
|
|
||||||
// V2MetadataService maps layer IDs to a set of known metadata for
|
// V2MetadataService maps layer IDs to a set of known metadata for
|
||||||
// the layer.
|
// the layer.
|
||||||
type V2MetadataService struct {
|
type V2MetadataService interface {
|
||||||
|
GetMetadata(diffID layer.DiffID) ([]V2Metadata, error)
|
||||||
|
GetDiffID(dgst digest.Digest) (layer.DiffID, error)
|
||||||
|
Add(diffID layer.DiffID, metadata V2Metadata) error
|
||||||
|
TagAndAdd(diffID layer.DiffID, hmacKey []byte, metadata V2Metadata) error
|
||||||
|
Remove(metadata V2Metadata) error
|
||||||
|
}
|
||||||
|
|
||||||
|
// v2MetadataService implements V2MetadataService
|
||||||
|
type v2MetadataService struct {
|
||||||
store Store
|
store Store
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var _ V2MetadataService = &v2MetadataService{}
|
||||||
|
|
||||||
// V2Metadata contains the digest and source repository information for a layer.
|
// V2Metadata contains the digest and source repository information for a layer.
|
||||||
type V2Metadata struct {
|
type V2Metadata struct {
|
||||||
Digest digest.Digest
|
Digest digest.Digest
|
||||||
SourceRepository string
|
SourceRepository string
|
||||||
|
// HMAC hashes above attributes with recent authconfig digest used as a key in order to determine matching
|
||||||
|
// metadata entries accompanied by the same credentials without actually exposing them.
|
||||||
|
HMAC string
|
||||||
|
}
|
||||||
|
|
||||||
|
// CheckV2MetadataHMAC return true if the given "meta" is tagged with a hmac hashed by the given "key".
|
||||||
|
func CheckV2MetadataHMAC(meta *V2Metadata, key []byte) bool {
|
||||||
|
if len(meta.HMAC) == 0 || len(key) == 0 {
|
||||||
|
return len(meta.HMAC) == 0 && len(key) == 0
|
||||||
|
}
|
||||||
|
mac := hmac.New(sha256.New, key)
|
||||||
|
mac.Write([]byte(meta.Digest))
|
||||||
|
mac.Write([]byte(meta.SourceRepository))
|
||||||
|
expectedMac := mac.Sum(nil)
|
||||||
|
|
||||||
|
storedMac, err := hex.DecodeString(meta.HMAC)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return hmac.Equal(storedMac, expectedMac)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ComputeV2MetadataHMAC returns a hmac for the given "meta" hash by the given key.
|
||||||
|
func ComputeV2MetadataHMAC(key []byte, meta *V2Metadata) string {
|
||||||
|
if len(key) == 0 || meta == nil {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
mac := hmac.New(sha256.New, key)
|
||||||
|
mac.Write([]byte(meta.Digest))
|
||||||
|
mac.Write([]byte(meta.SourceRepository))
|
||||||
|
return hex.EncodeToString(mac.Sum(nil))
|
||||||
|
}
|
||||||
|
|
||||||
|
// ComputeV2MetadataHMACKey returns a key for the given "authConfig" that can be used to hash v2 metadata
|
||||||
|
// entries.
|
||||||
|
func ComputeV2MetadataHMACKey(authConfig *types.AuthConfig) ([]byte, error) {
|
||||||
|
if authConfig == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
key := authConfigKeyInput{
|
||||||
|
Username: authConfig.Username,
|
||||||
|
Password: authConfig.Password,
|
||||||
|
Auth: authConfig.Auth,
|
||||||
|
IdentityToken: authConfig.IdentityToken,
|
||||||
|
RegistryToken: authConfig.RegistryToken,
|
||||||
|
}
|
||||||
|
buf, err := json.Marshal(&key)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return []byte(digest.FromBytes([]byte(buf))), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// authConfigKeyInput is a reduced AuthConfig structure holding just relevant credential data eligible for
|
||||||
|
// hmac key creation.
|
||||||
|
type authConfigKeyInput struct {
|
||||||
|
Username string `json:"username,omitempty"`
|
||||||
|
Password string `json:"password,omitempty"`
|
||||||
|
Auth string `json:"auth,omitempty"`
|
||||||
|
|
||||||
|
IdentityToken string `json:"identitytoken,omitempty"`
|
||||||
|
RegistryToken string `json:"registrytoken,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// maxMetadata is the number of metadata entries to keep per layer DiffID.
|
// maxMetadata is the number of metadata entries to keep per layer DiffID.
|
||||||
const maxMetadata = 50
|
const maxMetadata = 50
|
||||||
|
|
||||||
// NewV2MetadataService creates a new diff ID to v2 metadata mapping service.
|
// NewV2MetadataService creates a new diff ID to v2 metadata mapping service.
|
||||||
func NewV2MetadataService(store Store) *V2MetadataService {
|
func NewV2MetadataService(store Store) V2MetadataService {
|
||||||
return &V2MetadataService{
|
return &v2MetadataService{
|
||||||
store: store,
|
store: store,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (serv *V2MetadataService) diffIDNamespace() string {
|
func (serv *v2MetadataService) diffIDNamespace() string {
|
||||||
return "v2metadata-by-diffid"
|
return "v2metadata-by-diffid"
|
||||||
}
|
}
|
||||||
|
|
||||||
func (serv *V2MetadataService) digestNamespace() string {
|
func (serv *v2MetadataService) digestNamespace() string {
|
||||||
return "diffid-by-digest"
|
return "diffid-by-digest"
|
||||||
}
|
}
|
||||||
|
|
||||||
func (serv *V2MetadataService) diffIDKey(diffID layer.DiffID) string {
|
func (serv *v2MetadataService) diffIDKey(diffID layer.DiffID) string {
|
||||||
return string(digest.Digest(diffID).Algorithm()) + "/" + digest.Digest(diffID).Hex()
|
return string(digest.Digest(diffID).Algorithm()) + "/" + digest.Digest(diffID).Hex()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (serv *V2MetadataService) digestKey(dgst digest.Digest) string {
|
func (serv *v2MetadataService) digestKey(dgst digest.Digest) string {
|
||||||
return string(dgst.Algorithm()) + "/" + dgst.Hex()
|
return string(dgst.Algorithm()) + "/" + dgst.Hex()
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetMetadata finds the metadata associated with a layer DiffID.
|
// GetMetadata finds the metadata associated with a layer DiffID.
|
||||||
func (serv *V2MetadataService) GetMetadata(diffID layer.DiffID) ([]V2Metadata, error) {
|
func (serv *v2MetadataService) GetMetadata(diffID layer.DiffID) ([]V2Metadata, error) {
|
||||||
jsonBytes, err := serv.store.Get(serv.diffIDNamespace(), serv.diffIDKey(diffID))
|
jsonBytes, err := serv.store.Get(serv.diffIDNamespace(), serv.diffIDKey(diffID))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
@ -61,7 +139,7 @@ func (serv *V2MetadataService) GetMetadata(diffID layer.DiffID) ([]V2Metadata, e
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetDiffID finds a layer DiffID from a digest.
|
// GetDiffID finds a layer DiffID from a digest.
|
||||||
func (serv *V2MetadataService) GetDiffID(dgst digest.Digest) (layer.DiffID, error) {
|
func (serv *v2MetadataService) GetDiffID(dgst digest.Digest) (layer.DiffID, error) {
|
||||||
diffIDBytes, err := serv.store.Get(serv.digestNamespace(), serv.digestKey(dgst))
|
diffIDBytes, err := serv.store.Get(serv.digestNamespace(), serv.digestKey(dgst))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return layer.DiffID(""), err
|
return layer.DiffID(""), err
|
||||||
|
@ -72,7 +150,7 @@ func (serv *V2MetadataService) GetDiffID(dgst digest.Digest) (layer.DiffID, erro
|
||||||
|
|
||||||
// Add associates metadata with a layer DiffID. If too many metadata entries are
|
// Add associates metadata with a layer DiffID. If too many metadata entries are
|
||||||
// present, the oldest one is dropped.
|
// present, the oldest one is dropped.
|
||||||
func (serv *V2MetadataService) Add(diffID layer.DiffID, metadata V2Metadata) error {
|
func (serv *v2MetadataService) Add(diffID layer.DiffID, metadata V2Metadata) error {
|
||||||
oldMetadata, err := serv.GetMetadata(diffID)
|
oldMetadata, err := serv.GetMetadata(diffID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
oldMetadata = nil
|
oldMetadata = nil
|
||||||
|
@ -105,8 +183,15 @@ func (serv *V2MetadataService) Add(diffID layer.DiffID, metadata V2Metadata) err
|
||||||
return serv.store.Set(serv.digestNamespace(), serv.digestKey(metadata.Digest), []byte(diffID))
|
return serv.store.Set(serv.digestNamespace(), serv.digestKey(metadata.Digest), []byte(diffID))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TagAndAdd amends the given "meta" for hmac hashed by the given "hmacKey" and associates it with a layer
|
||||||
|
// DiffID. If too many metadata entries are present, the oldest one is dropped.
|
||||||
|
func (serv *v2MetadataService) TagAndAdd(diffID layer.DiffID, hmacKey []byte, meta V2Metadata) error {
|
||||||
|
meta.HMAC = ComputeV2MetadataHMAC(hmacKey, &meta)
|
||||||
|
return serv.Add(diffID, meta)
|
||||||
|
}
|
||||||
|
|
||||||
// Remove unassociates a metadata entry from a layer DiffID.
|
// Remove unassociates a metadata entry from a layer DiffID.
|
||||||
func (serv *V2MetadataService) Remove(metadata V2Metadata) error {
|
func (serv *v2MetadataService) Remove(metadata V2Metadata) error {
|
||||||
diffID, err := serv.GetDiffID(metadata.Digest)
|
diffID, err := serv.GetDiffID(metadata.Digest)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
|
|
@ -50,7 +50,7 @@ func (e ImageConfigPullError) Error() string {
|
||||||
}
|
}
|
||||||
|
|
||||||
type v2Puller struct {
|
type v2Puller struct {
|
||||||
V2MetadataService *metadata.V2MetadataService
|
V2MetadataService metadata.V2MetadataService
|
||||||
endpoint registry.APIEndpoint
|
endpoint registry.APIEndpoint
|
||||||
config *ImagePullConfig
|
config *ImagePullConfig
|
||||||
repoInfo *registry.RepositoryInfo
|
repoInfo *registry.RepositoryInfo
|
||||||
|
@ -134,7 +134,7 @@ type v2LayerDescriptor struct {
|
||||||
digest digest.Digest
|
digest digest.Digest
|
||||||
repoInfo *registry.RepositoryInfo
|
repoInfo *registry.RepositoryInfo
|
||||||
repo distribution.Repository
|
repo distribution.Repository
|
||||||
V2MetadataService *metadata.V2MetadataService
|
V2MetadataService metadata.V2MetadataService
|
||||||
tmpFile *os.File
|
tmpFile *os.File
|
||||||
verifier digest.Verifier
|
verifier digest.Verifier
|
||||||
src distribution.Descriptor
|
src distribution.Descriptor
|
||||||
|
|
|
@ -5,8 +5,12 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
"sort"
|
||||||
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
|
"golang.org/x/net/context"
|
||||||
|
|
||||||
"github.com/Sirupsen/logrus"
|
"github.com/Sirupsen/logrus"
|
||||||
"github.com/docker/distribution"
|
"github.com/docker/distribution"
|
||||||
"github.com/docker/distribution/digest"
|
"github.com/docker/distribution/digest"
|
||||||
|
@ -23,7 +27,11 @@ import (
|
||||||
"github.com/docker/docker/pkg/stringid"
|
"github.com/docker/docker/pkg/stringid"
|
||||||
"github.com/docker/docker/reference"
|
"github.com/docker/docker/reference"
|
||||||
"github.com/docker/docker/registry"
|
"github.com/docker/docker/registry"
|
||||||
"golang.org/x/net/context"
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
smallLayerMaximumSize = 100 * (1 << 10) // 100KB
|
||||||
|
middleLayerMaximumSize = 10 * (1 << 20) // 10MB
|
||||||
)
|
)
|
||||||
|
|
||||||
// PushResult contains the tag, manifest digest, and manifest size from the
|
// PushResult contains the tag, manifest digest, and manifest size from the
|
||||||
|
@ -36,7 +44,7 @@ type PushResult struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
type v2Pusher struct {
|
type v2Pusher struct {
|
||||||
v2MetadataService *metadata.V2MetadataService
|
v2MetadataService metadata.V2MetadataService
|
||||||
ref reference.Named
|
ref reference.Named
|
||||||
endpoint registry.APIEndpoint
|
endpoint registry.APIEndpoint
|
||||||
repoInfo *registry.RepositoryInfo
|
repoInfo *registry.RepositoryInfo
|
||||||
|
@ -133,10 +141,16 @@ func (p *v2Pusher) pushV2Tag(ctx context.Context, ref reference.NamedTagged, id
|
||||||
defer layer.ReleaseAndLog(p.config.LayerStore, l)
|
defer layer.ReleaseAndLog(p.config.LayerStore, l)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
hmacKey, err := metadata.ComputeV2MetadataHMACKey(p.config.AuthConfig)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to compute hmac key of auth config: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
var descriptors []xfer.UploadDescriptor
|
var descriptors []xfer.UploadDescriptor
|
||||||
|
|
||||||
descriptorTemplate := v2PushDescriptor{
|
descriptorTemplate := v2PushDescriptor{
|
||||||
v2MetadataService: p.v2MetadataService,
|
v2MetadataService: p.v2MetadataService,
|
||||||
|
hmacKey: hmacKey,
|
||||||
repoInfo: p.repoInfo,
|
repoInfo: p.repoInfo,
|
||||||
ref: p.ref,
|
ref: p.ref,
|
||||||
repo: p.repo,
|
repo: p.repo,
|
||||||
|
@ -147,6 +161,7 @@ func (p *v2Pusher) pushV2Tag(ctx context.Context, ref reference.NamedTagged, id
|
||||||
for i := 0; i < len(img.RootFS.DiffIDs); i++ {
|
for i := 0; i < len(img.RootFS.DiffIDs); i++ {
|
||||||
descriptor := descriptorTemplate
|
descriptor := descriptorTemplate
|
||||||
descriptor.layer = l
|
descriptor.layer = l
|
||||||
|
descriptor.checkedDigests = make(map[digest.Digest]struct{})
|
||||||
descriptors = append(descriptors, &descriptor)
|
descriptors = append(descriptors, &descriptor)
|
||||||
|
|
||||||
l = l.Parent()
|
l = l.Parent()
|
||||||
|
@ -232,12 +247,15 @@ func manifestFromBuilder(ctx context.Context, builder distribution.ManifestBuild
|
||||||
|
|
||||||
type v2PushDescriptor struct {
|
type v2PushDescriptor struct {
|
||||||
layer layer.Layer
|
layer layer.Layer
|
||||||
v2MetadataService *metadata.V2MetadataService
|
v2MetadataService metadata.V2MetadataService
|
||||||
|
hmacKey []byte
|
||||||
repoInfo reference.Named
|
repoInfo reference.Named
|
||||||
ref reference.Named
|
ref reference.Named
|
||||||
repo distribution.Repository
|
repo distribution.Repository
|
||||||
pushState *pushState
|
pushState *pushState
|
||||||
remoteDescriptor distribution.Descriptor
|
remoteDescriptor distribution.Descriptor
|
||||||
|
// a set of digests whose presence has been checked in a target repository
|
||||||
|
checkedDigests map[digest.Digest]struct{}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (pd *v2PushDescriptor) Key() string {
|
func (pd *v2PushDescriptor) Key() string {
|
||||||
|
@ -272,71 +290,61 @@ func (pd *v2PushDescriptor) Upload(ctx context.Context, progressOutput progress.
|
||||||
}
|
}
|
||||||
pd.pushState.Unlock()
|
pd.pushState.Unlock()
|
||||||
|
|
||||||
|
maxMountAttempts, maxExistenceChecks, checkOtherRepositories := getMaxMountAndExistenceCheckAttempts(pd.layer)
|
||||||
|
|
||||||
// Do we have any metadata associated with this layer's DiffID?
|
// Do we have any metadata associated with this layer's DiffID?
|
||||||
v2Metadata, err := pd.v2MetadataService.GetMetadata(diffID)
|
v2Metadata, err := pd.v2MetadataService.GetMetadata(diffID)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
descriptor, exists, err := layerAlreadyExists(ctx, v2Metadata, pd.repoInfo, pd.repo, pd.pushState)
|
// check for blob existence in the target repository if we have a mapping with it
|
||||||
if err != nil {
|
descriptor, exists, err := pd.layerAlreadyExists(ctx, progressOutput, diffID, false, 1, v2Metadata)
|
||||||
progress.Update(progressOutput, pd.ID(), "Image push failed")
|
if exists || err != nil {
|
||||||
return distribution.Descriptor{}, retryOnError(err)
|
return descriptor, err
|
||||||
}
|
|
||||||
if exists {
|
|
||||||
progress.Update(progressOutput, pd.ID(), "Layer already exists")
|
|
||||||
pd.pushState.Lock()
|
|
||||||
pd.pushState.remoteLayers[diffID] = descriptor
|
|
||||||
pd.pushState.Unlock()
|
|
||||||
return descriptor, nil
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
logrus.Debugf("Pushing layer: %s", diffID)
|
|
||||||
|
|
||||||
// if digest was empty or not saved, or if blob does not exist on the remote repository,
|
// if digest was empty or not saved, or if blob does not exist on the remote repository,
|
||||||
// then push the blob.
|
// then push the blob.
|
||||||
bs := pd.repo.Blobs(ctx)
|
bs := pd.repo.Blobs(ctx)
|
||||||
|
|
||||||
var layerUpload distribution.BlobWriter
|
var layerUpload distribution.BlobWriter
|
||||||
mountAttemptsRemaining := 3
|
|
||||||
|
|
||||||
// Attempt to find another repository in the same registry to mount the layer
|
// Attempt to find another repository in the same registry to mount the layer from to avoid an unnecessary upload
|
||||||
// from to avoid an unnecessary upload.
|
candidates := getRepositoryMountCandidates(pd.repoInfo, pd.hmacKey, maxMountAttempts, v2Metadata)
|
||||||
// Note: metadata is stored from oldest to newest, so we iterate through this
|
for _, mountCandidate := range candidates {
|
||||||
// slice in reverse to maximize our chances of the blob still existing in the
|
logrus.Debugf("attempting to mount layer %s (%s) from %s", diffID, mountCandidate.Digest, mountCandidate.SourceRepository)
|
||||||
// remote repository.
|
createOpts := []distribution.BlobCreateOption{}
|
||||||
for i := len(v2Metadata) - 1; i >= 0 && mountAttemptsRemaining > 0; i-- {
|
|
||||||
mountFrom := v2Metadata[i]
|
|
||||||
|
|
||||||
sourceRepo, err := reference.ParseNamed(mountFrom.SourceRepository)
|
if len(mountCandidate.SourceRepository) > 0 {
|
||||||
if err != nil {
|
namedRef, err := reference.WithName(mountCandidate.SourceRepository)
|
||||||
continue
|
if err != nil {
|
||||||
}
|
logrus.Errorf("failed to parse source repository reference %v: %v", namedRef.String(), err)
|
||||||
if pd.repoInfo.Hostname() != sourceRepo.Hostname() {
|
pd.v2MetadataService.Remove(mountCandidate)
|
||||||
// don't mount blobs from another registry
|
continue
|
||||||
continue
|
}
|
||||||
|
|
||||||
|
// TODO (brianbland): We need to construct a reference where the Name is
|
||||||
|
// only the full remote name, so clean this up when distribution has a
|
||||||
|
// richer reference package
|
||||||
|
remoteRef, err := distreference.WithName(namedRef.RemoteName())
|
||||||
|
if err != nil {
|
||||||
|
logrus.Errorf("failed to make remote reference out of %q: %v", namedRef.RemoteName(), namedRef.RemoteName())
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
canonicalRef, err := distreference.WithDigest(remoteRef, mountCandidate.Digest)
|
||||||
|
if err != nil {
|
||||||
|
logrus.Errorf("failed to make canonical reference: %v", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
createOpts = append(createOpts, client.WithMountFrom(canonicalRef))
|
||||||
}
|
}
|
||||||
|
|
||||||
namedRef, err := reference.WithName(mountFrom.SourceRepository)
|
// send the layer
|
||||||
if err != nil {
|
lu, err := bs.Create(ctx, createOpts...)
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO (brianbland): We need to construct a reference where the Name is
|
|
||||||
// only the full remote name, so clean this up when distribution has a
|
|
||||||
// richer reference package
|
|
||||||
remoteRef, err := distreference.WithName(namedRef.RemoteName())
|
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
canonicalRef, err := distreference.WithDigest(remoteRef, mountFrom.Digest)
|
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
logrus.Debugf("attempting to mount layer %s (%s) from %s", diffID, mountFrom.Digest, sourceRepo.FullName())
|
|
||||||
|
|
||||||
layerUpload, err = bs.Create(ctx, client.WithMountFrom(canonicalRef))
|
|
||||||
switch err := err.(type) {
|
switch err := err.(type) {
|
||||||
|
case nil:
|
||||||
|
// noop
|
||||||
case distribution.ErrBlobMounted:
|
case distribution.ErrBlobMounted:
|
||||||
progress.Updatef(progressOutput, pd.ID(), "Mounted from %s", err.From.Name())
|
progress.Updatef(progressOutput, pd.ID(), "Mounted from %s", err.From.Name())
|
||||||
|
|
||||||
|
@ -348,21 +356,44 @@ func (pd *v2PushDescriptor) Upload(ctx context.Context, progressOutput progress.
|
||||||
pd.pushState.Unlock()
|
pd.pushState.Unlock()
|
||||||
|
|
||||||
// Cache mapping from this layer's DiffID to the blobsum
|
// Cache mapping from this layer's DiffID to the blobsum
|
||||||
if err := pd.v2MetadataService.Add(diffID, metadata.V2Metadata{Digest: mountFrom.Digest, SourceRepository: pd.repoInfo.FullName()}); err != nil {
|
if err := pd.v2MetadataService.TagAndAdd(diffID, pd.hmacKey, metadata.V2Metadata{
|
||||||
|
Digest: err.Descriptor.Digest,
|
||||||
|
SourceRepository: pd.repoInfo.FullName(),
|
||||||
|
}); err != nil {
|
||||||
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
||||||
}
|
}
|
||||||
return err.Descriptor, nil
|
return err.Descriptor, nil
|
||||||
case nil:
|
|
||||||
// blob upload session created successfully, so begin the upload
|
|
||||||
mountAttemptsRemaining = 0
|
|
||||||
default:
|
default:
|
||||||
// unable to mount layer from this repository, so this source mapping is no longer valid
|
logrus.Infof("failed to mount layer %s (%s) from %s: %v", diffID, mountCandidate.Digest, mountCandidate.SourceRepository, err)
|
||||||
logrus.Debugf("unassociating layer %s (%s) with %s", diffID, mountFrom.Digest, mountFrom.SourceRepository)
|
}
|
||||||
pd.v2MetadataService.Remove(mountFrom)
|
|
||||||
mountAttemptsRemaining--
|
if len(mountCandidate.SourceRepository) > 0 &&
|
||||||
|
(metadata.CheckV2MetadataHMAC(&mountCandidate, pd.hmacKey) ||
|
||||||
|
len(mountCandidate.HMAC) == 0) {
|
||||||
|
cause := "blob mount failure"
|
||||||
|
if err != nil {
|
||||||
|
cause = fmt.Sprintf("an error: %v", err.Error())
|
||||||
|
}
|
||||||
|
logrus.Debugf("removing association between layer %s and %s due to %s", mountCandidate.Digest, mountCandidate.SourceRepository, cause)
|
||||||
|
pd.v2MetadataService.Remove(mountCandidate)
|
||||||
|
}
|
||||||
|
|
||||||
|
if lu != nil {
|
||||||
|
// cancel previous upload
|
||||||
|
cancelLayerUpload(ctx, mountCandidate.Digest, layerUpload)
|
||||||
|
layerUpload = lu
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if maxExistenceChecks-len(pd.checkedDigests) > 0 {
|
||||||
|
// do additional layer existence checks with other known digests if any
|
||||||
|
descriptor, exists, err := pd.layerAlreadyExists(ctx, progressOutput, diffID, checkOtherRepositories, maxExistenceChecks-len(pd.checkedDigests), v2Metadata)
|
||||||
|
if exists || err != nil {
|
||||||
|
return descriptor, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
logrus.Debugf("Pushing layer: %s", diffID)
|
||||||
if layerUpload == nil {
|
if layerUpload == nil {
|
||||||
layerUpload, err = bs.Create(ctx)
|
layerUpload, err = bs.Create(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -371,6 +402,29 @@ func (pd *v2PushDescriptor) Upload(ctx context.Context, progressOutput progress.
|
||||||
}
|
}
|
||||||
defer layerUpload.Close()
|
defer layerUpload.Close()
|
||||||
|
|
||||||
|
// upload the blob
|
||||||
|
desc, err := pd.uploadUsingSession(ctx, progressOutput, diffID, layerUpload)
|
||||||
|
if err != nil {
|
||||||
|
return desc, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return desc, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pd *v2PushDescriptor) SetRemoteDescriptor(descriptor distribution.Descriptor) {
|
||||||
|
pd.remoteDescriptor = descriptor
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pd *v2PushDescriptor) Descriptor() distribution.Descriptor {
|
||||||
|
return pd.remoteDescriptor
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pd *v2PushDescriptor) uploadUsingSession(
|
||||||
|
ctx context.Context,
|
||||||
|
progressOutput progress.Output,
|
||||||
|
diffID layer.DiffID,
|
||||||
|
layerUpload distribution.BlobWriter,
|
||||||
|
) (distribution.Descriptor, error) {
|
||||||
arch, err := pd.layer.TarStream()
|
arch, err := pd.layer.TarStream()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
||||||
|
@ -404,55 +458,237 @@ func (pd *v2PushDescriptor) Upload(ctx context.Context, progressOutput progress.
|
||||||
progress.Update(progressOutput, pd.ID(), "Pushed")
|
progress.Update(progressOutput, pd.ID(), "Pushed")
|
||||||
|
|
||||||
// Cache mapping from this layer's DiffID to the blobsum
|
// Cache mapping from this layer's DiffID to the blobsum
|
||||||
if err := pd.v2MetadataService.Add(diffID, metadata.V2Metadata{Digest: pushDigest, SourceRepository: pd.repoInfo.FullName()}); err != nil {
|
if err := pd.v2MetadataService.TagAndAdd(diffID, pd.hmacKey, metadata.V2Metadata{
|
||||||
|
Digest: pushDigest,
|
||||||
|
SourceRepository: pd.repoInfo.FullName(),
|
||||||
|
}); err != nil {
|
||||||
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
return distribution.Descriptor{}, xfer.DoNotRetry{Err: err}
|
||||||
}
|
}
|
||||||
|
|
||||||
pd.pushState.Lock()
|
desc := distribution.Descriptor{
|
||||||
|
|
||||||
// If Commit succeeded, that's an indication that the remote registry
|
|
||||||
// speaks the v2 protocol.
|
|
||||||
pd.pushState.confirmedV2 = true
|
|
||||||
|
|
||||||
descriptor := distribution.Descriptor{
|
|
||||||
Digest: pushDigest,
|
Digest: pushDigest,
|
||||||
MediaType: schema2.MediaTypeLayer,
|
MediaType: schema2.MediaTypeLayer,
|
||||||
Size: nn,
|
Size: nn,
|
||||||
}
|
}
|
||||||
pd.pushState.remoteLayers[diffID] = descriptor
|
|
||||||
|
|
||||||
|
pd.pushState.Lock()
|
||||||
|
// If Commit succeeded, that's an indication that the remote registry speaks the v2 protocol.
|
||||||
|
pd.pushState.confirmedV2 = true
|
||||||
|
pd.pushState.remoteLayers[diffID] = desc
|
||||||
pd.pushState.Unlock()
|
pd.pushState.Unlock()
|
||||||
|
|
||||||
return descriptor, nil
|
return desc, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (pd *v2PushDescriptor) SetRemoteDescriptor(descriptor distribution.Descriptor) {
|
// layerAlreadyExists checks if the registry already knows about any of the metadata passed in the "metadata"
|
||||||
pd.remoteDescriptor = descriptor
|
// slice. If it finds one that the registry knows about, it returns the known digest and "true". If
|
||||||
}
|
// "checkOtherRepositories" is true, stat will be performed also with digests mapped to any other repository
|
||||||
|
// (not just the target one).
|
||||||
func (pd *v2PushDescriptor) Descriptor() distribution.Descriptor {
|
func (pd *v2PushDescriptor) layerAlreadyExists(
|
||||||
return pd.remoteDescriptor
|
ctx context.Context,
|
||||||
}
|
progressOutput progress.Output,
|
||||||
|
diffID layer.DiffID,
|
||||||
// layerAlreadyExists checks if the registry already know about any of the
|
checkOtherRepositories bool,
|
||||||
// metadata passed in the "metadata" slice. If it finds one that the registry
|
maxExistenceCheckAttempts int,
|
||||||
// knows about, it returns the known digest and "true".
|
v2Metadata []metadata.V2Metadata,
|
||||||
func layerAlreadyExists(ctx context.Context, metadata []metadata.V2Metadata, repoInfo reference.Named, repo distribution.Repository, pushState *pushState) (distribution.Descriptor, bool, error) {
|
) (desc distribution.Descriptor, exists bool, err error) {
|
||||||
for _, meta := range metadata {
|
// filter the metadata
|
||||||
// Only check blobsums that are known to this repository or have an unknown source
|
candidates := []metadata.V2Metadata{}
|
||||||
if meta.SourceRepository != "" && meta.SourceRepository != repoInfo.FullName() {
|
for _, meta := range v2Metadata {
|
||||||
|
if len(meta.SourceRepository) > 0 && !checkOtherRepositories && meta.SourceRepository != pd.repoInfo.FullName() {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
descriptor, err := repo.Blobs(ctx).Stat(ctx, meta.Digest)
|
candidates = append(candidates, meta)
|
||||||
|
}
|
||||||
|
// sort the candidates by similarity
|
||||||
|
sortV2MetadataByLikenessAndAge(pd.repoInfo, pd.hmacKey, candidates)
|
||||||
|
|
||||||
|
digestToMetadata := make(map[digest.Digest]*metadata.V2Metadata)
|
||||||
|
// an array of unique blob digests ordered from the best mount candidates to worst
|
||||||
|
layerDigests := []digest.Digest{}
|
||||||
|
for i := 0; i < len(candidates); i++ {
|
||||||
|
if len(layerDigests) >= maxExistenceCheckAttempts {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
meta := &candidates[i]
|
||||||
|
if _, exists := digestToMetadata[meta.Digest]; exists {
|
||||||
|
// keep reference just to the first mapping (the best mount candidate)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if _, exists := pd.checkedDigests[meta.Digest]; exists {
|
||||||
|
// existence of this digest has already been tested
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
digestToMetadata[meta.Digest] = meta
|
||||||
|
layerDigests = append(layerDigests, meta.Digest)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, dgst := range layerDigests {
|
||||||
|
meta := digestToMetadata[dgst]
|
||||||
|
logrus.Debugf("Checking for presence of layer %s (%s) in %s", diffID, dgst, pd.repoInfo.FullName())
|
||||||
|
desc, err = pd.repo.Blobs(ctx).Stat(ctx, dgst)
|
||||||
|
pd.checkedDigests[meta.Digest] = struct{}{}
|
||||||
switch err {
|
switch err {
|
||||||
case nil:
|
case nil:
|
||||||
descriptor.MediaType = schema2.MediaTypeLayer
|
if m, ok := digestToMetadata[desc.Digest]; !ok || m.SourceRepository != pd.repoInfo.FullName() || !metadata.CheckV2MetadataHMAC(m, pd.hmacKey) {
|
||||||
return descriptor, true, nil
|
// cache mapping from this layer's DiffID to the blobsum
|
||||||
|
if err := pd.v2MetadataService.TagAndAdd(diffID, pd.hmacKey, metadata.V2Metadata{
|
||||||
|
Digest: desc.Digest,
|
||||||
|
SourceRepository: pd.repoInfo.FullName(),
|
||||||
|
}); err != nil {
|
||||||
|
return distribution.Descriptor{}, false, xfer.DoNotRetry{Err: err}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
desc.MediaType = schema2.MediaTypeLayer
|
||||||
|
exists = true
|
||||||
|
break
|
||||||
case distribution.ErrBlobUnknown:
|
case distribution.ErrBlobUnknown:
|
||||||
// nop
|
if meta.SourceRepository == pd.repoInfo.FullName() {
|
||||||
|
// remove the mapping to the target repository
|
||||||
|
pd.v2MetadataService.Remove(*meta)
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
return distribution.Descriptor{}, false, err
|
progress.Update(progressOutput, pd.ID(), "Image push failed")
|
||||||
|
return desc, false, retryOnError(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if exists {
|
||||||
|
progress.Update(progressOutput, pd.ID(), "Layer already exists")
|
||||||
|
pd.pushState.Lock()
|
||||||
|
pd.pushState.remoteLayers[diffID] = desc
|
||||||
|
pd.pushState.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
return desc, exists, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// getMaxMountAndExistenceCheckAttempts returns a maximum number of cross repository mount attempts from
|
||||||
|
// source repositories of target registry, maximum number of layer existence checks performed on the target
|
||||||
|
// repository and whether the check shall be done also with digests mapped to different repositories. The
|
||||||
|
// decision is based on layer size. The smaller the layer, the fewer attempts shall be made because the cost
|
||||||
|
// of upload does not outweigh a latency.
|
||||||
|
func getMaxMountAndExistenceCheckAttempts(layer layer.Layer) (maxMountAttempts, maxExistenceCheckAttempts int, checkOtherRepositories bool) {
|
||||||
|
size, err := layer.DiffSize()
|
||||||
|
switch {
|
||||||
|
// big blob
|
||||||
|
case size > middleLayerMaximumSize:
|
||||||
|
// 1st attempt to mount the blob few times
|
||||||
|
// 2nd few existence checks with digests associated to any repository
|
||||||
|
// then fallback to upload
|
||||||
|
return 4, 3, true
|
||||||
|
|
||||||
|
// middle sized blobs; if we could not get the size, assume we deal with middle sized blob
|
||||||
|
case size > smallLayerMaximumSize, err != nil:
|
||||||
|
// 1st attempt to mount blobs of average size few times
|
||||||
|
// 2nd try at most 1 existence check if there's an existing mapping to the target repository
|
||||||
|
// then fallback to upload
|
||||||
|
return 3, 1, false
|
||||||
|
|
||||||
|
// small blobs, do a minimum number of checks
|
||||||
|
default:
|
||||||
|
return 1, 1, false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getRepositoryMountCandidates returns an array of v2 metadata items belonging to the given registry. The
|
||||||
|
// array is sorted from youngest to oldest. If requireReigstryMatch is true, the resulting array will contain
|
||||||
|
// only metadata entries having registry part of SourceRepository matching the part of repoInfo.
|
||||||
|
func getRepositoryMountCandidates(
|
||||||
|
repoInfo reference.Named,
|
||||||
|
hmacKey []byte,
|
||||||
|
max int,
|
||||||
|
v2Metadata []metadata.V2Metadata,
|
||||||
|
) []metadata.V2Metadata {
|
||||||
|
candidates := []metadata.V2Metadata{}
|
||||||
|
for _, meta := range v2Metadata {
|
||||||
|
sourceRepo, err := reference.ParseNamed(meta.SourceRepository)
|
||||||
|
if err != nil || repoInfo.Hostname() != sourceRepo.Hostname() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// target repository is not a viable candidate
|
||||||
|
if meta.SourceRepository == repoInfo.FullName() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
candidates = append(candidates, meta)
|
||||||
|
}
|
||||||
|
|
||||||
|
sortV2MetadataByLikenessAndAge(repoInfo, hmacKey, candidates)
|
||||||
|
if max >= 0 && len(candidates) > max {
|
||||||
|
// select the youngest metadata
|
||||||
|
candidates = candidates[:max]
|
||||||
|
}
|
||||||
|
|
||||||
|
return candidates
|
||||||
|
}
|
||||||
|
|
||||||
|
// byLikeness is a sorting container for v2 metadata candidates for cross repository mount. The
|
||||||
|
// candidate "a" is preferred over "b":
|
||||||
|
//
|
||||||
|
// 1. if it was hashed using the same AuthConfig as the one used to authenticate to target repository and the
|
||||||
|
// "b" was not
|
||||||
|
// 2. if a number of its repository path components exactly matching path components of target repository is higher
|
||||||
|
type byLikeness struct {
|
||||||
|
arr []metadata.V2Metadata
|
||||||
|
hmacKey []byte
|
||||||
|
pathComponents []string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (bla byLikeness) Less(i, j int) bool {
|
||||||
|
aMacMatch := metadata.CheckV2MetadataHMAC(&bla.arr[i], bla.hmacKey)
|
||||||
|
bMacMatch := metadata.CheckV2MetadataHMAC(&bla.arr[j], bla.hmacKey)
|
||||||
|
if aMacMatch != bMacMatch {
|
||||||
|
return aMacMatch
|
||||||
|
}
|
||||||
|
aMatch := numOfMatchingPathComponents(bla.arr[i].SourceRepository, bla.pathComponents)
|
||||||
|
bMatch := numOfMatchingPathComponents(bla.arr[j].SourceRepository, bla.pathComponents)
|
||||||
|
return aMatch > bMatch
|
||||||
|
}
|
||||||
|
func (bla byLikeness) Swap(i, j int) {
|
||||||
|
bla.arr[i], bla.arr[j] = bla.arr[j], bla.arr[i]
|
||||||
|
}
|
||||||
|
func (bla byLikeness) Len() int { return len(bla.arr) }
|
||||||
|
|
||||||
|
func sortV2MetadataByLikenessAndAge(repoInfo reference.Named, hmacKey []byte, marr []metadata.V2Metadata) {
|
||||||
|
// reverse the metadata array to shift the newest entries to the beginning
|
||||||
|
for i := 0; i < len(marr)/2; i++ {
|
||||||
|
marr[i], marr[len(marr)-i-1] = marr[len(marr)-i-1], marr[i]
|
||||||
|
}
|
||||||
|
// keep equal entries ordered from the youngest to the oldest
|
||||||
|
sort.Stable(byLikeness{
|
||||||
|
arr: marr,
|
||||||
|
hmacKey: hmacKey,
|
||||||
|
pathComponents: getPathComponents(repoInfo.FullName()),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// numOfMatchingPathComponents returns a number of path components in "pth" that exactly match "matchComponents".
|
||||||
|
func numOfMatchingPathComponents(pth string, matchComponents []string) int {
|
||||||
|
pthComponents := getPathComponents(pth)
|
||||||
|
i := 0
|
||||||
|
for ; i < len(pthComponents) && i < len(matchComponents); i++ {
|
||||||
|
if matchComponents[i] != pthComponents[i] {
|
||||||
|
return i
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return i
|
||||||
|
}
|
||||||
|
|
||||||
|
func getPathComponents(path string) []string {
|
||||||
|
// make sure to add docker.io/ prefix to the path
|
||||||
|
named, err := reference.ParseNamed(path)
|
||||||
|
if err == nil {
|
||||||
|
path = named.FullName()
|
||||||
|
}
|
||||||
|
return strings.Split(path, "/")
|
||||||
|
}
|
||||||
|
|
||||||
|
func cancelLayerUpload(ctx context.Context, dgst digest.Digest, layerUpload distribution.BlobWriter) {
|
||||||
|
if layerUpload != nil {
|
||||||
|
logrus.Debugf("cancelling upload of blob %s", dgst)
|
||||||
|
err := layerUpload.Cancel(ctx)
|
||||||
|
if err != nil {
|
||||||
|
logrus.Warnf("failed to cancel upload: %v", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return distribution.Descriptor{}, false, nil
|
|
||||||
}
|
}
|
||||||
|
|
574
distribution/push_v2_test.go
Normal file
574
distribution/push_v2_test.go
Normal file
|
@ -0,0 +1,574 @@
|
||||||
|
package distribution
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net/http"
|
||||||
|
"reflect"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/docker/distribution"
|
||||||
|
"github.com/docker/distribution/context"
|
||||||
|
"github.com/docker/distribution/digest"
|
||||||
|
"github.com/docker/distribution/manifest/schema2"
|
||||||
|
distreference "github.com/docker/distribution/reference"
|
||||||
|
"github.com/docker/docker/distribution/metadata"
|
||||||
|
"github.com/docker/docker/layer"
|
||||||
|
"github.com/docker/docker/pkg/progress"
|
||||||
|
"github.com/docker/docker/reference"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestGetRepositoryMountCandidates(t *testing.T) {
|
||||||
|
for _, tc := range []struct {
|
||||||
|
name string
|
||||||
|
hmacKey string
|
||||||
|
targetRepo string
|
||||||
|
maxCandidates int
|
||||||
|
metadata []metadata.V2Metadata
|
||||||
|
candidates []metadata.V2Metadata
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "empty metadata",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{},
|
||||||
|
candidates: []metadata.V2Metadata{},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one item not matching",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{taggedMetadata("key", "dgst", "127.0.0.1/repo")},
|
||||||
|
candidates: []metadata.V2Metadata{},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "one item matching",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{taggedMetadata("hash", "1", "hello-world")},
|
||||||
|
candidates: []metadata.V2Metadata{taggedMetadata("hash", "1", "hello-world")},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "allow missing SourceRepository",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("1")},
|
||||||
|
{Digest: digest.Digest("3")},
|
||||||
|
{Digest: digest.Digest("2")},
|
||||||
|
},
|
||||||
|
candidates: []metadata.V2Metadata{},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "handle docker.io",
|
||||||
|
targetRepo: "user/app",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("1"), SourceRepository: "docker.io/user/foo"},
|
||||||
|
{Digest: digest.Digest("3"), SourceRepository: "user/bar"},
|
||||||
|
{Digest: digest.Digest("2"), SourceRepository: "app"},
|
||||||
|
},
|
||||||
|
candidates: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("3"), SourceRepository: "user/bar"},
|
||||||
|
{Digest: digest.Digest("1"), SourceRepository: "docker.io/user/foo"},
|
||||||
|
{Digest: digest.Digest("2"), SourceRepository: "app"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "sort more items",
|
||||||
|
hmacKey: "abcd",
|
||||||
|
targetRepo: "127.0.0.1/foo/bar",
|
||||||
|
maxCandidates: -1,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("hash", "1", "hello-world"),
|
||||||
|
taggedMetadata("efgh", "2", "127.0.0.1/hello-world"),
|
||||||
|
taggedMetadata("abcd", "3", "busybox"),
|
||||||
|
taggedMetadata("hash", "4", "busybox"),
|
||||||
|
taggedMetadata("hash", "5", "127.0.0.1/foo"),
|
||||||
|
taggedMetadata("hash", "6", "127.0.0.1/bar"),
|
||||||
|
taggedMetadata("efgh", "7", "127.0.0.1/foo/bar"),
|
||||||
|
taggedMetadata("abcd", "8", "127.0.0.1/xyz"),
|
||||||
|
taggedMetadata("hash", "9", "127.0.0.1/foo/app"),
|
||||||
|
},
|
||||||
|
candidates: []metadata.V2Metadata{
|
||||||
|
// first by matching hash
|
||||||
|
taggedMetadata("abcd", "8", "127.0.0.1/xyz"),
|
||||||
|
// then by longest matching prefix
|
||||||
|
taggedMetadata("hash", "9", "127.0.0.1/foo/app"),
|
||||||
|
taggedMetadata("hash", "5", "127.0.0.1/foo"),
|
||||||
|
// sort the rest of the matching items in reversed order
|
||||||
|
taggedMetadata("hash", "6", "127.0.0.1/bar"),
|
||||||
|
taggedMetadata("efgh", "2", "127.0.0.1/hello-world"),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "limit max candidates",
|
||||||
|
hmacKey: "abcd",
|
||||||
|
targetRepo: "user/app",
|
||||||
|
maxCandidates: 3,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("abcd", "1", "user/app1"),
|
||||||
|
taggedMetadata("abcd", "2", "user/app/base"),
|
||||||
|
taggedMetadata("hash", "3", "user/app"),
|
||||||
|
taggedMetadata("abcd", "4", "127.0.0.1/user/app"),
|
||||||
|
taggedMetadata("hash", "5", "user/foo"),
|
||||||
|
taggedMetadata("hash", "6", "app/bar"),
|
||||||
|
},
|
||||||
|
candidates: []metadata.V2Metadata{
|
||||||
|
// first by matching hash
|
||||||
|
taggedMetadata("abcd", "2", "user/app/base"),
|
||||||
|
taggedMetadata("abcd", "1", "user/app1"),
|
||||||
|
// then by longest matching prefix
|
||||||
|
taggedMetadata("hash", "3", "user/app"),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
} {
|
||||||
|
repoInfo, err := reference.ParseNamed(tc.targetRepo)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("[%s] failed to parse reference name: %v", tc.name, err)
|
||||||
|
}
|
||||||
|
candidates := getRepositoryMountCandidates(repoInfo, []byte(tc.hmacKey), tc.maxCandidates, tc.metadata)
|
||||||
|
if len(candidates) != len(tc.candidates) {
|
||||||
|
t.Errorf("[%s] got unexpected number of candidates: %d != %d", tc.name, len(candidates), len(tc.candidates))
|
||||||
|
}
|
||||||
|
for i := 0; i < len(candidates) && i < len(tc.candidates); i++ {
|
||||||
|
if !reflect.DeepEqual(candidates[i], tc.candidates[i]) {
|
||||||
|
t.Errorf("[%s] candidate %d does not match expected: %#+v != %#+v", tc.name, i, candidates[i], tc.candidates[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i := len(candidates); i < len(tc.candidates); i++ {
|
||||||
|
t.Errorf("[%s] missing expected candidate at position %d (%#+v)", tc.name, i, tc.candidates[i])
|
||||||
|
}
|
||||||
|
for i := len(tc.candidates); i < len(candidates); i++ {
|
||||||
|
t.Errorf("[%s] got unexpected candidate at position %d (%#+v)", tc.name, i, candidates[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestLayerAlreadyExists(t *testing.T) {
|
||||||
|
for _, tc := range []struct {
|
||||||
|
name string
|
||||||
|
metadata []metadata.V2Metadata
|
||||||
|
targetRepo string
|
||||||
|
hmacKey string
|
||||||
|
maxExistenceChecks int
|
||||||
|
checkOtherRepositories bool
|
||||||
|
remoteBlobs map[digest.Digest]distribution.Descriptor
|
||||||
|
remoteErrors map[digest.Digest]error
|
||||||
|
expectedDescriptor distribution.Descriptor
|
||||||
|
expectedExists bool
|
||||||
|
expectedError error
|
||||||
|
expectedRequests []string
|
||||||
|
expectedAdditions []metadata.V2Metadata
|
||||||
|
expectedRemovals []metadata.V2Metadata
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "empty metadata",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
checkOtherRepositories: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "single not existent metadata",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{{Digest: digest.Digest("pear"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
expectedRequests: []string{"pear"},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{{Digest: digest.Digest("pear"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "access denied",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxExistenceChecks: 1,
|
||||||
|
metadata: []metadata.V2Metadata{{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
remoteErrors: map[digest.Digest]error{digest.Digest("apple"): distribution.ErrAccessDenied},
|
||||||
|
expectedError: distribution.ErrAccessDenied,
|
||||||
|
expectedRequests: []string{"apple"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "not matching reposies",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/hello-world"},
|
||||||
|
{Digest: digest.Digest("orange"), SourceRepository: "docker.io/library/busybox/subapp"},
|
||||||
|
{Digest: digest.Digest("pear"), SourceRepository: "docker.io/busybox"},
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "busybox"},
|
||||||
|
{Digest: digest.Digest("banana"), SourceRepository: "127.0.0.1/busybox"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "check other repositories",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
maxExistenceChecks: 10,
|
||||||
|
checkOtherRepositories: true,
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/hello-world"},
|
||||||
|
{Digest: digest.Digest("orange"), SourceRepository: "docker.io/library/busybox/subapp"},
|
||||||
|
{Digest: digest.Digest("pear"), SourceRepository: "docker.io/busybox"},
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "busybox"},
|
||||||
|
{Digest: digest.Digest("banana"), SourceRepository: "127.0.0.1/busybox"},
|
||||||
|
},
|
||||||
|
expectedRequests: []string{"plum", "pear", "apple", "orange", "banana"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "find existing blob",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("apple"): {Digest: digest.Digest("apple")}},
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("apple"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"apple"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "find existing blob with different hmac",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{{SourceRepository: "docker.io/library/busybox", Digest: digest.Digest("apple"), HMAC: "dummyhmac"}},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("apple"): {Digest: digest.Digest("apple")}},
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("apple"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"apple"},
|
||||||
|
expectedAdditions: []metadata.V2Metadata{{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "overwrite media types",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{{Digest: digest.Digest("apple"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
hmacKey: "key",
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("apple"): {Digest: digest.Digest("apple"), MediaType: "custom-media-type"}},
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("apple"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"apple"},
|
||||||
|
expectedAdditions: []metadata.V2Metadata{taggedMetadata("key", "apple", "docker.io/library/busybox")},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "find existing blob among many",
|
||||||
|
targetRepo: "127.0.0.1/myapp",
|
||||||
|
hmacKey: "key",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("someotherkey", "pear", "127.0.0.1/myapp"),
|
||||||
|
taggedMetadata("key", "apple", "127.0.0.1/myapp"),
|
||||||
|
taggedMetadata("", "plum", "127.0.0.1/myapp"),
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("pear"): {Digest: digest.Digest("pear")}},
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("pear"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"apple", "plum", "pear"},
|
||||||
|
expectedAdditions: []metadata.V2Metadata{taggedMetadata("key", "pear", "127.0.0.1/myapp")},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("key", "apple", "127.0.0.1/myapp"),
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "127.0.0.1/myapp"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "reach maximum existence checks",
|
||||||
|
targetRepo: "user/app",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("pear"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("apple"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("banana"), SourceRepository: "docker.io/user/app"},
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("pear"): {Digest: digest.Digest("pear")}},
|
||||||
|
expectedExists: false,
|
||||||
|
expectedRequests: []string{"banana", "plum", "apple"},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("banana"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("apple"), SourceRepository: "docker.io/user/app"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "zero allowed existence checks",
|
||||||
|
targetRepo: "user/app",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("pear"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("apple"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "docker.io/user/app"},
|
||||||
|
{Digest: digest.Digest("banana"), SourceRepository: "docker.io/user/app"},
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 0,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("pear"): {Digest: digest.Digest("pear")}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "stat single digest just once",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("key1", "pear", "docker.io/library/busybox"),
|
||||||
|
taggedMetadata("key2", "apple", "docker.io/library/busybox"),
|
||||||
|
taggedMetadata("key3", "apple", "docker.io/library/busybox"),
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("pear"): {Digest: digest.Digest("pear")}},
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("pear"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"apple", "pear"},
|
||||||
|
expectedAdditions: []metadata.V2Metadata{{Digest: digest.Digest("pear"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{taggedMetadata("key3", "apple", "docker.io/library/busybox")},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "stop on first error",
|
||||||
|
targetRepo: "user/app",
|
||||||
|
hmacKey: "key",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
taggedMetadata("key", "banana", "docker.io/user/app"),
|
||||||
|
taggedMetadata("key", "orange", "docker.io/user/app"),
|
||||||
|
taggedMetadata("key", "plum", "docker.io/user/app"),
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteErrors: map[digest.Digest]error{"orange": distribution.ErrAccessDenied},
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("apple"): {}},
|
||||||
|
expectedError: distribution.ErrAccessDenied,
|
||||||
|
expectedRequests: []string{"plum", "orange"},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{taggedMetadata("key", "plum", "docker.io/user/app")},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "remove outdated metadata",
|
||||||
|
targetRepo: "docker.io/user/app",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("plum"), SourceRepository: "docker.io/library/busybox"},
|
||||||
|
{Digest: digest.Digest("orange"), SourceRepository: "docker.io/user/app"},
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
remoteErrors: map[digest.Digest]error{"orange": distribution.ErrBlobUnknown},
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("plum"): {}},
|
||||||
|
expectedExists: false,
|
||||||
|
expectedRequests: []string{"orange"},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{{Digest: digest.Digest("orange"), SourceRepository: "docker.io/user/app"}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "missing SourceRepository",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("1")},
|
||||||
|
{Digest: digest.Digest("3")},
|
||||||
|
{Digest: digest.Digest("2")},
|
||||||
|
},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
expectedExists: false,
|
||||||
|
expectedRequests: []string{"2", "3", "1"},
|
||||||
|
},
|
||||||
|
|
||||||
|
{
|
||||||
|
name: "with and without SourceRepository",
|
||||||
|
targetRepo: "busybox",
|
||||||
|
metadata: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("1")},
|
||||||
|
{Digest: digest.Digest("2"), SourceRepository: "docker.io/library/busybox"},
|
||||||
|
{Digest: digest.Digest("3")},
|
||||||
|
},
|
||||||
|
remoteBlobs: map[digest.Digest]distribution.Descriptor{digest.Digest("1"): {Digest: digest.Digest("1")}},
|
||||||
|
maxExistenceChecks: 3,
|
||||||
|
expectedDescriptor: distribution.Descriptor{Digest: digest.Digest("1"), MediaType: schema2.MediaTypeLayer},
|
||||||
|
expectedExists: true,
|
||||||
|
expectedRequests: []string{"2", "3", "1"},
|
||||||
|
expectedAdditions: []metadata.V2Metadata{{Digest: digest.Digest("1"), SourceRepository: "docker.io/library/busybox"}},
|
||||||
|
expectedRemovals: []metadata.V2Metadata{
|
||||||
|
{Digest: digest.Digest("2"), SourceRepository: "docker.io/library/busybox"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
} {
|
||||||
|
repoInfo, err := reference.ParseNamed(tc.targetRepo)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("[%s] failed to parse reference name: %v", tc.name, err)
|
||||||
|
}
|
||||||
|
repo := &mockRepo{
|
||||||
|
t: t,
|
||||||
|
errors: tc.remoteErrors,
|
||||||
|
blobs: tc.remoteBlobs,
|
||||||
|
requests: []string{},
|
||||||
|
}
|
||||||
|
ctx := context.Background()
|
||||||
|
ms := &mockV2MetadataService{}
|
||||||
|
pd := &v2PushDescriptor{
|
||||||
|
hmacKey: []byte(tc.hmacKey),
|
||||||
|
repoInfo: repoInfo,
|
||||||
|
layer: layer.EmptyLayer,
|
||||||
|
repo: repo,
|
||||||
|
v2MetadataService: ms,
|
||||||
|
pushState: &pushState{remoteLayers: make(map[layer.DiffID]distribution.Descriptor)},
|
||||||
|
checkedDigests: make(map[digest.Digest]struct{}),
|
||||||
|
}
|
||||||
|
|
||||||
|
desc, exists, err := pd.layerAlreadyExists(ctx, &progressSink{t}, layer.EmptyLayer.DiffID(), tc.checkOtherRepositories, tc.maxExistenceChecks, tc.metadata)
|
||||||
|
|
||||||
|
if !reflect.DeepEqual(desc, tc.expectedDescriptor) {
|
||||||
|
t.Errorf("[%s] got unexpected descriptor: %#+v != %#+v", tc.name, desc, tc.expectedDescriptor)
|
||||||
|
}
|
||||||
|
if exists != tc.expectedExists {
|
||||||
|
t.Errorf("[%s] got unexpected exists: %t != %t", tc.name, exists, tc.expectedExists)
|
||||||
|
}
|
||||||
|
if !reflect.DeepEqual(err, tc.expectedError) {
|
||||||
|
t.Errorf("[%s] got unexpected error: %#+v != %#+v", tc.name, err, tc.expectedError)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(repo.requests) != len(tc.expectedRequests) {
|
||||||
|
t.Errorf("[%s] got unexpected number of requests: %d != %d", tc.name, len(repo.requests), len(tc.expectedRequests))
|
||||||
|
}
|
||||||
|
for i := 0; i < len(repo.requests) && i < len(tc.expectedRequests); i++ {
|
||||||
|
if repo.requests[i] != tc.expectedRequests[i] {
|
||||||
|
t.Errorf("[%s] request %d does not match expected: %q != %q", tc.name, i, repo.requests[i], tc.expectedRequests[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i := len(repo.requests); i < len(tc.expectedRequests); i++ {
|
||||||
|
t.Errorf("[%s] missing expected request at position %d (%q)", tc.name, i, tc.expectedRequests[i])
|
||||||
|
}
|
||||||
|
for i := len(tc.expectedRequests); i < len(repo.requests); i++ {
|
||||||
|
t.Errorf("[%s] got unexpected request at position %d (%q)", tc.name, i, repo.requests[i])
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(ms.added) != len(tc.expectedAdditions) {
|
||||||
|
t.Errorf("[%s] got unexpected number of additions: %d != %d", tc.name, len(ms.added), len(tc.expectedAdditions))
|
||||||
|
}
|
||||||
|
for i := 0; i < len(ms.added) && i < len(tc.expectedAdditions); i++ {
|
||||||
|
if ms.added[i] != tc.expectedAdditions[i] {
|
||||||
|
t.Errorf("[%s] added metadata at %d does not match expected: %q != %q", tc.name, i, ms.added[i], tc.expectedAdditions[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i := len(ms.added); i < len(tc.expectedAdditions); i++ {
|
||||||
|
t.Errorf("[%s] missing expected addition at position %d (%q)", tc.name, i, tc.expectedAdditions[i])
|
||||||
|
}
|
||||||
|
for i := len(tc.expectedAdditions); i < len(ms.added); i++ {
|
||||||
|
t.Errorf("[%s] unexpected metadata addition at position %d (%q)", tc.name, i, ms.added[i])
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(ms.removed) != len(tc.expectedRemovals) {
|
||||||
|
t.Errorf("[%s] got unexpected number of removals: %d != %d", tc.name, len(ms.removed), len(tc.expectedRemovals))
|
||||||
|
}
|
||||||
|
for i := 0; i < len(ms.removed) && i < len(tc.expectedRemovals); i++ {
|
||||||
|
if ms.removed[i] != tc.expectedRemovals[i] {
|
||||||
|
t.Errorf("[%s] removed metadata at %d does not match expected: %q != %q", tc.name, i, ms.removed[i], tc.expectedRemovals[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i := len(ms.removed); i < len(tc.expectedRemovals); i++ {
|
||||||
|
t.Errorf("[%s] missing expected removal at position %d (%q)", tc.name, i, tc.expectedRemovals[i])
|
||||||
|
}
|
||||||
|
for i := len(tc.expectedRemovals); i < len(ms.removed); i++ {
|
||||||
|
t.Errorf("[%s] removed unexpected metadata at position %d (%q)", tc.name, i, ms.removed[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func taggedMetadata(key string, dgst string, sourceRepo string) metadata.V2Metadata {
|
||||||
|
meta := metadata.V2Metadata{
|
||||||
|
Digest: digest.Digest(dgst),
|
||||||
|
SourceRepository: sourceRepo,
|
||||||
|
}
|
||||||
|
|
||||||
|
meta.HMAC = metadata.ComputeV2MetadataHMAC([]byte(key), &meta)
|
||||||
|
return meta
|
||||||
|
}
|
||||||
|
|
||||||
|
type mockRepo struct {
|
||||||
|
t *testing.T
|
||||||
|
errors map[digest.Digest]error
|
||||||
|
blobs map[digest.Digest]distribution.Descriptor
|
||||||
|
requests []string
|
||||||
|
}
|
||||||
|
|
||||||
|
var _ distribution.Repository = &mockRepo{}
|
||||||
|
|
||||||
|
func (m *mockRepo) Named() distreference.Named {
|
||||||
|
m.t.Fatalf("Named() not implemented")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (m *mockRepo) Manifests(ctc context.Context, options ...distribution.ManifestServiceOption) (distribution.ManifestService, error) {
|
||||||
|
m.t.Fatalf("Manifests() not implemented")
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
func (m *mockRepo) Tags(ctc context.Context) distribution.TagService {
|
||||||
|
m.t.Fatalf("Tags() not implemented")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (m *mockRepo) Blobs(ctx context.Context) distribution.BlobStore {
|
||||||
|
return &mockBlobStore{
|
||||||
|
repo: m,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type mockBlobStore struct {
|
||||||
|
repo *mockRepo
|
||||||
|
}
|
||||||
|
|
||||||
|
var _ distribution.BlobStore = &mockBlobStore{}
|
||||||
|
|
||||||
|
func (m *mockBlobStore) Stat(ctx context.Context, dgst digest.Digest) (distribution.Descriptor, error) {
|
||||||
|
m.repo.requests = append(m.repo.requests, dgst.String())
|
||||||
|
if err, exists := m.repo.errors[dgst]; exists {
|
||||||
|
return distribution.Descriptor{}, err
|
||||||
|
}
|
||||||
|
if desc, exists := m.repo.blobs[dgst]; exists {
|
||||||
|
return desc, nil
|
||||||
|
}
|
||||||
|
return distribution.Descriptor{}, distribution.ErrBlobUnknown
|
||||||
|
}
|
||||||
|
func (m *mockBlobStore) Get(ctx context.Context, dgst digest.Digest) ([]byte, error) {
|
||||||
|
m.repo.t.Fatal("Get() not implemented")
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *mockBlobStore) Open(ctx context.Context, dgst digest.Digest) (distribution.ReadSeekCloser, error) {
|
||||||
|
m.repo.t.Fatal("Open() not implemented")
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *mockBlobStore) Put(ctx context.Context, mediaType string, p []byte) (distribution.Descriptor, error) {
|
||||||
|
m.repo.t.Fatal("Put() not implemented")
|
||||||
|
return distribution.Descriptor{}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *mockBlobStore) Create(ctx context.Context, options ...distribution.BlobCreateOption) (distribution.BlobWriter, error) {
|
||||||
|
m.repo.t.Fatal("Create() not implemented")
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
func (m *mockBlobStore) Resume(ctx context.Context, id string) (distribution.BlobWriter, error) {
|
||||||
|
m.repo.t.Fatal("Resume() not implemented")
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
func (m *mockBlobStore) Delete(ctx context.Context, dgst digest.Digest) error {
|
||||||
|
m.repo.t.Fatal("Delete() not implemented")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (m *mockBlobStore) ServeBlob(ctx context.Context, w http.ResponseWriter, r *http.Request, dgst digest.Digest) error {
|
||||||
|
m.repo.t.Fatalf("ServeBlob() not implemented")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type mockV2MetadataService struct {
|
||||||
|
added []metadata.V2Metadata
|
||||||
|
removed []metadata.V2Metadata
|
||||||
|
}
|
||||||
|
|
||||||
|
var _ metadata.V2MetadataService = &mockV2MetadataService{}
|
||||||
|
|
||||||
|
func (*mockV2MetadataService) GetMetadata(diffID layer.DiffID) ([]metadata.V2Metadata, error) {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
func (*mockV2MetadataService) GetDiffID(dgst digest.Digest) (layer.DiffID, error) {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
func (m *mockV2MetadataService) Add(diffID layer.DiffID, metadata metadata.V2Metadata) error {
|
||||||
|
m.added = append(m.added, metadata)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (m *mockV2MetadataService) TagAndAdd(diffID layer.DiffID, hmacKey []byte, meta metadata.V2Metadata) error {
|
||||||
|
meta.HMAC = metadata.ComputeV2MetadataHMAC(hmacKey, &meta)
|
||||||
|
m.Add(diffID, meta)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (m *mockV2MetadataService) Remove(metadata metadata.V2Metadata) error {
|
||||||
|
m.removed = append(m.removed, metadata)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type progressSink struct {
|
||||||
|
t *testing.T
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *progressSink) WriteProgress(p progress.Progress) error {
|
||||||
|
s.t.Logf("progress update: %#+v", p)
|
||||||
|
return nil
|
||||||
|
}
|
Loading…
Add table
Reference in a new issue