// +build linux package plugin import ( "archive/tar" "compress/gzip" "encoding/json" "fmt" "io" "io/ioutil" "net/http" "os" "path" "path/filepath" "strings" "github.com/Sirupsen/logrus" "github.com/docker/distribution/manifest/schema2" "github.com/docker/docker/api/types" "github.com/docker/docker/distribution" progressutils "github.com/docker/docker/distribution/utils" "github.com/docker/docker/distribution/xfer" "github.com/docker/docker/image" "github.com/docker/docker/layer" "github.com/docker/docker/pkg/chrootarchive" "github.com/docker/docker/pkg/pools" "github.com/docker/docker/pkg/progress" "github.com/docker/docker/plugin/v2" "github.com/docker/docker/reference" "github.com/opencontainers/go-digest" "github.com/pkg/errors" "golang.org/x/net/context" ) // Disable deactivates a plugin. This means resources (volumes, networks) cant use them. func (pm *Manager) Disable(refOrID string, config *types.PluginDisableConfig) error { p, err := pm.config.Store.GetV2Plugin(refOrID) if err != nil { return err } pm.mu.RLock() c := pm.cMap[p] pm.mu.RUnlock() if !config.ForceDisable && p.GetRefCount() > 0 { return fmt.Errorf("plugin %s is in use", p.Name()) } if err := pm.disable(p, c); err != nil { return err } pm.config.LogPluginEvent(p.GetID(), refOrID, "disable") return nil } // Enable activates a plugin, which implies that they are ready to be used by containers. func (pm *Manager) Enable(refOrID string, config *types.PluginEnableConfig) error { p, err := pm.config.Store.GetV2Plugin(refOrID) if err != nil { return err } c := &controller{timeoutInSecs: config.Timeout} if err := pm.enable(p, c, false); err != nil { return err } pm.config.LogPluginEvent(p.GetID(), refOrID, "enable") return nil } // Inspect examines a plugin config func (pm *Manager) Inspect(refOrID string) (tp *types.Plugin, err error) { p, err := pm.config.Store.GetV2Plugin(refOrID) if err != nil { return nil, err } return &p.PluginObj, nil } func (pm *Manager) pull(ctx context.Context, ref reference.Named, config *distribution.ImagePullConfig, outStream io.Writer) error { if outStream != nil { // Include a buffer so that slow client connections don't affect // transfer performance. progressChan := make(chan progress.Progress, 100) writesDone := make(chan struct{}) defer func() { close(progressChan) <-writesDone }() var cancelFunc context.CancelFunc ctx, cancelFunc = context.WithCancel(ctx) go func() { progressutils.WriteDistributionProgress(cancelFunc, outStream, progressChan) close(writesDone) }() config.ProgressOutput = progress.ChanOutput(progressChan) } else { config.ProgressOutput = progress.DiscardOutput() } return distribution.Pull(ctx, ref, config) } type tempConfigStore struct { config []byte configDigest digest.Digest } func (s *tempConfigStore) Put(c []byte) (digest.Digest, error) { dgst := digest.FromBytes(c) s.config = c s.configDigest = dgst return dgst, nil } func (s *tempConfigStore) Get(d digest.Digest) ([]byte, error) { if d != s.configDigest { return nil, fmt.Errorf("digest not found") } return s.config, nil } func (s *tempConfigStore) RootFSFromConfig(c []byte) (*image.RootFS, error) { return configToRootFS(c) } func computePrivileges(c types.PluginConfig) (types.PluginPrivileges, error) { var privileges types.PluginPrivileges if c.Network.Type != "null" && c.Network.Type != "bridge" && c.Network.Type != "" { privileges = append(privileges, types.PluginPrivilege{ Name: "network", Description: "permissions to access a network", Value: []string{c.Network.Type}, }) } for _, mount := range c.Mounts { if mount.Source != nil { privileges = append(privileges, types.PluginPrivilege{ Name: "mount", Description: "host path to mount", Value: []string{*mount.Source}, }) } } for _, device := range c.Linux.Devices { if device.Path != nil { privileges = append(privileges, types.PluginPrivilege{ Name: "device", Description: "host device to access", Value: []string{*device.Path}, }) } } if c.Linux.DeviceCreation { privileges = append(privileges, types.PluginPrivilege{ Name: "device-creation", Description: "allow creating devices inside plugin", Value: []string{"true"}, }) } if len(c.Linux.Capabilities) > 0 { privileges = append(privileges, types.PluginPrivilege{ Name: "capabilities", Description: "list of additional capabilities required", Value: c.Linux.Capabilities, }) } return privileges, nil } // Privileges pulls a plugin config and computes the privileges required to install it. func (pm *Manager) Privileges(ctx context.Context, ref reference.Named, metaHeader http.Header, authConfig *types.AuthConfig) (types.PluginPrivileges, error) { // create image store instance cs := &tempConfigStore{} // DownloadManager not defined because only pulling configuration. pluginPullConfig := &distribution.ImagePullConfig{ Config: distribution.Config{ MetaHeaders: metaHeader, AuthConfig: authConfig, RegistryService: pm.config.RegistryService, ImageEventLogger: func(string, string, string) {}, ImageStore: cs, }, Schema2Types: distribution.PluginTypes, } if err := pm.pull(ctx, ref, pluginPullConfig, nil); err != nil { return nil, err } if cs.config == nil { return nil, errors.New("no configuration pulled") } var config types.PluginConfig if err := json.Unmarshal(cs.config, &config); err != nil { return nil, err } return computePrivileges(config) } // Pull pulls a plugin, check if the correct privileges are provided and install the plugin. func (pm *Manager) Pull(ctx context.Context, ref reference.Named, name string, metaHeader http.Header, authConfig *types.AuthConfig, privileges types.PluginPrivileges, outStream io.Writer) (err error) { pm.muGC.RLock() defer pm.muGC.RUnlock() // revalidate because Pull is public nameref, err := reference.ParseNamed(name) if err != nil { return errors.Wrapf(err, "failed to parse %q", name) } name = reference.WithDefaultTag(nameref).String() if err := pm.config.Store.validateName(name); err != nil { return err } tmpRootFSDir, err := ioutil.TempDir(pm.tmpDir(), ".rootfs") defer os.RemoveAll(tmpRootFSDir) dm := &downloadManager{ tmpDir: tmpRootFSDir, blobStore: pm.blobStore, } pluginPullConfig := &distribution.ImagePullConfig{ Config: distribution.Config{ MetaHeaders: metaHeader, AuthConfig: authConfig, RegistryService: pm.config.RegistryService, ImageEventLogger: pm.config.LogPluginEvent, ImageStore: dm, }, DownloadManager: dm, // todo: reevaluate if possible to substitute distribution/xfer dependencies instead Schema2Types: distribution.PluginTypes, } err = pm.pull(ctx, ref, pluginPullConfig, outStream) if err != nil { go pm.GC() return err } if _, err := pm.createPlugin(name, dm.configDigest, dm.blobs, tmpRootFSDir, &privileges); err != nil { return err } return nil } // List displays the list of plugins and associated metadata. func (pm *Manager) List() ([]types.Plugin, error) { plugins := pm.config.Store.GetAll() out := make([]types.Plugin, 0, len(plugins)) for _, p := range plugins { out = append(out, p.PluginObj) } return out, nil } // Push pushes a plugin to the store. func (pm *Manager) Push(ctx context.Context, name string, metaHeader http.Header, authConfig *types.AuthConfig, outStream io.Writer) error { p, err := pm.config.Store.GetV2Plugin(name) if err != nil { return err } ref, err := reference.ParseNamed(p.Name()) if err != nil { return errors.Wrapf(err, "plugin has invalid name %v for push", p.Name()) } var po progress.Output if outStream != nil { // Include a buffer so that slow client connections don't affect // transfer performance. progressChan := make(chan progress.Progress, 100) writesDone := make(chan struct{}) defer func() { close(progressChan) <-writesDone }() var cancelFunc context.CancelFunc ctx, cancelFunc = context.WithCancel(ctx) go func() { progressutils.WriteDistributionProgress(cancelFunc, outStream, progressChan) close(writesDone) }() po = progress.ChanOutput(progressChan) } else { po = progress.DiscardOutput() } // TODO: replace these with manager is := &pluginConfigStore{ pm: pm, plugin: p, } ls := &pluginLayerProvider{ pm: pm, plugin: p, } rs := &pluginReference{ name: ref, pluginID: p.Config, } uploadManager := xfer.NewLayerUploadManager(3) imagePushConfig := &distribution.ImagePushConfig{ Config: distribution.Config{ MetaHeaders: metaHeader, AuthConfig: authConfig, ProgressOutput: po, RegistryService: pm.config.RegistryService, ReferenceStore: rs, ImageEventLogger: pm.config.LogPluginEvent, ImageStore: is, RequireSchema2: true, }, ConfigMediaType: schema2.MediaTypePluginConfig, LayerStore: ls, UploadManager: uploadManager, } return distribution.Push(ctx, ref, imagePushConfig) } type pluginReference struct { name reference.Named pluginID digest.Digest } func (r *pluginReference) References(id digest.Digest) []reference.Named { if r.pluginID != id { return nil } return []reference.Named{r.name} } func (r *pluginReference) ReferencesByName(ref reference.Named) []reference.Association { return []reference.Association{ { Ref: r.name, ID: r.pluginID, }, } } func (r *pluginReference) Get(ref reference.Named) (digest.Digest, error) { if r.name.String() != ref.String() { return digest.Digest(""), reference.ErrDoesNotExist } return r.pluginID, nil } func (r *pluginReference) AddTag(ref reference.Named, id digest.Digest, force bool) error { // Read only, ignore return nil } func (r *pluginReference) AddDigest(ref reference.Canonical, id digest.Digest, force bool) error { // Read only, ignore return nil } func (r *pluginReference) Delete(ref reference.Named) (bool, error) { // Read only, ignore return false, nil } type pluginConfigStore struct { pm *Manager plugin *v2.Plugin } func (s *pluginConfigStore) Put([]byte) (digest.Digest, error) { return digest.Digest(""), errors.New("cannot store config on push") } func (s *pluginConfigStore) Get(d digest.Digest) ([]byte, error) { if s.plugin.Config != d { return nil, errors.New("plugin not found") } rwc, err := s.pm.blobStore.Get(d) if err != nil { return nil, err } defer rwc.Close() return ioutil.ReadAll(rwc) } func (s *pluginConfigStore) RootFSFromConfig(c []byte) (*image.RootFS, error) { return configToRootFS(c) } type pluginLayerProvider struct { pm *Manager plugin *v2.Plugin } func (p *pluginLayerProvider) Get(id layer.ChainID) (distribution.PushLayer, error) { rootFS := rootFSFromPlugin(p.plugin.PluginObj.Config.Rootfs) var i int for i = 1; i <= len(rootFS.DiffIDs); i++ { if layer.CreateChainID(rootFS.DiffIDs[:i]) == id { break } } if i > len(rootFS.DiffIDs) { return nil, errors.New("layer not found") } return &pluginLayer{ pm: p.pm, diffIDs: rootFS.DiffIDs[:i], blobs: p.plugin.Blobsums[:i], }, nil } type pluginLayer struct { pm *Manager diffIDs []layer.DiffID blobs []digest.Digest } func (l *pluginLayer) ChainID() layer.ChainID { return layer.CreateChainID(l.diffIDs) } func (l *pluginLayer) DiffID() layer.DiffID { return l.diffIDs[len(l.diffIDs)-1] } func (l *pluginLayer) Parent() distribution.PushLayer { if len(l.diffIDs) == 1 { return nil } return &pluginLayer{ pm: l.pm, diffIDs: l.diffIDs[:len(l.diffIDs)-1], blobs: l.blobs[:len(l.diffIDs)-1], } } func (l *pluginLayer) Open() (io.ReadCloser, error) { return l.pm.blobStore.Get(l.blobs[len(l.diffIDs)-1]) } func (l *pluginLayer) Size() (int64, error) { return l.pm.blobStore.Size(l.blobs[len(l.diffIDs)-1]) } func (l *pluginLayer) MediaType() string { return schema2.MediaTypeLayer } func (l *pluginLayer) Release() { // Nothing needs to be release, no references held } // Remove deletes plugin's root directory. func (pm *Manager) Remove(name string, config *types.PluginRmConfig) error { p, err := pm.config.Store.GetV2Plugin(name) pm.mu.RLock() c := pm.cMap[p] pm.mu.RUnlock() if err != nil { return err } if !config.ForceRemove { if p.GetRefCount() > 0 { return fmt.Errorf("plugin %s is in use", p.Name()) } if p.IsEnabled() { return fmt.Errorf("plugin %s is enabled", p.Name()) } } if p.IsEnabled() { if err := pm.disable(p, c); err != nil { logrus.Errorf("failed to disable plugin '%s': %s", p.Name(), err) } } defer func() { go pm.GC() }() id := p.GetID() pm.config.Store.Remove(p) pluginDir := filepath.Join(pm.config.Root, id) if err := os.RemoveAll(pluginDir); err != nil { logrus.Warnf("unable to remove %q from plugin remove: %v", pluginDir, err) } pm.config.LogPluginEvent(id, name, "remove") return nil } // Set sets plugin args func (pm *Manager) Set(name string, args []string) error { p, err := pm.config.Store.GetV2Plugin(name) if err != nil { return err } if err := p.Set(args); err != nil { return err } return pm.save(p) } // CreateFromContext creates a plugin from the given pluginDir which contains // both the rootfs and the config.json and a repoName with optional tag. func (pm *Manager) CreateFromContext(ctx context.Context, tarCtx io.ReadCloser, options *types.PluginCreateOptions) (err error) { pm.muGC.RLock() defer pm.muGC.RUnlock() ref, err := reference.ParseNamed(options.RepoName) if err != nil { return errors.Wrapf(err, "failed to parse reference %v", options.RepoName) } if _, ok := ref.(reference.Canonical); ok { return errors.Errorf("canonical references are not permitted") } name := reference.WithDefaultTag(ref).String() if err := pm.config.Store.validateName(name); err != nil { // fast check, real check is in createPlugin() return err } tmpRootFSDir, err := ioutil.TempDir(pm.tmpDir(), ".rootfs") defer os.RemoveAll(tmpRootFSDir) if err != nil { return errors.Wrap(err, "failed to create temp directory") } var configJSON []byte rootFS := splitConfigRootFSFromTar(tarCtx, &configJSON) rootFSBlob, err := pm.blobStore.New() if err != nil { return err } defer rootFSBlob.Close() gzw := gzip.NewWriter(rootFSBlob) layerDigester := digest.Canonical.Digester() rootFSReader := io.TeeReader(rootFS, io.MultiWriter(gzw, layerDigester.Hash())) if err := chrootarchive.Untar(rootFSReader, tmpRootFSDir, nil); err != nil { return err } if err := rootFS.Close(); err != nil { return err } if configJSON == nil { return errors.New("config not found") } if err := gzw.Close(); err != nil { return errors.Wrap(err, "error closing gzip writer") } var config types.PluginConfig if err := json.Unmarshal(configJSON, &config); err != nil { return errors.Wrap(err, "failed to parse config") } if err := pm.validateConfig(config); err != nil { return err } pm.mu.Lock() defer pm.mu.Unlock() rootFSBlobsum, err := rootFSBlob.Commit() if err != nil { return err } defer func() { if err != nil { go pm.GC() } }() config.Rootfs = &types.PluginConfigRootfs{ Type: "layers", DiffIds: []string{layerDigester.Digest().String()}, } configBlob, err := pm.blobStore.New() if err != nil { return err } defer configBlob.Close() if err := json.NewEncoder(configBlob).Encode(config); err != nil { return errors.Wrap(err, "error encoding json config") } configBlobsum, err := configBlob.Commit() if err != nil { return err } p, err := pm.createPlugin(name, configBlobsum, []digest.Digest{rootFSBlobsum}, tmpRootFSDir, nil) if err != nil { return err } pm.config.LogPluginEvent(p.PluginObj.ID, name, "create") return nil } func (pm *Manager) validateConfig(config types.PluginConfig) error { return nil // TODO: } func splitConfigRootFSFromTar(in io.ReadCloser, config *[]byte) io.ReadCloser { pr, pw := io.Pipe() go func() { tarReader := tar.NewReader(in) tarWriter := tar.NewWriter(pw) defer in.Close() hasRootFS := false for { hdr, err := tarReader.Next() if err == io.EOF { if !hasRootFS { pw.CloseWithError(errors.Wrap(err, "no rootfs found")) return } // Signals end of archive. tarWriter.Close() pw.Close() return } if err != nil { pw.CloseWithError(errors.Wrap(err, "failed to read from tar")) return } content := io.Reader(tarReader) name := path.Clean(hdr.Name) if path.IsAbs(name) { name = name[1:] } if name == configFileName { dt, err := ioutil.ReadAll(content) if err != nil { pw.CloseWithError(errors.Wrapf(err, "failed to read %s", configFileName)) return } *config = dt } if parts := strings.Split(name, "/"); len(parts) != 0 && parts[0] == rootFSFileName { hdr.Name = path.Clean(path.Join(parts[1:]...)) if hdr.Typeflag == tar.TypeLink && strings.HasPrefix(strings.ToLower(hdr.Linkname), rootFSFileName+"/") { hdr.Linkname = hdr.Linkname[len(rootFSFileName)+1:] } if err := tarWriter.WriteHeader(hdr); err != nil { pw.CloseWithError(errors.Wrap(err, "error writing tar header")) return } if _, err := pools.Copy(tarWriter, content); err != nil { pw.CloseWithError(errors.Wrap(err, "error copying tar data")) return } hasRootFS = true } else { io.Copy(ioutil.Discard, content) } } }() return pr }