mirror of
https://github.com/moby/moby.git
synced 2022-11-09 12:21:53 -05:00
5d5c89398c
This commit makes the Docker daemon call UpdateSuffixarray only after it finishes registering all containers. This lowers the amount of time required for the Docker daemon to start up. Docker-DCO-1.1-Signed-off-by: Cristian Staretu <cristian.staretu@gmail.com> (github: unclejack)
1043 lines
30 KiB
Go
1043 lines
30 KiB
Go
package daemon
|
|
|
|
import (
|
|
"container/list"
|
|
"fmt"
|
|
"io"
|
|
"io/ioutil"
|
|
"log"
|
|
"os"
|
|
"path"
|
|
"regexp"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/dotcloud/docker/archive"
|
|
"github.com/dotcloud/docker/daemon/execdriver"
|
|
"github.com/dotcloud/docker/daemon/execdriver/execdrivers"
|
|
"github.com/dotcloud/docker/daemon/execdriver/lxc"
|
|
"github.com/dotcloud/docker/daemon/graphdriver"
|
|
_ "github.com/dotcloud/docker/daemon/graphdriver/vfs"
|
|
_ "github.com/dotcloud/docker/daemon/networkdriver/bridge"
|
|
"github.com/dotcloud/docker/daemon/networkdriver/portallocator"
|
|
"github.com/dotcloud/docker/daemonconfig"
|
|
"github.com/dotcloud/docker/dockerversion"
|
|
"github.com/dotcloud/docker/engine"
|
|
"github.com/dotcloud/docker/graph"
|
|
"github.com/dotcloud/docker/image"
|
|
"github.com/dotcloud/docker/pkg/graphdb"
|
|
"github.com/dotcloud/docker/pkg/label"
|
|
"github.com/dotcloud/docker/pkg/mount"
|
|
"github.com/dotcloud/docker/pkg/networkfs/resolvconf"
|
|
"github.com/dotcloud/docker/pkg/selinux"
|
|
"github.com/dotcloud/docker/pkg/sysinfo"
|
|
"github.com/dotcloud/docker/runconfig"
|
|
"github.com/dotcloud/docker/utils"
|
|
)
|
|
|
|
// Set the max depth to the aufs default that most
|
|
// kernels are compiled with
|
|
// For more information see: http://sourceforge.net/p/aufs/aufs3-standalone/ci/aufs3.12/tree/config.mk
|
|
const MaxImageDepth = 127
|
|
|
|
var (
|
|
DefaultDns = []string{"8.8.8.8", "8.8.4.4"}
|
|
validContainerNameChars = `[a-zA-Z0-9_.-]`
|
|
validContainerNamePattern = regexp.MustCompile(`^/?` + validContainerNameChars + `+$`)
|
|
)
|
|
|
|
type Daemon struct {
|
|
repository string
|
|
sysInitPath string
|
|
containers *list.List
|
|
graph *graph.Graph
|
|
repositories *graph.TagStore
|
|
idIndex *utils.TruncIndex
|
|
sysInfo *sysinfo.SysInfo
|
|
volumes *graph.Graph
|
|
srv Server
|
|
eng *engine.Engine
|
|
config *daemonconfig.Config
|
|
containerGraph *graphdb.Database
|
|
driver graphdriver.Driver
|
|
execDriver execdriver.Driver
|
|
}
|
|
|
|
// Mountpoints should be private to the container
|
|
func remountPrivate(mountPoint string) error {
|
|
mounted, err := mount.Mounted(mountPoint)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if !mounted {
|
|
if err := mount.Mount(mountPoint, mountPoint, "none", "bind,rw"); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return mount.ForceMount("", mountPoint, "none", "private")
|
|
}
|
|
|
|
// List returns an array of all containers registered in the daemon.
|
|
func (daemon *Daemon) List() []*Container {
|
|
containers := new(History)
|
|
for e := daemon.containers.Front(); e != nil; e = e.Next() {
|
|
containers.Add(e.Value.(*Container))
|
|
}
|
|
return *containers
|
|
}
|
|
|
|
func (daemon *Daemon) getContainerElement(id string) *list.Element {
|
|
for e := daemon.containers.Front(); e != nil; e = e.Next() {
|
|
container := e.Value.(*Container)
|
|
if container.ID == id {
|
|
return e
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Get looks for a container by the specified ID or name, and returns it.
|
|
// If the container is not found, or if an error occurs, nil is returned.
|
|
func (daemon *Daemon) Get(name string) *Container {
|
|
if c, _ := daemon.GetByName(name); c != nil {
|
|
return c
|
|
}
|
|
|
|
id, err := daemon.idIndex.Get(name)
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
|
|
e := daemon.getContainerElement(id)
|
|
if e == nil {
|
|
return nil
|
|
}
|
|
return e.Value.(*Container)
|
|
}
|
|
|
|
// Exists returns a true if a container of the specified ID or name exists,
|
|
// false otherwise.
|
|
func (daemon *Daemon) Exists(id string) bool {
|
|
return daemon.Get(id) != nil
|
|
}
|
|
|
|
func (daemon *Daemon) containerRoot(id string) string {
|
|
return path.Join(daemon.repository, id)
|
|
}
|
|
|
|
// Load reads the contents of a container from disk
|
|
// This is typically done at startup.
|
|
func (daemon *Daemon) load(id string) (*Container, error) {
|
|
container := &Container{root: daemon.containerRoot(id)}
|
|
if err := container.FromDisk(); err != nil {
|
|
return nil, err
|
|
}
|
|
if container.ID != id {
|
|
return container, fmt.Errorf("Container %s is stored at %s", container.ID, id)
|
|
}
|
|
return container, nil
|
|
}
|
|
|
|
// Register makes a container object usable by the daemon as <container.ID>
|
|
// This is a wrapper for register
|
|
func (daemon *Daemon) Register(container *Container) error {
|
|
return daemon.register(container, true)
|
|
}
|
|
|
|
// register makes a container object usable by the daemon as <container.ID>
|
|
func (daemon *Daemon) register(container *Container, updateSuffixarray bool) error {
|
|
if container.daemon != nil || daemon.Exists(container.ID) {
|
|
return fmt.Errorf("Container is already loaded")
|
|
}
|
|
if err := validateID(container.ID); err != nil {
|
|
return err
|
|
}
|
|
if err := daemon.ensureName(container); err != nil {
|
|
return err
|
|
}
|
|
|
|
container.daemon = daemon
|
|
|
|
// Attach to stdout and stderr
|
|
container.stderr = utils.NewWriteBroadcaster()
|
|
container.stdout = utils.NewWriteBroadcaster()
|
|
// Attach to stdin
|
|
if container.Config.OpenStdin {
|
|
container.stdin, container.stdinPipe = io.Pipe()
|
|
} else {
|
|
container.stdinPipe = utils.NopWriteCloser(ioutil.Discard) // Silently drop stdin
|
|
}
|
|
// done
|
|
daemon.containers.PushBack(container)
|
|
|
|
// don't update the Suffixarray if we're starting up
|
|
// we'll waste time if we update it for every container
|
|
if updateSuffixarray {
|
|
daemon.idIndex.Add(container.ID)
|
|
} else {
|
|
daemon.idIndex.AddWithoutSuffixarrayUpdate(container.ID)
|
|
}
|
|
|
|
// FIXME: if the container is supposed to be running but is not, auto restart it?
|
|
// if so, then we need to restart monitor and init a new lock
|
|
// If the container is supposed to be running, make sure of it
|
|
if container.State.IsRunning() {
|
|
utils.Debugf("killing old running container %s", container.ID)
|
|
|
|
existingPid := container.State.Pid
|
|
container.State.SetStopped(0)
|
|
|
|
// We only have to handle this for lxc because the other drivers will ensure that
|
|
// no processes are left when docker dies
|
|
if container.ExecDriver == "" || strings.Contains(container.ExecDriver, "lxc") {
|
|
lxc.KillLxc(container.ID, 9)
|
|
} else {
|
|
// use the current driver and ensure that the container is dead x.x
|
|
cmd := &execdriver.Command{
|
|
ID: container.ID,
|
|
}
|
|
var err error
|
|
cmd.Process, err = os.FindProcess(existingPid)
|
|
if err != nil {
|
|
utils.Debugf("cannot find existing process for %d", existingPid)
|
|
}
|
|
daemon.execDriver.Terminate(cmd)
|
|
}
|
|
if err := container.Unmount(); err != nil {
|
|
utils.Debugf("unmount error %s", err)
|
|
}
|
|
if err := container.ToDisk(); err != nil {
|
|
utils.Debugf("saving stopped state to disk %s", err)
|
|
}
|
|
|
|
info := daemon.execDriver.Info(container.ID)
|
|
if !info.IsRunning() {
|
|
utils.Debugf("Container %s was supposed to be running but is not.", container.ID)
|
|
if daemon.config.AutoRestart {
|
|
utils.Debugf("Restarting")
|
|
if err := container.Unmount(); err != nil {
|
|
utils.Debugf("restart unmount error %s", err)
|
|
}
|
|
|
|
if err := container.Start(); err != nil {
|
|
return err
|
|
}
|
|
} else {
|
|
utils.Debugf("Marking as stopped")
|
|
container.State.SetStopped(-127)
|
|
if err := container.ToDisk(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
// When the container is not running, we still initialize the waitLock
|
|
// chan and close it. Receiving on nil chan blocks whereas receiving on a
|
|
// closed chan does not. In this case we do not want to block.
|
|
container.waitLock = make(chan struct{})
|
|
close(container.waitLock)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) ensureName(container *Container) error {
|
|
if container.Name == "" {
|
|
name, err := generateRandomName(daemon)
|
|
if err != nil {
|
|
name = utils.TruncateID(container.ID)
|
|
}
|
|
container.Name = name
|
|
|
|
if err := container.ToDisk(); err != nil {
|
|
utils.Debugf("Error saving container name %s", err)
|
|
}
|
|
if !daemon.containerGraph.Exists(name) {
|
|
if _, err := daemon.containerGraph.Set(name, container.ID); err != nil {
|
|
utils.Debugf("Setting default id - %s", err)
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) LogToDisk(src *utils.WriteBroadcaster, dst, stream string) error {
|
|
log, err := os.OpenFile(dst, os.O_RDWR|os.O_APPEND|os.O_CREATE, 0600)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
src.AddWriter(log, stream)
|
|
return nil
|
|
}
|
|
|
|
// Destroy unregisters a container from the daemon and cleanly removes its contents from the filesystem.
|
|
func (daemon *Daemon) Destroy(container *Container) error {
|
|
if container == nil {
|
|
return fmt.Errorf("The given container is <nil>")
|
|
}
|
|
|
|
element := daemon.getContainerElement(container.ID)
|
|
if element == nil {
|
|
return fmt.Errorf("Container %v not found - maybe it was already destroyed?", container.ID)
|
|
}
|
|
|
|
if err := container.Stop(3); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Deregister the container before removing its directory, to avoid race conditions
|
|
daemon.idIndex.Delete(container.ID)
|
|
daemon.containers.Remove(element)
|
|
|
|
if err := daemon.driver.Remove(container.ID); err != nil {
|
|
return fmt.Errorf("Driver %s failed to remove root filesystem %s: %s", daemon.driver, container.ID, err)
|
|
}
|
|
|
|
initID := fmt.Sprintf("%s-init", container.ID)
|
|
if err := daemon.driver.Remove(initID); err != nil {
|
|
return fmt.Errorf("Driver %s failed to remove init filesystem %s: %s", daemon.driver, initID, err)
|
|
}
|
|
|
|
if _, err := daemon.containerGraph.Purge(container.ID); err != nil {
|
|
utils.Debugf("Unable to remove container from link graph: %s", err)
|
|
}
|
|
|
|
if err := os.RemoveAll(container.root); err != nil {
|
|
return fmt.Errorf("Unable to remove filesystem for %v: %v", container.ID, err)
|
|
}
|
|
selinux.FreeLxcContexts(container.ProcessLabel)
|
|
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) restore() error {
|
|
if os.Getenv("DEBUG") == "" && os.Getenv("TEST") == "" {
|
|
fmt.Printf("Loading containers: ")
|
|
}
|
|
dir, err := ioutil.ReadDir(daemon.repository)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
containers := make(map[string]*Container)
|
|
currentDriver := daemon.driver.String()
|
|
|
|
for _, v := range dir {
|
|
id := v.Name()
|
|
container, err := daemon.load(id)
|
|
if os.Getenv("DEBUG") == "" && os.Getenv("TEST") == "" {
|
|
fmt.Print(".")
|
|
}
|
|
if err != nil {
|
|
utils.Errorf("Failed to load container %v: %v", id, err)
|
|
continue
|
|
}
|
|
|
|
// Ignore the container if it does not support the current driver being used by the graph
|
|
if container.Driver == "" && currentDriver == "aufs" || container.Driver == currentDriver {
|
|
utils.Debugf("Loaded container %v", container.ID)
|
|
containers[container.ID] = container
|
|
} else {
|
|
utils.Debugf("Cannot load container %s because it was created with another graph driver.", container.ID)
|
|
}
|
|
}
|
|
|
|
registerContainer := func(container *Container) {
|
|
if err := daemon.register(container, false); err != nil {
|
|
utils.Debugf("Failed to register container %s: %s", container.ID, err)
|
|
}
|
|
}
|
|
|
|
if entities := daemon.containerGraph.List("/", -1); entities != nil {
|
|
for _, p := range entities.Paths() {
|
|
if os.Getenv("DEBUG") == "" && os.Getenv("TEST") == "" {
|
|
fmt.Print(".")
|
|
}
|
|
e := entities[p]
|
|
if container, ok := containers[e.ID()]; ok {
|
|
registerContainer(container)
|
|
delete(containers, e.ID())
|
|
}
|
|
}
|
|
}
|
|
|
|
// Any containers that are left over do not exist in the graph
|
|
for _, container := range containers {
|
|
// Try to set the default name for a container if it exists prior to links
|
|
container.Name, err = generateRandomName(daemon)
|
|
if err != nil {
|
|
container.Name = utils.TruncateID(container.ID)
|
|
}
|
|
|
|
if _, err := daemon.containerGraph.Set(container.Name, container.ID); err != nil {
|
|
utils.Debugf("Setting default id - %s", err)
|
|
}
|
|
registerContainer(container)
|
|
}
|
|
|
|
daemon.idIndex.UpdateSuffixarray()
|
|
if os.Getenv("DEBUG") == "" && os.Getenv("TEST") == "" {
|
|
fmt.Printf(": done.\n")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Create creates a new container from the given configuration with a given name.
|
|
func (daemon *Daemon) Create(config *runconfig.Config, name string) (*Container, []string, error) {
|
|
var (
|
|
container *Container
|
|
warnings []string
|
|
)
|
|
|
|
img, err := daemon.repositories.LookupImage(config.Image)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if err := daemon.checkImageDepth(img); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if warnings, err = daemon.mergeAndVerifyConfig(config, img); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if container, err = daemon.newContainer(name, config, img); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if err := daemon.createRootfs(container, img); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if err := container.ToDisk(); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if err := daemon.Register(container); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
return container, warnings, nil
|
|
}
|
|
|
|
func (daemon *Daemon) checkImageDepth(img *image.Image) error {
|
|
// We add 2 layers to the depth because the container's rw and
|
|
// init layer add to the restriction
|
|
depth, err := img.Depth()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if depth+2 >= MaxImageDepth {
|
|
return fmt.Errorf("Cannot create container with more than %d parents", MaxImageDepth)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) checkDeprecatedExpose(config *runconfig.Config) bool {
|
|
if config != nil {
|
|
if config.PortSpecs != nil {
|
|
for _, p := range config.PortSpecs {
|
|
if strings.Contains(p, ":") {
|
|
return true
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (daemon *Daemon) mergeAndVerifyConfig(config *runconfig.Config, img *image.Image) ([]string, error) {
|
|
warnings := []string{}
|
|
if daemon.checkDeprecatedExpose(img.Config) || daemon.checkDeprecatedExpose(config) {
|
|
warnings = append(warnings, "The mapping to public ports on your host via Dockerfile EXPOSE (host:port:port) has been deprecated. Use -p to publish the ports.")
|
|
}
|
|
if img.Config != nil {
|
|
if err := runconfig.Merge(config, img.Config); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
if len(config.Entrypoint) == 0 && len(config.Cmd) == 0 {
|
|
return nil, fmt.Errorf("No command specified")
|
|
}
|
|
return warnings, nil
|
|
}
|
|
|
|
func (daemon *Daemon) generateIdAndName(name string) (string, string, error) {
|
|
var (
|
|
err error
|
|
id = utils.GenerateRandomID()
|
|
)
|
|
|
|
if name == "" {
|
|
name, err = generateRandomName(daemon)
|
|
if err != nil {
|
|
name = utils.TruncateID(id)
|
|
}
|
|
} else {
|
|
if !validContainerNamePattern.MatchString(name) {
|
|
return "", "", fmt.Errorf("Invalid container name (%s), only %s are allowed", name, validContainerNameChars)
|
|
}
|
|
}
|
|
if name[0] != '/' {
|
|
name = "/" + name
|
|
}
|
|
// Set the enitity in the graph using the default name specified
|
|
if _, err := daemon.containerGraph.Set(name, id); err != nil {
|
|
if !graphdb.IsNonUniqueNameError(err) {
|
|
return "", "", err
|
|
}
|
|
|
|
conflictingContainer, err := daemon.GetByName(name)
|
|
if err != nil {
|
|
if strings.Contains(err.Error(), "Could not find entity") {
|
|
return "", "", err
|
|
}
|
|
|
|
// Remove name and continue starting the container
|
|
if err := daemon.containerGraph.Delete(name); err != nil {
|
|
return "", "", err
|
|
}
|
|
} else {
|
|
nameAsKnownByUser := strings.TrimPrefix(name, "/")
|
|
return "", "", fmt.Errorf(
|
|
"Conflict, The name %s is already assigned to %s. You have to delete (or rename) that container to be able to assign %s to a container again.", nameAsKnownByUser,
|
|
utils.TruncateID(conflictingContainer.ID), nameAsKnownByUser)
|
|
}
|
|
}
|
|
return id, name, nil
|
|
}
|
|
|
|
func (daemon *Daemon) generateHostname(id string, config *runconfig.Config) {
|
|
// Generate default hostname
|
|
// FIXME: the lxc template no longer needs to set a default hostname
|
|
if config.Hostname == "" {
|
|
config.Hostname = id[:12]
|
|
}
|
|
}
|
|
|
|
func (daemon *Daemon) getEntrypointAndArgs(config *runconfig.Config) (string, []string) {
|
|
var (
|
|
entrypoint string
|
|
args []string
|
|
)
|
|
if len(config.Entrypoint) != 0 {
|
|
entrypoint = config.Entrypoint[0]
|
|
args = append(config.Entrypoint[1:], config.Cmd...)
|
|
} else {
|
|
entrypoint = config.Cmd[0]
|
|
args = config.Cmd[1:]
|
|
}
|
|
return entrypoint, args
|
|
}
|
|
|
|
func (daemon *Daemon) newContainer(name string, config *runconfig.Config, img *image.Image) (*Container, error) {
|
|
var (
|
|
id string
|
|
err error
|
|
)
|
|
id, name, err = daemon.generateIdAndName(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
daemon.generateHostname(id, config)
|
|
entrypoint, args := daemon.getEntrypointAndArgs(config)
|
|
|
|
container := &Container{
|
|
// FIXME: we should generate the ID here instead of receiving it as an argument
|
|
ID: id,
|
|
Created: time.Now().UTC(),
|
|
Path: entrypoint,
|
|
Args: args, //FIXME: de-duplicate from config
|
|
Config: config,
|
|
hostConfig: &runconfig.HostConfig{},
|
|
Image: img.ID, // Always use the resolved image id
|
|
NetworkSettings: &NetworkSettings{},
|
|
Name: name,
|
|
Driver: daemon.driver.String(),
|
|
ExecDriver: daemon.execDriver.Name(),
|
|
}
|
|
container.root = daemon.containerRoot(container.ID)
|
|
|
|
if container.ProcessLabel, container.MountLabel, err = label.GenLabels(""); err != nil {
|
|
return nil, err
|
|
}
|
|
return container, nil
|
|
}
|
|
|
|
func (daemon *Daemon) createRootfs(container *Container, img *image.Image) error {
|
|
// Step 1: create the container directory.
|
|
// This doubles as a barrier to avoid race conditions.
|
|
if err := os.Mkdir(container.root, 0700); err != nil {
|
|
return err
|
|
}
|
|
initID := fmt.Sprintf("%s-init", container.ID)
|
|
if err := daemon.driver.Create(initID, img.ID); err != nil {
|
|
return err
|
|
}
|
|
initPath, err := daemon.driver.Get(initID, "")
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer daemon.driver.Put(initID)
|
|
|
|
if err := graph.SetupInitLayer(initPath); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := daemon.driver.Create(container.ID, initID); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Commit creates a new filesystem image from the current state of a container.
|
|
// The image can optionally be tagged into a repository
|
|
func (daemon *Daemon) Commit(container *Container, repository, tag, comment, author string, config *runconfig.Config) (*image.Image, error) {
|
|
// FIXME: freeze the container before copying it to avoid data corruption?
|
|
if err := container.Mount(); err != nil {
|
|
return nil, err
|
|
}
|
|
defer container.Unmount()
|
|
|
|
rwTar, err := container.ExportRw()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer rwTar.Close()
|
|
|
|
// Create a new image from the container's base layers + a new layer from container changes
|
|
var (
|
|
containerID, containerImage string
|
|
containerConfig *runconfig.Config
|
|
)
|
|
if container != nil {
|
|
containerID = container.ID
|
|
containerImage = container.Image
|
|
containerConfig = container.Config
|
|
}
|
|
img, err := daemon.graph.Create(rwTar, containerID, containerImage, comment, author, containerConfig, config)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// Register the image if needed
|
|
if repository != "" {
|
|
if err := daemon.repositories.Set(repository, tag, img.ID, true); err != nil {
|
|
return img, err
|
|
}
|
|
}
|
|
return img, nil
|
|
}
|
|
|
|
func GetFullContainerName(name string) (string, error) {
|
|
if name == "" {
|
|
return "", fmt.Errorf("Container name cannot be empty")
|
|
}
|
|
if name[0] != '/' {
|
|
name = "/" + name
|
|
}
|
|
return name, nil
|
|
}
|
|
|
|
func (daemon *Daemon) GetByName(name string) (*Container, error) {
|
|
fullName, err := GetFullContainerName(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
entity := daemon.containerGraph.Get(fullName)
|
|
if entity == nil {
|
|
return nil, fmt.Errorf("Could not find entity for %s", name)
|
|
}
|
|
e := daemon.getContainerElement(entity.ID())
|
|
if e == nil {
|
|
return nil, fmt.Errorf("Could not find container for entity id %s", entity.ID())
|
|
}
|
|
return e.Value.(*Container), nil
|
|
}
|
|
|
|
func (daemon *Daemon) Children(name string) (map[string]*Container, error) {
|
|
name, err := GetFullContainerName(name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
children := make(map[string]*Container)
|
|
|
|
err = daemon.containerGraph.Walk(name, func(p string, e *graphdb.Entity) error {
|
|
c := daemon.Get(e.ID())
|
|
if c == nil {
|
|
return fmt.Errorf("Could not get container for name %s and id %s", e.ID(), p)
|
|
}
|
|
children[p] = c
|
|
return nil
|
|
}, 0)
|
|
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return children, nil
|
|
}
|
|
|
|
func (daemon *Daemon) RegisterLink(parent, child *Container, alias string) error {
|
|
fullName := path.Join(parent.Name, alias)
|
|
if !daemon.containerGraph.Exists(fullName) {
|
|
_, err := daemon.containerGraph.Set(fullName, child.ID)
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) RegisterLinks(container *Container, hostConfig *runconfig.HostConfig) error {
|
|
if hostConfig != nil && hostConfig.Links != nil {
|
|
for _, l := range hostConfig.Links {
|
|
parts, err := utils.PartParser("name:alias", l)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
child, err := daemon.GetByName(parts["name"])
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if child == nil {
|
|
return fmt.Errorf("Could not get container for %s", parts["name"])
|
|
}
|
|
if err := daemon.RegisterLink(container, child, parts["alias"]); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// After we load all the links into the daemon
|
|
// set them to nil on the hostconfig
|
|
hostConfig.Links = nil
|
|
if err := container.WriteHostConfig(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// FIXME: harmonize with NewGraph()
|
|
func NewDaemon(config *daemonconfig.Config, eng *engine.Engine) (*Daemon, error) {
|
|
daemon, err := NewDaemonFromDirectory(config, eng)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return daemon, nil
|
|
}
|
|
|
|
func NewDaemonFromDirectory(config *daemonconfig.Config, eng *engine.Engine) (*Daemon, error) {
|
|
if !config.EnableSelinuxSupport {
|
|
selinux.SetDisabled()
|
|
}
|
|
|
|
// Create the root directory if it doesn't exists
|
|
if err := os.MkdirAll(config.Root, 0700); err != nil && !os.IsExist(err) {
|
|
return nil, err
|
|
}
|
|
|
|
// Set the default driver
|
|
graphdriver.DefaultDriver = config.GraphDriver
|
|
|
|
// Load storage driver
|
|
driver, err := graphdriver.New(config.Root)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
utils.Debugf("Using graph driver %s", driver)
|
|
|
|
if err := remountPrivate(config.Root); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
daemonRepo := path.Join(config.Root, "containers")
|
|
|
|
if err := os.MkdirAll(daemonRepo, 0700); err != nil && !os.IsExist(err) {
|
|
return nil, err
|
|
}
|
|
|
|
// Migrate the container if it is aufs and aufs is enabled
|
|
if err = migrateIfAufs(driver, config.Root); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
utils.Debugf("Creating images graph")
|
|
g, err := graph.NewGraph(path.Join(config.Root, "graph"), driver)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// We don't want to use a complex driver like aufs or devmapper
|
|
// for volumes, just a plain filesystem
|
|
volumesDriver, err := graphdriver.GetDriver("vfs", config.Root)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
utils.Debugf("Creating volumes graph")
|
|
volumes, err := graph.NewGraph(path.Join(config.Root, "volumes"), volumesDriver)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
utils.Debugf("Creating repository list")
|
|
repositories, err := graph.NewTagStore(path.Join(config.Root, "repositories-"+driver.String()), g)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("Couldn't create Tag store: %s", err)
|
|
}
|
|
|
|
if !config.DisableNetwork {
|
|
job := eng.Job("init_networkdriver")
|
|
|
|
job.SetenvBool("EnableIptables", config.EnableIptables)
|
|
job.SetenvBool("InterContainerCommunication", config.InterContainerCommunication)
|
|
job.SetenvBool("EnableIpForward", config.EnableIpForward)
|
|
job.Setenv("BridgeIface", config.BridgeIface)
|
|
job.Setenv("BridgeIP", config.BridgeIP)
|
|
job.Setenv("DefaultBindingIP", config.DefaultIp.String())
|
|
|
|
if err := job.Run(); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
graphdbPath := path.Join(config.Root, "linkgraph.db")
|
|
graph, err := graphdb.NewSqliteConn(graphdbPath)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
localCopy := path.Join(config.Root, "init", fmt.Sprintf("dockerinit-%s", dockerversion.VERSION))
|
|
sysInitPath := utils.DockerInitPath(localCopy)
|
|
if sysInitPath == "" {
|
|
return nil, fmt.Errorf("Could not locate dockerinit: This usually means docker was built incorrectly. See http://docs.docker.io/en/latest/contributing/devenvironment for official build instructions.")
|
|
}
|
|
|
|
if sysInitPath != localCopy {
|
|
// When we find a suitable dockerinit binary (even if it's our local binary), we copy it into config.Root at localCopy for future use (so that the original can go away without that being a problem, for example during a package upgrade).
|
|
if err := os.Mkdir(path.Dir(localCopy), 0700); err != nil && !os.IsExist(err) {
|
|
return nil, err
|
|
}
|
|
if _, err := utils.CopyFile(sysInitPath, localCopy); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := os.Chmod(localCopy, 0700); err != nil {
|
|
return nil, err
|
|
}
|
|
sysInitPath = localCopy
|
|
}
|
|
|
|
sysInfo := sysinfo.New(false)
|
|
ed, err := execdrivers.NewDriver(config.ExecDriver, config.Root, sysInitPath, sysInfo)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
daemon := &Daemon{
|
|
repository: daemonRepo,
|
|
containers: list.New(),
|
|
graph: g,
|
|
repositories: repositories,
|
|
idIndex: utils.NewTruncIndex([]string{}),
|
|
sysInfo: sysInfo,
|
|
volumes: volumes,
|
|
config: config,
|
|
containerGraph: graph,
|
|
driver: driver,
|
|
sysInitPath: sysInitPath,
|
|
execDriver: ed,
|
|
eng: eng,
|
|
}
|
|
|
|
if err := daemon.checkLocaldns(); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := daemon.restore(); err != nil {
|
|
return nil, err
|
|
}
|
|
return daemon, nil
|
|
}
|
|
|
|
func (daemon *Daemon) shutdown() error {
|
|
group := sync.WaitGroup{}
|
|
utils.Debugf("starting clean shutdown of all containers...")
|
|
for _, container := range daemon.List() {
|
|
c := container
|
|
if c.State.IsRunning() {
|
|
utils.Debugf("stopping %s", c.ID)
|
|
group.Add(1)
|
|
|
|
go func() {
|
|
defer group.Done()
|
|
if err := c.KillSig(15); err != nil {
|
|
utils.Debugf("kill 15 error for %s - %s", c.ID, err)
|
|
}
|
|
c.Wait()
|
|
utils.Debugf("container stopped %s", c.ID)
|
|
}()
|
|
}
|
|
}
|
|
group.Wait()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) Close() error {
|
|
errorsStrings := []string{}
|
|
if err := daemon.shutdown(); err != nil {
|
|
utils.Errorf("daemon.shutdown(): %s", err)
|
|
errorsStrings = append(errorsStrings, err.Error())
|
|
}
|
|
if err := portallocator.ReleaseAll(); err != nil {
|
|
utils.Errorf("portallocator.ReleaseAll(): %s", err)
|
|
errorsStrings = append(errorsStrings, err.Error())
|
|
}
|
|
if err := daemon.driver.Cleanup(); err != nil {
|
|
utils.Errorf("daemon.driver.Cleanup(): %s", err.Error())
|
|
errorsStrings = append(errorsStrings, err.Error())
|
|
}
|
|
if err := daemon.containerGraph.Close(); err != nil {
|
|
utils.Errorf("daemon.containerGraph.Close(): %s", err.Error())
|
|
errorsStrings = append(errorsStrings, err.Error())
|
|
}
|
|
if len(errorsStrings) > 0 {
|
|
return fmt.Errorf("%s", strings.Join(errorsStrings, ", "))
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) Mount(container *Container) error {
|
|
dir, err := daemon.driver.Get(container.ID, container.GetMountLabel())
|
|
if err != nil {
|
|
return fmt.Errorf("Error getting container %s from driver %s: %s", container.ID, daemon.driver, err)
|
|
}
|
|
if container.basefs == "" {
|
|
container.basefs = dir
|
|
} else if container.basefs != dir {
|
|
return fmt.Errorf("Error: driver %s is returning inconsistent paths for container %s ('%s' then '%s')",
|
|
daemon.driver, container.ID, container.basefs, dir)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) Unmount(container *Container) error {
|
|
daemon.driver.Put(container.ID)
|
|
return nil
|
|
}
|
|
|
|
func (daemon *Daemon) Changes(container *Container) ([]archive.Change, error) {
|
|
if differ, ok := daemon.driver.(graphdriver.Differ); ok {
|
|
return differ.Changes(container.ID)
|
|
}
|
|
cDir, err := daemon.driver.Get(container.ID, "")
|
|
if err != nil {
|
|
return nil, fmt.Errorf("Error getting container rootfs %s from driver %s: %s", container.ID, container.daemon.driver, err)
|
|
}
|
|
defer daemon.driver.Put(container.ID)
|
|
initDir, err := daemon.driver.Get(container.ID+"-init", "")
|
|
if err != nil {
|
|
return nil, fmt.Errorf("Error getting container init rootfs %s from driver %s: %s", container.ID, container.daemon.driver, err)
|
|
}
|
|
defer daemon.driver.Put(container.ID + "-init")
|
|
return archive.ChangesDirs(cDir, initDir)
|
|
}
|
|
|
|
func (daemon *Daemon) Diff(container *Container) (archive.Archive, error) {
|
|
if differ, ok := daemon.driver.(graphdriver.Differ); ok {
|
|
return differ.Diff(container.ID)
|
|
}
|
|
|
|
changes, err := daemon.Changes(container)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
cDir, err := daemon.driver.Get(container.ID, "")
|
|
if err != nil {
|
|
return nil, fmt.Errorf("Error getting container rootfs %s from driver %s: %s", container.ID, container.daemon.driver, err)
|
|
}
|
|
|
|
archive, err := archive.ExportChanges(cDir, changes)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return utils.NewReadCloserWrapper(archive, func() error {
|
|
err := archive.Close()
|
|
daemon.driver.Put(container.ID)
|
|
return err
|
|
}), nil
|
|
}
|
|
|
|
func (daemon *Daemon) Run(c *Container, pipes *execdriver.Pipes, startCallback execdriver.StartCallback) (int, error) {
|
|
return daemon.execDriver.Run(c.command, pipes, startCallback)
|
|
}
|
|
|
|
func (daemon *Daemon) Kill(c *Container, sig int) error {
|
|
return daemon.execDriver.Kill(c.command, sig)
|
|
}
|
|
|
|
// Nuke kills all containers then removes all content
|
|
// from the content root, including images, volumes and
|
|
// container filesystems.
|
|
// Again: this will remove your entire docker daemon!
|
|
func (daemon *Daemon) Nuke() error {
|
|
var wg sync.WaitGroup
|
|
for _, container := range daemon.List() {
|
|
wg.Add(1)
|
|
go func(c *Container) {
|
|
c.Kill()
|
|
wg.Done()
|
|
}(container)
|
|
}
|
|
wg.Wait()
|
|
daemon.Close()
|
|
|
|
return os.RemoveAll(daemon.config.Root)
|
|
}
|
|
|
|
// FIXME: this is a convenience function for integration tests
|
|
// which need direct access to daemon.graph.
|
|
// Once the tests switch to using engine and jobs, this method
|
|
// can go away.
|
|
func (daemon *Daemon) Graph() *graph.Graph {
|
|
return daemon.graph
|
|
}
|
|
|
|
func (daemon *Daemon) Repositories() *graph.TagStore {
|
|
return daemon.repositories
|
|
}
|
|
|
|
func (daemon *Daemon) Config() *daemonconfig.Config {
|
|
return daemon.config
|
|
}
|
|
|
|
func (daemon *Daemon) SystemConfig() *sysinfo.SysInfo {
|
|
return daemon.sysInfo
|
|
}
|
|
|
|
func (daemon *Daemon) SystemInitPath() string {
|
|
return daemon.sysInitPath
|
|
}
|
|
|
|
func (daemon *Daemon) GraphDriver() graphdriver.Driver {
|
|
return daemon.driver
|
|
}
|
|
|
|
func (daemon *Daemon) ExecutionDriver() execdriver.Driver {
|
|
return daemon.execDriver
|
|
}
|
|
|
|
func (daemon *Daemon) Volumes() *graph.Graph {
|
|
return daemon.volumes
|
|
}
|
|
|
|
func (daemon *Daemon) ContainerGraph() *graphdb.Database {
|
|
return daemon.containerGraph
|
|
}
|
|
|
|
func (daemon *Daemon) SetServer(server Server) {
|
|
daemon.srv = server
|
|
}
|
|
|
|
func (daemon *Daemon) checkLocaldns() error {
|
|
resolvConf, err := resolvconf.Get()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if len(daemon.config.Dns) == 0 && utils.CheckLocalDns(resolvConf) {
|
|
log.Printf("Local (127.0.0.1) DNS resolver found in resolv.conf and containers can't use it. Using default external servers : %v\n", DefaultDns)
|
|
daemon.config.Dns = DefaultDns
|
|
}
|
|
return nil
|
|
}
|