mirror of
				https://github.com/moby/moby.git
				synced 2022-11-09 12:21:53 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			669 lines
		
	
	
	
		
			18 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			669 lines
		
	
	
	
		
			18 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
package dockerfile
 | 
						|
 | 
						|
// internals for handling commands. Covers many areas and a lot of
 | 
						|
// non-contiguous functionality. Please read the comments.
 | 
						|
 | 
						|
import (
 | 
						|
	"crypto/sha256"
 | 
						|
	"encoding/hex"
 | 
						|
	"errors"
 | 
						|
	"fmt"
 | 
						|
	"io"
 | 
						|
	"io/ioutil"
 | 
						|
	"net/http"
 | 
						|
	"net/url"
 | 
						|
	"os"
 | 
						|
	"path/filepath"
 | 
						|
	"sort"
 | 
						|
	"strings"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/Sirupsen/logrus"
 | 
						|
	"github.com/docker/docker/api/types"
 | 
						|
	"github.com/docker/docker/api/types/backend"
 | 
						|
	"github.com/docker/docker/api/types/container"
 | 
						|
	"github.com/docker/docker/api/types/strslice"
 | 
						|
	"github.com/docker/docker/builder"
 | 
						|
	"github.com/docker/docker/builder/dockerfile/parser"
 | 
						|
	"github.com/docker/docker/pkg/archive"
 | 
						|
	"github.com/docker/docker/pkg/httputils"
 | 
						|
	"github.com/docker/docker/pkg/ioutils"
 | 
						|
	"github.com/docker/docker/pkg/jsonmessage"
 | 
						|
	"github.com/docker/docker/pkg/progress"
 | 
						|
	"github.com/docker/docker/pkg/streamformatter"
 | 
						|
	"github.com/docker/docker/pkg/stringid"
 | 
						|
	"github.com/docker/docker/pkg/system"
 | 
						|
	"github.com/docker/docker/pkg/tarsum"
 | 
						|
	"github.com/docker/docker/pkg/urlutil"
 | 
						|
	"github.com/docker/docker/runconfig/opts"
 | 
						|
)
 | 
						|
 | 
						|
func (b *Builder) commit(id string, autoCmd strslice.StrSlice, comment string) error {
 | 
						|
	if b.disableCommit {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
	if b.image == "" && !b.noBaseImage {
 | 
						|
		return fmt.Errorf("Please provide a source image with `from` prior to commit")
 | 
						|
	}
 | 
						|
	b.runConfig.Image = b.image
 | 
						|
 | 
						|
	if id == "" {
 | 
						|
		cmd := b.runConfig.Cmd
 | 
						|
		b.runConfig.Cmd = strslice.StrSlice(append(getShell(b.runConfig), "#(nop) ", comment))
 | 
						|
		defer func(cmd strslice.StrSlice) { b.runConfig.Cmd = cmd }(cmd)
 | 
						|
 | 
						|
		hit, err := b.probeCache()
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		} else if hit {
 | 
						|
			return nil
 | 
						|
		}
 | 
						|
		id, err = b.create()
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Note: Actually copy the struct
 | 
						|
	autoConfig := *b.runConfig
 | 
						|
	autoConfig.Cmd = autoCmd
 | 
						|
 | 
						|
	commitCfg := &backend.ContainerCommitConfig{
 | 
						|
		ContainerCommitConfig: types.ContainerCommitConfig{
 | 
						|
			Author: b.maintainer,
 | 
						|
			Pause:  true,
 | 
						|
			Config: &autoConfig,
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	// Commit the container
 | 
						|
	imageID, err := b.docker.Commit(id, commitCfg)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	b.image = imageID
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
type copyInfo struct {
 | 
						|
	builder.FileInfo
 | 
						|
	decompress bool
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) runContextCommand(args []string, allowRemote bool, allowLocalDecompression bool, cmdName string) error {
 | 
						|
	if b.context == nil {
 | 
						|
		return fmt.Errorf("No context given. Impossible to use %s", cmdName)
 | 
						|
	}
 | 
						|
 | 
						|
	if len(args) < 2 {
 | 
						|
		return fmt.Errorf("Invalid %s format - at least two arguments required", cmdName)
 | 
						|
	}
 | 
						|
 | 
						|
	// Work in daemon-specific filepath semantics
 | 
						|
	dest := filepath.FromSlash(args[len(args)-1]) // last one is always the dest
 | 
						|
 | 
						|
	b.runConfig.Image = b.image
 | 
						|
 | 
						|
	var infos []copyInfo
 | 
						|
 | 
						|
	// Loop through each src file and calculate the info we need to
 | 
						|
	// do the copy (e.g. hash value if cached).  Don't actually do
 | 
						|
	// the copy until we've looked at all src files
 | 
						|
	var err error
 | 
						|
	for _, orig := range args[0 : len(args)-1] {
 | 
						|
		var fi builder.FileInfo
 | 
						|
		decompress := allowLocalDecompression
 | 
						|
		if urlutil.IsURL(orig) {
 | 
						|
			if !allowRemote {
 | 
						|
				return fmt.Errorf("Source can't be a URL for %s", cmdName)
 | 
						|
			}
 | 
						|
			fi, err = b.download(orig)
 | 
						|
			if err != nil {
 | 
						|
				return err
 | 
						|
			}
 | 
						|
			defer os.RemoveAll(filepath.Dir(fi.Path()))
 | 
						|
			decompress = false
 | 
						|
			infos = append(infos, copyInfo{fi, decompress})
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		// not a URL
 | 
						|
		subInfos, err := b.calcCopyInfo(cmdName, orig, allowLocalDecompression, true)
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
 | 
						|
		infos = append(infos, subInfos...)
 | 
						|
	}
 | 
						|
 | 
						|
	if len(infos) == 0 {
 | 
						|
		return fmt.Errorf("No source files were specified")
 | 
						|
	}
 | 
						|
	if len(infos) > 1 && !strings.HasSuffix(dest, string(os.PathSeparator)) {
 | 
						|
		return fmt.Errorf("When using %s with more than one source file, the destination must be a directory and end with a /", cmdName)
 | 
						|
	}
 | 
						|
 | 
						|
	// For backwards compat, if there's just one info then use it as the
 | 
						|
	// cache look-up string, otherwise hash 'em all into one
 | 
						|
	var srcHash string
 | 
						|
	var origPaths string
 | 
						|
 | 
						|
	if len(infos) == 1 {
 | 
						|
		fi := infos[0].FileInfo
 | 
						|
		origPaths = fi.Name()
 | 
						|
		if hfi, ok := fi.(builder.Hashed); ok {
 | 
						|
			srcHash = hfi.Hash()
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		var hashs []string
 | 
						|
		var origs []string
 | 
						|
		for _, info := range infos {
 | 
						|
			fi := info.FileInfo
 | 
						|
			origs = append(origs, fi.Name())
 | 
						|
			if hfi, ok := fi.(builder.Hashed); ok {
 | 
						|
				hashs = append(hashs, hfi.Hash())
 | 
						|
			}
 | 
						|
		}
 | 
						|
		hasher := sha256.New()
 | 
						|
		hasher.Write([]byte(strings.Join(hashs, ",")))
 | 
						|
		srcHash = "multi:" + hex.EncodeToString(hasher.Sum(nil))
 | 
						|
		origPaths = strings.Join(origs, " ")
 | 
						|
	}
 | 
						|
 | 
						|
	cmd := b.runConfig.Cmd
 | 
						|
	b.runConfig.Cmd = strslice.StrSlice(append(getShell(b.runConfig), fmt.Sprintf("#(nop) %s %s in %s ", cmdName, srcHash, dest)))
 | 
						|
	defer func(cmd strslice.StrSlice) { b.runConfig.Cmd = cmd }(cmd)
 | 
						|
 | 
						|
	if hit, err := b.probeCache(); err != nil {
 | 
						|
		return err
 | 
						|
	} else if hit {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	container, err := b.docker.ContainerCreate(types.ContainerCreateConfig{Config: b.runConfig}, true)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	b.tmpContainers[container.ID] = struct{}{}
 | 
						|
 | 
						|
	comment := fmt.Sprintf("%s %s in %s", cmdName, origPaths, dest)
 | 
						|
 | 
						|
	// Twiddle the destination when its a relative path - meaning, make it
 | 
						|
	// relative to the WORKINGDIR
 | 
						|
	if dest, err = normaliseDest(cmdName, b.runConfig.WorkingDir, dest); err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	for _, info := range infos {
 | 
						|
		if err := b.docker.CopyOnBuild(container.ID, dest, info.FileInfo, info.decompress); err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return b.commit(container.ID, cmd, comment)
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) download(srcURL string) (fi builder.FileInfo, err error) {
 | 
						|
	// get filename from URL
 | 
						|
	u, err := url.Parse(srcURL)
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
	path := filepath.FromSlash(u.Path) // Ensure in platform semantics
 | 
						|
	if strings.HasSuffix(path, string(os.PathSeparator)) {
 | 
						|
		path = path[:len(path)-1]
 | 
						|
	}
 | 
						|
	parts := strings.Split(path, string(os.PathSeparator))
 | 
						|
	filename := parts[len(parts)-1]
 | 
						|
	if filename == "" {
 | 
						|
		err = fmt.Errorf("cannot determine filename from url: %s", u)
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Initiate the download
 | 
						|
	resp, err := httputils.Download(srcURL)
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Prepare file in a tmp dir
 | 
						|
	tmpDir, err := ioutils.TempDir("", "docker-remote")
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
	defer func() {
 | 
						|
		if err != nil {
 | 
						|
			os.RemoveAll(tmpDir)
 | 
						|
		}
 | 
						|
	}()
 | 
						|
	tmpFileName := filepath.Join(tmpDir, filename)
 | 
						|
	tmpFile, err := os.OpenFile(tmpFileName, os.O_RDWR|os.O_CREATE|os.O_EXCL, 0600)
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	stdoutFormatter := b.Stdout.(*streamformatter.StdoutFormatter)
 | 
						|
	progressOutput := stdoutFormatter.StreamFormatter.NewProgressOutput(stdoutFormatter.Writer, true)
 | 
						|
	progressReader := progress.NewProgressReader(resp.Body, progressOutput, resp.ContentLength, "", "Downloading")
 | 
						|
	// Download and dump result to tmp file
 | 
						|
	if _, err = io.Copy(tmpFile, progressReader); err != nil {
 | 
						|
		tmpFile.Close()
 | 
						|
		return
 | 
						|
	}
 | 
						|
	fmt.Fprintln(b.Stdout)
 | 
						|
	// ignoring error because the file was already opened successfully
 | 
						|
	tmpFileSt, err := tmpFile.Stat()
 | 
						|
	if err != nil {
 | 
						|
		tmpFile.Close()
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Set the mtime to the Last-Modified header value if present
 | 
						|
	// Otherwise just remove atime and mtime
 | 
						|
	mTime := time.Time{}
 | 
						|
 | 
						|
	lastMod := resp.Header.Get("Last-Modified")
 | 
						|
	if lastMod != "" {
 | 
						|
		// If we can't parse it then just let it default to 'zero'
 | 
						|
		// otherwise use the parsed time value
 | 
						|
		if parsedMTime, err := http.ParseTime(lastMod); err == nil {
 | 
						|
			mTime = parsedMTime
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	tmpFile.Close()
 | 
						|
 | 
						|
	if err = system.Chtimes(tmpFileName, mTime, mTime); err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Calc the checksum, even if we're using the cache
 | 
						|
	r, err := archive.Tar(tmpFileName, archive.Uncompressed)
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
	tarSum, err := tarsum.NewTarSum(r, true, tarsum.Version1)
 | 
						|
	if err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
	if _, err = io.Copy(ioutil.Discard, tarSum); err != nil {
 | 
						|
		return
 | 
						|
	}
 | 
						|
	hash := tarSum.Sum(nil)
 | 
						|
	r.Close()
 | 
						|
	return &builder.HashedFileInfo{FileInfo: builder.PathFileInfo{FileInfo: tmpFileSt, FilePath: tmpFileName}, FileHash: hash}, nil
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) calcCopyInfo(cmdName, origPath string, allowLocalDecompression, allowWildcards bool) ([]copyInfo, error) {
 | 
						|
 | 
						|
	// Work in daemon-specific OS filepath semantics
 | 
						|
	origPath = filepath.FromSlash(origPath)
 | 
						|
 | 
						|
	if origPath != "" && origPath[0] == os.PathSeparator && len(origPath) > 1 {
 | 
						|
		origPath = origPath[1:]
 | 
						|
	}
 | 
						|
	origPath = strings.TrimPrefix(origPath, "."+string(os.PathSeparator))
 | 
						|
 | 
						|
	// Deal with wildcards
 | 
						|
	if allowWildcards && containsWildcards(origPath) {
 | 
						|
		var copyInfos []copyInfo
 | 
						|
		if err := b.context.Walk("", func(path string, info builder.FileInfo, err error) error {
 | 
						|
			if err != nil {
 | 
						|
				return err
 | 
						|
			}
 | 
						|
			if info.Name() == "" {
 | 
						|
				// Why are we doing this check?
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
			if match, _ := filepath.Match(origPath, path); !match {
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
 | 
						|
			// Note we set allowWildcards to false in case the name has
 | 
						|
			// a * in it
 | 
						|
			subInfos, err := b.calcCopyInfo(cmdName, path, allowLocalDecompression, false)
 | 
						|
			if err != nil {
 | 
						|
				return err
 | 
						|
			}
 | 
						|
			copyInfos = append(copyInfos, subInfos...)
 | 
						|
			return nil
 | 
						|
		}); err != nil {
 | 
						|
			return nil, err
 | 
						|
		}
 | 
						|
		return copyInfos, nil
 | 
						|
	}
 | 
						|
 | 
						|
	// Must be a dir or a file
 | 
						|
 | 
						|
	statPath, fi, err := b.context.Stat(origPath)
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	copyInfos := []copyInfo{{FileInfo: fi, decompress: allowLocalDecompression}}
 | 
						|
 | 
						|
	hfi, handleHash := fi.(builder.Hashed)
 | 
						|
	if !handleHash {
 | 
						|
		return copyInfos, nil
 | 
						|
	}
 | 
						|
 | 
						|
	// Deal with the single file case
 | 
						|
	if !fi.IsDir() {
 | 
						|
		hfi.SetHash("file:" + hfi.Hash())
 | 
						|
		return copyInfos, nil
 | 
						|
	}
 | 
						|
	// Must be a dir
 | 
						|
	var subfiles []string
 | 
						|
	err = b.context.Walk(statPath, func(path string, info builder.FileInfo, err error) error {
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
		// we already checked handleHash above
 | 
						|
		subfiles = append(subfiles, info.(builder.Hashed).Hash())
 | 
						|
		return nil
 | 
						|
	})
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	}
 | 
						|
 | 
						|
	sort.Strings(subfiles)
 | 
						|
	hasher := sha256.New()
 | 
						|
	hasher.Write([]byte(strings.Join(subfiles, ",")))
 | 
						|
	hfi.SetHash("dir:" + hex.EncodeToString(hasher.Sum(nil)))
 | 
						|
 | 
						|
	return copyInfos, nil
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) processImageFrom(img builder.Image) error {
 | 
						|
	if img != nil {
 | 
						|
		b.image = img.ImageID()
 | 
						|
 | 
						|
		if img.RunConfig() != nil {
 | 
						|
			b.runConfig = img.RunConfig()
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Check to see if we have a default PATH, note that windows won't
 | 
						|
	// have one as its set by HCS
 | 
						|
	if system.DefaultPathEnv != "" {
 | 
						|
		// Convert the slice of strings that represent the current list
 | 
						|
		// of env vars into a map so we can see if PATH is already set.
 | 
						|
		// If its not set then go ahead and give it our default value
 | 
						|
		configEnv := opts.ConvertKVStringsToMap(b.runConfig.Env)
 | 
						|
		if _, ok := configEnv["PATH"]; !ok {
 | 
						|
			b.runConfig.Env = append(b.runConfig.Env,
 | 
						|
				"PATH="+system.DefaultPathEnv)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if img == nil {
 | 
						|
		// Typically this means they used "FROM scratch"
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	// Process ONBUILD triggers if they exist
 | 
						|
	if nTriggers := len(b.runConfig.OnBuild); nTriggers != 0 {
 | 
						|
		word := "trigger"
 | 
						|
		if nTriggers > 1 {
 | 
						|
			word = "triggers"
 | 
						|
		}
 | 
						|
		fmt.Fprintf(b.Stderr, "# Executing %d build %s...\n", nTriggers, word)
 | 
						|
	}
 | 
						|
 | 
						|
	// Copy the ONBUILD triggers, and remove them from the config, since the config will be committed.
 | 
						|
	onBuildTriggers := b.runConfig.OnBuild
 | 
						|
	b.runConfig.OnBuild = []string{}
 | 
						|
 | 
						|
	// parse the ONBUILD triggers by invoking the parser
 | 
						|
	for _, step := range onBuildTriggers {
 | 
						|
		ast, err := parser.Parse(strings.NewReader(step), &b.directive)
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
 | 
						|
		total := len(ast.Children)
 | 
						|
		for _, n := range ast.Children {
 | 
						|
			if err := b.checkDispatch(n, true); err != nil {
 | 
						|
				return err
 | 
						|
			}
 | 
						|
		}
 | 
						|
		for i, n := range ast.Children {
 | 
						|
			if err := b.dispatch(i, total, n); err != nil {
 | 
						|
				return err
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// probeCache checks if cache match can be found for current build instruction.
 | 
						|
// If an image is found, probeCache returns `(true, nil)`.
 | 
						|
// If no image is found, it returns `(false, nil)`.
 | 
						|
// If there is any error, it returns `(false, err)`.
 | 
						|
func (b *Builder) probeCache() (bool, error) {
 | 
						|
	c := b.imageCache
 | 
						|
	if c == nil || b.options.NoCache || b.cacheBusted {
 | 
						|
		return false, nil
 | 
						|
	}
 | 
						|
	cache, err := c.GetCache(b.image, b.runConfig)
 | 
						|
	if err != nil {
 | 
						|
		return false, err
 | 
						|
	}
 | 
						|
	if len(cache) == 0 {
 | 
						|
		logrus.Debugf("[BUILDER] Cache miss: %s", b.runConfig.Cmd)
 | 
						|
		b.cacheBusted = true
 | 
						|
		return false, nil
 | 
						|
	}
 | 
						|
 | 
						|
	fmt.Fprintf(b.Stdout, " ---> Using cache\n")
 | 
						|
	logrus.Debugf("[BUILDER] Use cached version: %s", b.runConfig.Cmd)
 | 
						|
	b.image = string(cache)
 | 
						|
 | 
						|
	return true, nil
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) create() (string, error) {
 | 
						|
	if b.image == "" && !b.noBaseImage {
 | 
						|
		return "", fmt.Errorf("Please provide a source image with `from` prior to run")
 | 
						|
	}
 | 
						|
	b.runConfig.Image = b.image
 | 
						|
 | 
						|
	resources := container.Resources{
 | 
						|
		CgroupParent: b.options.CgroupParent,
 | 
						|
		CPUShares:    b.options.CPUShares,
 | 
						|
		CPUPeriod:    b.options.CPUPeriod,
 | 
						|
		CPUQuota:     b.options.CPUQuota,
 | 
						|
		CpusetCpus:   b.options.CPUSetCPUs,
 | 
						|
		CpusetMems:   b.options.CPUSetMems,
 | 
						|
		Memory:       b.options.Memory,
 | 
						|
		MemorySwap:   b.options.MemorySwap,
 | 
						|
		Ulimits:      b.options.Ulimits,
 | 
						|
	}
 | 
						|
 | 
						|
	// TODO: why not embed a hostconfig in builder?
 | 
						|
	hostConfig := &container.HostConfig{
 | 
						|
		SecurityOpt: b.options.SecurityOpt,
 | 
						|
		Isolation:   b.options.Isolation,
 | 
						|
		ShmSize:     b.options.ShmSize,
 | 
						|
		Resources:   resources,
 | 
						|
		NetworkMode: container.NetworkMode(b.options.NetworkMode),
 | 
						|
	}
 | 
						|
 | 
						|
	config := *b.runConfig
 | 
						|
 | 
						|
	// Create the container
 | 
						|
	c, err := b.docker.ContainerCreate(types.ContainerCreateConfig{
 | 
						|
		Config:     b.runConfig,
 | 
						|
		HostConfig: hostConfig,
 | 
						|
	}, true)
 | 
						|
	if err != nil {
 | 
						|
		return "", err
 | 
						|
	}
 | 
						|
	for _, warning := range c.Warnings {
 | 
						|
		fmt.Fprintf(b.Stdout, " ---> [Warning] %s\n", warning)
 | 
						|
	}
 | 
						|
 | 
						|
	b.tmpContainers[c.ID] = struct{}{}
 | 
						|
	fmt.Fprintf(b.Stdout, " ---> Running in %s\n", stringid.TruncateID(c.ID))
 | 
						|
 | 
						|
	// override the entry point that may have been picked up from the base image
 | 
						|
	if err := b.docker.ContainerUpdateCmdOnBuild(c.ID, config.Cmd); err != nil {
 | 
						|
		return "", err
 | 
						|
	}
 | 
						|
 | 
						|
	return c.ID, nil
 | 
						|
}
 | 
						|
 | 
						|
var errCancelled = errors.New("build cancelled")
 | 
						|
 | 
						|
func (b *Builder) run(cID string) (err error) {
 | 
						|
	errCh := make(chan error)
 | 
						|
	go func() {
 | 
						|
		errCh <- b.docker.ContainerAttachRaw(cID, nil, b.Stdout, b.Stderr, true)
 | 
						|
	}()
 | 
						|
 | 
						|
	finished := make(chan struct{})
 | 
						|
	cancelErrCh := make(chan error, 1)
 | 
						|
	go func() {
 | 
						|
		select {
 | 
						|
		case <-b.clientCtx.Done():
 | 
						|
			logrus.Debugln("Build cancelled, killing and removing container:", cID)
 | 
						|
			b.docker.ContainerKill(cID, 0)
 | 
						|
			b.removeContainer(cID)
 | 
						|
			cancelErrCh <- errCancelled
 | 
						|
		case <-finished:
 | 
						|
			cancelErrCh <- nil
 | 
						|
		}
 | 
						|
	}()
 | 
						|
 | 
						|
	if err := b.docker.ContainerStart(cID, nil, true, "", ""); err != nil {
 | 
						|
		close(finished)
 | 
						|
		if cancelErr := <-cancelErrCh; cancelErr != nil {
 | 
						|
			logrus.Debugf("Build cancelled (%v) and got an error from ContainerStart: %v",
 | 
						|
				cancelErr, err)
 | 
						|
		}
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	// Block on reading output from container, stop on err or chan closed
 | 
						|
	if err := <-errCh; err != nil {
 | 
						|
		close(finished)
 | 
						|
		if cancelErr := <-cancelErrCh; cancelErr != nil {
 | 
						|
			logrus.Debugf("Build cancelled (%v) and got an error from errCh: %v",
 | 
						|
				cancelErr, err)
 | 
						|
		}
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	if ret, _ := b.docker.ContainerWait(cID, -1); ret != 0 {
 | 
						|
		close(finished)
 | 
						|
		if cancelErr := <-cancelErrCh; cancelErr != nil {
 | 
						|
			logrus.Debugf("Build cancelled (%v) and got a non-zero code from ContainerWait: %d",
 | 
						|
				cancelErr, ret)
 | 
						|
		}
 | 
						|
		// TODO: change error type, because jsonmessage.JSONError assumes HTTP
 | 
						|
		return &jsonmessage.JSONError{
 | 
						|
			Message: fmt.Sprintf("The command '%s' returned a non-zero code: %d", strings.Join(b.runConfig.Cmd, " "), ret),
 | 
						|
			Code:    ret,
 | 
						|
		}
 | 
						|
	}
 | 
						|
	close(finished)
 | 
						|
	return <-cancelErrCh
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) removeContainer(c string) error {
 | 
						|
	rmConfig := &types.ContainerRmConfig{
 | 
						|
		ForceRemove:  true,
 | 
						|
		RemoveVolume: true,
 | 
						|
	}
 | 
						|
	if err := b.docker.ContainerRm(c, rmConfig); err != nil {
 | 
						|
		fmt.Fprintf(b.Stdout, "Error removing intermediate container %s: %v\n", stringid.TruncateID(c), err)
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) clearTmp() {
 | 
						|
	for c := range b.tmpContainers {
 | 
						|
		if err := b.removeContainer(c); err != nil {
 | 
						|
			return
 | 
						|
		}
 | 
						|
		delete(b.tmpContainers, c)
 | 
						|
		fmt.Fprintf(b.Stdout, "Removing intermediate container %s\n", stringid.TruncateID(c))
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// readDockerfile reads a Dockerfile from the current context.
 | 
						|
func (b *Builder) readDockerfile() error {
 | 
						|
	// If no -f was specified then look for 'Dockerfile'. If we can't find
 | 
						|
	// that then look for 'dockerfile'.  If neither are found then default
 | 
						|
	// back to 'Dockerfile' and use that in the error message.
 | 
						|
	if b.options.Dockerfile == "" {
 | 
						|
		b.options.Dockerfile = builder.DefaultDockerfileName
 | 
						|
		if _, _, err := b.context.Stat(b.options.Dockerfile); os.IsNotExist(err) {
 | 
						|
			lowercase := strings.ToLower(b.options.Dockerfile)
 | 
						|
			if _, _, err := b.context.Stat(lowercase); err == nil {
 | 
						|
				b.options.Dockerfile = lowercase
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	err := b.parseDockerfile()
 | 
						|
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	// After the Dockerfile has been parsed, we need to check the .dockerignore
 | 
						|
	// file for either "Dockerfile" or ".dockerignore", and if either are
 | 
						|
	// present then erase them from the build context. These files should never
 | 
						|
	// have been sent from the client but we did send them to make sure that
 | 
						|
	// we had the Dockerfile to actually parse, and then we also need the
 | 
						|
	// .dockerignore file to know whether either file should be removed.
 | 
						|
	// Note that this assumes the Dockerfile has been read into memory and
 | 
						|
	// is now safe to be removed.
 | 
						|
	if dockerIgnore, ok := b.context.(builder.DockerIgnoreContext); ok {
 | 
						|
		dockerIgnore.Process([]string{b.options.Dockerfile})
 | 
						|
	}
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (b *Builder) parseDockerfile() error {
 | 
						|
	f, err := b.context.Open(b.options.Dockerfile)
 | 
						|
	if err != nil {
 | 
						|
		if os.IsNotExist(err) {
 | 
						|
			return fmt.Errorf("Cannot locate specified Dockerfile: %s", b.options.Dockerfile)
 | 
						|
		}
 | 
						|
		return err
 | 
						|
	}
 | 
						|
	defer f.Close()
 | 
						|
	if f, ok := f.(*os.File); ok {
 | 
						|
		// ignoring error because Open already succeeded
 | 
						|
		fi, err := f.Stat()
 | 
						|
		if err != nil {
 | 
						|
			return fmt.Errorf("Unexpected error reading Dockerfile: %v", err)
 | 
						|
		}
 | 
						|
		if fi.Size() == 0 {
 | 
						|
			return fmt.Errorf("The Dockerfile (%s) cannot be empty", b.options.Dockerfile)
 | 
						|
		}
 | 
						|
	}
 | 
						|
	b.dockerfile, err = parser.Parse(f, &b.directive)
 | 
						|
	if err != nil {
 | 
						|
		return err
 | 
						|
	}
 | 
						|
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
// determine if build arg is part of built-in args or user
 | 
						|
// defined args in Dockerfile at any point in time.
 | 
						|
func (b *Builder) isBuildArgAllowed(arg string) bool {
 | 
						|
	if _, ok := BuiltinAllowedBuildArgs[arg]; ok {
 | 
						|
		return true
 | 
						|
	}
 | 
						|
	if _, ok := b.allowedBuildArgs[arg]; ok {
 | 
						|
		return true
 | 
						|
	}
 | 
						|
	return false
 | 
						|
}
 |