2014-08-05 01:29:14 -04:00
|
|
|
package graph
|
|
|
|
|
|
|
|
import (
|
|
|
|
"net/http"
|
|
|
|
"net/url"
|
|
|
|
|
|
|
|
"github.com/docker/docker/engine"
|
2014-09-30 02:23:36 -04:00
|
|
|
"github.com/docker/docker/pkg/archive"
|
2014-08-05 01:29:14 -04:00
|
|
|
"github.com/docker/docker/utils"
|
|
|
|
)
|
|
|
|
|
|
|
|
func (s *TagStore) CmdImport(job *engine.Job) engine.Status {
|
|
|
|
if n := len(job.Args); n != 2 && n != 3 {
|
|
|
|
return job.Errorf("Usage: %s SRC REPO [TAG]", job.Name)
|
|
|
|
}
|
|
|
|
var (
|
|
|
|
src = job.Args[0]
|
|
|
|
repo = job.Args[1]
|
|
|
|
tag string
|
|
|
|
sf = utils.NewStreamFormatter(job.GetenvBool("json"))
|
|
|
|
archive archive.ArchiveReader
|
|
|
|
resp *http.Response
|
|
|
|
)
|
|
|
|
if len(job.Args) > 2 {
|
|
|
|
tag = job.Args[2]
|
|
|
|
}
|
|
|
|
|
|
|
|
if src == "-" {
|
|
|
|
archive = job.Stdin
|
|
|
|
} else {
|
|
|
|
u, err := url.Parse(src)
|
|
|
|
if err != nil {
|
|
|
|
return job.Error(err)
|
|
|
|
}
|
|
|
|
if u.Scheme == "" {
|
|
|
|
u.Scheme = "http"
|
|
|
|
u.Host = src
|
|
|
|
u.Path = ""
|
|
|
|
}
|
|
|
|
job.Stdout.Write(sf.FormatStatus("", "Downloading from %s", u))
|
|
|
|
resp, err = utils.Download(u.String())
|
|
|
|
if err != nil {
|
|
|
|
return job.Error(err)
|
|
|
|
}
|
|
|
|
progressReader := utils.ProgressReader(resp.Body, int(resp.ContentLength), job.Stdout, sf, true, "", "Importing")
|
|
|
|
defer progressReader.Close()
|
|
|
|
archive = progressReader
|
|
|
|
}
|
|
|
|
img, err := s.graph.Create(archive, "", "", "Imported from "+src, "", nil, nil)
|
|
|
|
if err != nil {
|
|
|
|
return job.Error(err)
|
|
|
|
}
|
|
|
|
// Optionally register the image at REPO/TAG
|
|
|
|
if repo != "" {
|
|
|
|
if err := s.Set(repo, tag, img.ID, true); err != nil {
|
|
|
|
return job.Error(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
job.Stdout.Write(sf.FormatStatus("", img.ID))
|
|
|
|
return engine.StatusOK
|
|
|
|
}
|