mirror of
https://github.com/moby/moby.git
synced 2022-11-09 12:21:53 -05:00
b3ee9ac74e
Docker-DCO-1.1-Signed-off-by: Victor Vieux <vieux@docker.com> (github: vieux)
352 lines
8.4 KiB
Go
352 lines
8.4 KiB
Go
package docker
|
|
|
|
import (
|
|
"bytes"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/docker/docker/engine"
|
|
"github.com/docker/docker/runconfig"
|
|
)
|
|
|
|
func TestCreateNumberHostname(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
defer mkDaemonFromEngine(eng, t).Nuke()
|
|
|
|
config, _, _, err := runconfig.Parse([]string{"-h", "web.0", unitTestImageID, "echo test"}, nil)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
createTestContainer(eng, config, t)
|
|
}
|
|
|
|
func TestCommit(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
defer mkDaemonFromEngine(eng, t).Nuke()
|
|
|
|
config, _, _, err := runconfig.Parse([]string{unitTestImageID, "/bin/cat"}, nil)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
id := createTestContainer(eng, config, t)
|
|
|
|
job := eng.Job("commit", id)
|
|
job.Setenv("repo", "testrepo")
|
|
job.Setenv("tag", "testtag")
|
|
job.SetenvJson("config", config)
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
func TestMergeConfigOnCommit(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
runtime := mkDaemonFromEngine(eng, t)
|
|
defer runtime.Nuke()
|
|
|
|
container1, _, _ := mkContainer(runtime, []string{"-e", "FOO=bar", unitTestImageID, "echo test > /tmp/foo"}, t)
|
|
defer runtime.Destroy(container1)
|
|
|
|
config, _, _, err := runconfig.Parse([]string{container1.ID, "cat /tmp/foo"}, nil)
|
|
if err != nil {
|
|
t.Error(err)
|
|
}
|
|
|
|
job := eng.Job("commit", container1.ID)
|
|
job.Setenv("repo", "testrepo")
|
|
job.Setenv("tag", "testtag")
|
|
job.SetenvJson("config", config)
|
|
var outputBuffer = bytes.NewBuffer(nil)
|
|
job.Stdout.Add(outputBuffer)
|
|
if err := job.Run(); err != nil {
|
|
t.Error(err)
|
|
}
|
|
|
|
container2, _, _ := mkContainer(runtime, []string{engine.Tail(outputBuffer, 1)}, t)
|
|
defer runtime.Destroy(container2)
|
|
|
|
job = eng.Job("container_inspect", container1.Name)
|
|
baseContainer, _ := job.Stdout.AddEnv()
|
|
if err := job.Run(); err != nil {
|
|
t.Error(err)
|
|
}
|
|
|
|
job = eng.Job("container_inspect", container2.Name)
|
|
commitContainer, _ := job.Stdout.AddEnv()
|
|
if err := job.Run(); err != nil {
|
|
t.Error(err)
|
|
}
|
|
|
|
baseConfig := baseContainer.GetSubEnv("Config")
|
|
commitConfig := commitContainer.GetSubEnv("Config")
|
|
|
|
if commitConfig.Get("Env") != baseConfig.Get("Env") {
|
|
t.Fatalf("Env config in committed container should be %v, was %v",
|
|
baseConfig.Get("Env"), commitConfig.Get("Env"))
|
|
}
|
|
|
|
if baseConfig.Get("Cmd") != "[\"echo test \\u003e /tmp/foo\"]" {
|
|
t.Fatalf("Cmd in base container should be [\"echo test \\u003e /tmp/foo\"], was %s",
|
|
baseConfig.Get("Cmd"))
|
|
}
|
|
|
|
if commitConfig.Get("Cmd") != "[\"cat /tmp/foo\"]" {
|
|
t.Fatalf("Cmd in committed container should be [\"cat /tmp/foo\"], was %s",
|
|
commitConfig.Get("Cmd"))
|
|
}
|
|
}
|
|
|
|
func TestRestartKillWait(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
srv := mkServerFromEngine(eng, t)
|
|
runtime := mkDaemonFromEngine(eng, t)
|
|
defer runtime.Nuke()
|
|
|
|
config, hostConfig, _, err := runconfig.Parse([]string{"-i", unitTestImageID, "/bin/cat"}, nil)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
id := createTestContainer(eng, config, t)
|
|
|
|
job := eng.Job("containers")
|
|
job.SetenvBool("all", true)
|
|
outs, err := job.Stdout.AddListTable()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(outs.Data) != 1 {
|
|
t.Errorf("Expected 1 container, %v found", len(outs.Data))
|
|
}
|
|
|
|
job = eng.Job("start", id)
|
|
if err := job.ImportEnv(hostConfig); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
job = eng.Job("kill", id)
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
eng = newTestEngine(t, false, runtime.Config().Root)
|
|
srv = mkServerFromEngine(eng, t)
|
|
|
|
job = srv.Eng.Job("containers")
|
|
job.SetenvBool("all", true)
|
|
outs, err = job.Stdout.AddListTable()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(outs.Data) != 1 {
|
|
t.Errorf("Expected 1 container, %v found", len(outs.Data))
|
|
}
|
|
|
|
setTimeout(t, "Waiting on stopped container timedout", 5*time.Second, func() {
|
|
job = srv.Eng.Job("wait", outs.Data[0].Get("Id"))
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
})
|
|
}
|
|
|
|
func TestCreateStartRestartStopStartKillRm(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
srv := mkServerFromEngine(eng, t)
|
|
defer mkDaemonFromEngine(eng, t).Nuke()
|
|
|
|
config, hostConfig, _, err := runconfig.Parse([]string{"-i", unitTestImageID, "/bin/cat"}, nil)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
id := createTestContainer(eng, config, t)
|
|
|
|
job := srv.Eng.Job("containers")
|
|
job.SetenvBool("all", true)
|
|
outs, err := job.Stdout.AddListTable()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(outs.Data) != 1 {
|
|
t.Errorf("Expected 1 container, %v found", len(outs.Data))
|
|
}
|
|
|
|
job = eng.Job("start", id)
|
|
if err := job.ImportEnv(hostConfig); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
job = eng.Job("restart", id)
|
|
job.SetenvInt("t", 15)
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
job = eng.Job("stop", id)
|
|
job.SetenvInt("t", 15)
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
job = eng.Job("start", id)
|
|
if err := job.ImportEnv(hostConfig); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := eng.Job("kill", id).Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// FIXME: this failed once with a race condition ("Unable to remove filesystem for xxx: directory not empty")
|
|
job = eng.Job("container_delete", id)
|
|
job.SetenvBool("removeVolume", true)
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
job = srv.Eng.Job("containers")
|
|
job.SetenvBool("all", true)
|
|
outs, err = job.Stdout.AddListTable()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(outs.Data) != 0 {
|
|
t.Errorf("Expected 0 container, %v found", len(outs.Data))
|
|
}
|
|
}
|
|
|
|
func TestRunWithTooLowMemoryLimit(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
defer mkDaemonFromEngine(eng, t).Nuke()
|
|
|
|
// Try to create a container with a memory limit of 1 byte less than the minimum allowed limit.
|
|
job := eng.Job("create")
|
|
job.Setenv("Image", unitTestImageID)
|
|
job.Setenv("Memory", "524287")
|
|
job.Setenv("CpuShares", "1000")
|
|
job.SetenvList("Cmd", []string{"/bin/cat"})
|
|
if err := job.Run(); err == nil {
|
|
t.Errorf("Memory limit is smaller than the allowed limit. Container creation should've failed!")
|
|
}
|
|
}
|
|
|
|
func TestImagesFilter(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
defer nuke(mkDaemonFromEngine(eng, t))
|
|
|
|
if err := eng.Job("tag", unitTestImageName, "utest", "tag1").Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := eng.Job("tag", unitTestImageName, "utest/docker", "tag2").Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := eng.Job("tag", unitTestImageName, "utest:5000/docker", "tag3").Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
images := getImages(eng, t, false, "utest*/*")
|
|
|
|
if len(images.Data[0].GetList("RepoTags")) != 2 {
|
|
t.Fatal("incorrect number of matches returned")
|
|
}
|
|
|
|
images = getImages(eng, t, false, "utest")
|
|
|
|
if len(images.Data[0].GetList("RepoTags")) != 1 {
|
|
t.Fatal("incorrect number of matches returned")
|
|
}
|
|
|
|
images = getImages(eng, t, false, "utest*")
|
|
|
|
if len(images.Data[0].GetList("RepoTags")) != 1 {
|
|
t.Fatal("incorrect number of matches returned")
|
|
}
|
|
|
|
images = getImages(eng, t, false, "*5000*/*")
|
|
|
|
if len(images.Data[0].GetList("RepoTags")) != 1 {
|
|
t.Fatal("incorrect number of matches returned")
|
|
}
|
|
}
|
|
|
|
// Regression test for being able to untag an image with an existing
|
|
// container
|
|
func TestDeleteTagWithExistingContainers(t *testing.T) {
|
|
eng := NewTestEngine(t)
|
|
defer nuke(mkDaemonFromEngine(eng, t))
|
|
|
|
srv := mkServerFromEngine(eng, t)
|
|
|
|
// Tag the image
|
|
if err := eng.Job("tag", unitTestImageID, "utest", "tag1").Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Create a container from the image
|
|
config, _, _, err := runconfig.Parse([]string{unitTestImageID, "echo test"}, nil)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
id := createNamedTestContainer(eng, config, t, "testingtags")
|
|
if id == "" {
|
|
t.Fatal("No id returned")
|
|
}
|
|
|
|
job := srv.Eng.Job("containers")
|
|
job.SetenvBool("all", true)
|
|
outs, err := job.Stdout.AddListTable()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := job.Run(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(outs.Data) != 1 {
|
|
t.Fatalf("Expected 1 container got %d", len(outs.Data))
|
|
}
|
|
|
|
// Try to remove the tag
|
|
imgs := engine.NewTable("", 0)
|
|
if err := srv.DeleteImage("utest:tag1", imgs, true, false, false); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if len(imgs.Data) != 1 {
|
|
t.Fatalf("Should only have deleted one untag %d", len(imgs.Data))
|
|
}
|
|
|
|
if untag := imgs.Data[0].Get("Untagged"); untag != "utest:tag1" {
|
|
t.Fatalf("Expected %s got %s", unitTestImageID, untag)
|
|
}
|
|
}
|