1
0
Fork 0
mirror of https://github.com/moby/moby.git synced 2022-11-09 12:21:53 -05:00
Moby Project - a collaborative project for the container ecosystem to assemble container-based systems
Find a file
Vincent Batts fb7ceeb170 cleaner handling of client socket access
In the go stdlib net/http Transport, the used connections are cached
when idled. This behaviour is intended for TCP connections and does not
behave correctly for unix sockets. Despite the
DefaultMaxIdleConnsPerHost being 2, the idled connections are held open
during a session. For large sessions like `docker rm $(docker ps -a -q)`
of thousands of containers, it will cause the client _and_ the server to
open too many fails and have failures.

Having keep alives not used for only unix sockets is a work around for
this stdlib issue.

Also this includes disabling compression when communicating over the
local unix socket too.

Signed-off-by: Vincent Batts <vbatts@redhat.com>
2014-10-10 19:58:49 -04:00
api cleaner handling of client socket access 2014-10-10 19:58:49 -04:00
builder Allow child to overwrite entrypoint from parent 2014-10-07 23:45:35 +00:00
builtins
contrib Fix system socket/service unit files 2014-10-07 14:09:08 -04:00
daemon Merge pull request #8457 from jfrazelle/pr_8455 2014-10-08 16:21:14 -07:00
docker Add libtrust key identity management 2014-09-26 15:52:08 -07:00
dockerinit
dockerversion
docs Merge pull request #8419 from nhsiehgit/dockerfiletut 2014-10-09 10:15:10 -07:00
engine Fix error paring null JSON - Issue7941 2014-09-28 19:51:00 -07:00
events
graph Increase the tag length limit from 30 to 128 2014-10-07 11:29:47 -04:00
hack Merge pull request #8320 from dmcgowan/provenance_pull 2014-10-03 10:56:54 -07:00
image Move archive package into pkg/archive 2014-09-29 23:23:36 -07:00
integration Fix #6231 - Accept chunked encoding on start 2014-10-03 14:18:25 -04:00
integration-cli Merge pull request #8457 from jfrazelle/pr_8455 2014-10-08 16:21:14 -07:00
links
nat
opts
pkg Merge pull request #8350 from erikh/add_erikh_maintainer_proxy 2014-10-01 15:50:41 -07:00
reexec
registry Add comment for permission and fix wrong format variable 2014-10-02 17:41:57 -07:00
runconfig Add --security-opts options to allow user to customize security configuration 2014-09-30 00:06:22 +00:00
trust Add provenance pull flow for official images 2014-10-01 18:26:06 -07:00
utils Move Matches() file path matching function into pkg/fileutils 2014-09-29 23:21:41 -07:00
vendor Merge pull request #8320 from dmcgowan/provenance_pull 2014-10-03 10:56:54 -07:00
volumes Restore volume refs after daemon restart 2014-10-08 14:17:27 -04:00
.dockerignore
.drone.yml
.gitignore
.mailmap
.travis.yml
AUTHORS
CHANGELOG.md
CONTRIBUTING.md
Dockerfile bump Go to 1.3.3 2014-10-01 17:14:48 +03:00
LICENSE
MAINTAINERS
Makefile
NOTICE
README.md
VERSION

Docker: the Linux container engine

Docker is an open source project to pack, ship and run any application as a lightweight container

Docker containers are both hardware-agnostic and platform-agnostic. This means that they can run anywhere, from your laptop to the largest EC2 compute instance and everything in between - and they don't require that you use a particular language, framework or packaging system. That makes them great building blocks for deploying and scaling web apps, databases and backend services without depending on a particular stack or provider.

Docker is an open-source implementation of the deployment engine which powers dotCloud, a popular Platform-as-a-Service. It benefits directly from the experience accumulated over several years of large-scale operation and support of hundreds of thousands of applications and databases.

Docker L

Security Disclosure

Security is very important to us. If you have any issue regarding security, please disclose the information responsibly by sending an email to security@docker.com and not by creating a github issue.

Better than VMs

A common method for distributing applications and sandboxing their execution is to use virtual machines, or VMs. Typical VM formats are VMWare's vmdk, Oracle Virtualbox's vdi, and Amazon EC2's ami. In theory these formats should allow every developer to automatically package their application into a "machine" for easy distribution and deployment. In practice, that almost never happens, for a few reasons:

  • Size: VMs are very large which makes them impractical to store and transfer.
  • Performance: running VMs consumes significant CPU and memory, which makes them impractical in many scenarios, for example local development of multi-tier applications, and large-scale deployment of cpu and memory-intensive applications on large numbers of machines.
  • Portability: competing VM environments don't play well with each other. Although conversion tools do exist, they are limited and add even more overhead.
  • Hardware-centric: VMs were designed with machine operators in mind, not software developers. As a result, they offer very limited tooling for what developers need most: building, testing and running their software. For example, VMs offer no facilities for application versioning, monitoring, configuration, logging or service discovery.

By contrast, Docker relies on a different sandboxing method known as containerization. Unlike traditional virtualization, containerization takes place at the kernel level. Most modern operating system kernels now support the primitives necessary for containerization, including Linux with openvz, vserver and more recently lxc, Solaris with zones and FreeBSD with Jails.

Docker builds on top of these low-level primitives to offer developers a portable format and runtime environment that solves all 4 problems. Docker containers are small (and their transfer can be optimized with layers), they have basically zero memory and cpu overhead, they are completely portable and are designed from the ground up with an application-centric design.

The best part: because Docker operates at the OS level, it can still be run inside a VM!

Plays well with others

Docker does not require that you buy into a particular programming language, framework, packaging system or configuration language.

Is your application a Unix process? Does it use files, tcp connections, environment variables, standard Unix streams and command-line arguments as inputs and outputs? Then Docker can run it.

Can your application's build be expressed as a sequence of such commands? Then Docker can build it.

Escape dependency hell

A common problem for developers is the difficulty of managing all their application's dependencies in a simple and automated way.

This is usually difficult for several reasons:

  • Cross-platform dependencies. Modern applications often depend on a combination of system libraries and binaries, language-specific packages, framework-specific modules, internal components developed for another project, etc. These dependencies live in different "worlds" and require different tools - these tools typically don't work well with each other, requiring awkward custom integrations.

  • Conflicting dependencies. Different applications may depend on different versions of the same dependency. Packaging tools handle these situations with various degrees of ease - but they all handle them in different and incompatible ways, which again forces the developer to do extra work.

  • Custom dependencies. A developer may need to prepare a custom version of their application's dependency. Some packaging systems can handle custom versions of a dependency, others can't - and all of them handle it differently.

Docker solves dependency hell by giving the developer a simple way to express all their application's dependencies in one place, and streamline the process of assembling them. If this makes you think of XKCD 927, don't worry. Docker doesn't replace your favorite packaging systems. It simply orchestrates their use in a simple and repeatable way. How does it do that? With layers.

Docker defines a build as running a sequence of Unix commands, one after the other, in the same container. Build commands modify the contents of the container (usually by installing new files on the filesystem), the next command modifies it some more, etc. Since each build command inherits the result of the previous commands, the order in which the commands are executed expresses dependencies.

Here's a typical Docker build process:

FROM ubuntu:12.04
RUN apt-get update && apt-get install -y python python-pip curl
RUN curl -sSL https://github.com/shykes/helloflask/archive/master.tar.gz | tar -xzv
RUN cd helloflask-master && pip install -r requirements.txt

Note that Docker doesn't care how dependencies are built - as long as they can be built by running a Unix command in a container.

Getting started

Docker can be installed on your local machine as well as servers - both bare metal and virtualized. It is available as a binary on most modern Linux systems, or as a VM on Windows, Mac and other systems.

We also offer an interactive tutorial for quickly learning the basics of using Docker.

For up-to-date install instructions, see the Docs.

Usage examples

Docker can be used to run short-lived commands, long-running daemons (app servers, databases etc.), interactive shell sessions, etc.

You can find a list of real-world examples in the documentation.

Under the hood

Under the hood, Docker is built on the following components:

Contributing to Docker

GoDoc Travis

Want to hack on Docker? Awesome! There are instructions to get you started here.

They are probably not perfect, please let us know if anything feels wrong or incomplete.

Brought to you courtesy of our legal counsel. For more context, please see the Notice document.

Use and transfer of Docker may be subject to certain restrictions by the United States and other governments.
It is your responsibility to ensure that your use and/or transfer does not violate applicable laws.

For more information, please see http://www.bis.doc.gov

Licensing

Docker is licensed under the Apache License, Version 2.0. See LICENSE for full license text.