mirror of
https://github.com/moby/moby.git
synced 2022-11-09 12:21:53 -05:00
Merge pull request #30834 from allencloud/refactor-daemon-cluster-handler-and-helper
refactor cluster/helper.go and split cluster.go into multiple files
This commit is contained in:
commit
b8a3595941
8 changed files with 1604 additions and 1527 deletions
File diff suppressed because it is too large
Load diff
|
@ -24,116 +24,178 @@ func getSwarm(ctx context.Context, c swarmapi.ControlClient) (*swarmapi.Cluster,
|
|||
|
||||
func getNode(ctx context.Context, c swarmapi.ControlClient, input string) (*swarmapi.Node, error) {
|
||||
// GetNode to match via full ID.
|
||||
rg, err := c.GetNode(ctx, &swarmapi.GetNodeRequest{NodeID: input})
|
||||
if err != nil {
|
||||
// If any error (including NotFound), ListNodes to match via full name.
|
||||
rl, err := c.ListNodes(ctx, &swarmapi.ListNodesRequest{Filters: &swarmapi.ListNodesRequest_Filters{Names: []string{input}}})
|
||||
|
||||
if err != nil || len(rl.Nodes) == 0 {
|
||||
// If any error or 0 result, ListNodes to match via ID prefix.
|
||||
rl, err = c.ListNodes(ctx, &swarmapi.ListNodesRequest{Filters: &swarmapi.ListNodesRequest_Filters{IDPrefixes: []string{input}}})
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Nodes) == 0 {
|
||||
err := fmt.Errorf("node %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Nodes); l > 1 {
|
||||
return nil, fmt.Errorf("node %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Nodes[0], nil
|
||||
if rg, err := c.GetNode(ctx, &swarmapi.GetNodeRequest{NodeID: input}); err == nil {
|
||||
return rg.Node, nil
|
||||
}
|
||||
return rg.Node, nil
|
||||
|
||||
// If any error (including NotFound), ListNodes to match via full name.
|
||||
rl, err := c.ListNodes(ctx, &swarmapi.ListNodesRequest{
|
||||
Filters: &swarmapi.ListNodesRequest_Filters{
|
||||
Names: []string{input},
|
||||
},
|
||||
})
|
||||
if err != nil || len(rl.Nodes) == 0 {
|
||||
// If any error or 0 result, ListNodes to match via ID prefix.
|
||||
rl, err = c.ListNodes(ctx, &swarmapi.ListNodesRequest{
|
||||
Filters: &swarmapi.ListNodesRequest_Filters{
|
||||
IDPrefixes: []string{input},
|
||||
},
|
||||
})
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Nodes) == 0 {
|
||||
err := fmt.Errorf("node %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Nodes); l > 1 {
|
||||
return nil, fmt.Errorf("node %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Nodes[0], nil
|
||||
}
|
||||
|
||||
func getService(ctx context.Context, c swarmapi.ControlClient, input string) (*swarmapi.Service, error) {
|
||||
// GetService to match via full ID.
|
||||
rg, err := c.GetService(ctx, &swarmapi.GetServiceRequest{ServiceID: input})
|
||||
if err != nil {
|
||||
// If any error (including NotFound), ListServices to match via full name.
|
||||
rl, err := c.ListServices(ctx, &swarmapi.ListServicesRequest{Filters: &swarmapi.ListServicesRequest_Filters{Names: []string{input}}})
|
||||
if err != nil || len(rl.Services) == 0 {
|
||||
// If any error or 0 result, ListServices to match via ID prefix.
|
||||
rl, err = c.ListServices(ctx, &swarmapi.ListServicesRequest{Filters: &swarmapi.ListServicesRequest_Filters{IDPrefixes: []string{input}}})
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Services) == 0 {
|
||||
err := fmt.Errorf("service %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Services); l > 1 {
|
||||
return nil, fmt.Errorf("service %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Services[0], nil
|
||||
if rg, err := c.GetService(ctx, &swarmapi.GetServiceRequest{ServiceID: input}); err == nil {
|
||||
return rg.Service, nil
|
||||
}
|
||||
return rg.Service, nil
|
||||
|
||||
// If any error (including NotFound), ListServices to match via full name.
|
||||
rl, err := c.ListServices(ctx, &swarmapi.ListServicesRequest{
|
||||
Filters: &swarmapi.ListServicesRequest_Filters{
|
||||
Names: []string{input},
|
||||
},
|
||||
})
|
||||
if err != nil || len(rl.Services) == 0 {
|
||||
// If any error or 0 result, ListServices to match via ID prefix.
|
||||
rl, err = c.ListServices(ctx, &swarmapi.ListServicesRequest{
|
||||
Filters: &swarmapi.ListServicesRequest_Filters{
|
||||
IDPrefixes: []string{input},
|
||||
},
|
||||
})
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Services) == 0 {
|
||||
err := fmt.Errorf("service %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Services); l > 1 {
|
||||
return nil, fmt.Errorf("service %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Services[0], nil
|
||||
}
|
||||
|
||||
func getTask(ctx context.Context, c swarmapi.ControlClient, input string) (*swarmapi.Task, error) {
|
||||
// GetTask to match via full ID.
|
||||
rg, err := c.GetTask(ctx, &swarmapi.GetTaskRequest{TaskID: input})
|
||||
if err != nil {
|
||||
// If any error (including NotFound), ListTasks to match via full name.
|
||||
rl, err := c.ListTasks(ctx, &swarmapi.ListTasksRequest{Filters: &swarmapi.ListTasksRequest_Filters{Names: []string{input}}})
|
||||
|
||||
if err != nil || len(rl.Tasks) == 0 {
|
||||
// If any error or 0 result, ListTasks to match via ID prefix.
|
||||
rl, err = c.ListTasks(ctx, &swarmapi.ListTasksRequest{Filters: &swarmapi.ListTasksRequest_Filters{IDPrefixes: []string{input}}})
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Tasks) == 0 {
|
||||
err := fmt.Errorf("task %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Tasks); l > 1 {
|
||||
return nil, fmt.Errorf("task %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Tasks[0], nil
|
||||
if rg, err := c.GetTask(ctx, &swarmapi.GetTaskRequest{TaskID: input}); err == nil {
|
||||
return rg.Task, nil
|
||||
}
|
||||
return rg.Task, nil
|
||||
|
||||
// If any error (including NotFound), ListTasks to match via full name.
|
||||
rl, err := c.ListTasks(ctx, &swarmapi.ListTasksRequest{
|
||||
Filters: &swarmapi.ListTasksRequest_Filters{
|
||||
Names: []string{input},
|
||||
},
|
||||
})
|
||||
if err != nil || len(rl.Tasks) == 0 {
|
||||
// If any error or 0 result, ListTasks to match via ID prefix.
|
||||
rl, err = c.ListTasks(ctx, &swarmapi.ListTasksRequest{
|
||||
Filters: &swarmapi.ListTasksRequest_Filters{
|
||||
IDPrefixes: []string{input},
|
||||
},
|
||||
})
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Tasks) == 0 {
|
||||
err := fmt.Errorf("task %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Tasks); l > 1 {
|
||||
return nil, fmt.Errorf("task %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Tasks[0], nil
|
||||
}
|
||||
|
||||
func getSecret(ctx context.Context, c swarmapi.ControlClient, input string) (*swarmapi.Secret, error) {
|
||||
// attempt to lookup secret by full ID
|
||||
if rg, err := c.GetSecret(ctx, &swarmapi.GetSecretRequest{SecretID: input}); err == nil {
|
||||
return rg.Secret, nil
|
||||
}
|
||||
|
||||
// If any error (including NotFound), ListSecrets to match via full name.
|
||||
rl, err := c.ListSecrets(ctx, &swarmapi.ListSecretsRequest{
|
||||
Filters: &swarmapi.ListSecretsRequest_Filters{
|
||||
Names: []string{input},
|
||||
},
|
||||
})
|
||||
if err != nil || len(rl.Secrets) == 0 {
|
||||
// If any error or 0 result, ListSecrets to match via ID prefix.
|
||||
rl, err = c.ListSecrets(ctx, &swarmapi.ListSecretsRequest{
|
||||
Filters: &swarmapi.ListSecretsRequest_Filters{
|
||||
IDPrefixes: []string{input},
|
||||
},
|
||||
})
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Secrets) == 0 {
|
||||
err := fmt.Errorf("secret %s not found", input)
|
||||
return nil, errors.NewRequestNotFoundError(err)
|
||||
}
|
||||
|
||||
if l := len(rl.Secrets); l > 1 {
|
||||
return nil, fmt.Errorf("secret %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Secrets[0], nil
|
||||
}
|
||||
|
||||
func getNetwork(ctx context.Context, c swarmapi.ControlClient, input string) (*swarmapi.Network, error) {
|
||||
// GetNetwork to match via full ID.
|
||||
rg, err := c.GetNetwork(ctx, &swarmapi.GetNetworkRequest{NetworkID: input})
|
||||
if err != nil {
|
||||
// If any error (including NotFound), ListNetworks to match via ID prefix and full name.
|
||||
rl, err := c.ListNetworks(ctx, &swarmapi.ListNetworksRequest{Filters: &swarmapi.ListNetworksRequest_Filters{Names: []string{input}}})
|
||||
if err != nil || len(rl.Networks) == 0 {
|
||||
rl, err = c.ListNetworks(ctx, &swarmapi.ListNetworksRequest{Filters: &swarmapi.ListNetworksRequest_Filters{IDPrefixes: []string{input}}})
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Networks) == 0 {
|
||||
return nil, fmt.Errorf("network %s not found", input)
|
||||
}
|
||||
|
||||
if l := len(rl.Networks); l > 1 {
|
||||
return nil, fmt.Errorf("network %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Networks[0], nil
|
||||
if rg, err := c.GetNetwork(ctx, &swarmapi.GetNetworkRequest{NetworkID: input}); err == nil {
|
||||
return rg.Network, nil
|
||||
}
|
||||
return rg.Network, nil
|
||||
|
||||
// If any error (including NotFound), ListNetworks to match via ID prefix and full name.
|
||||
rl, err := c.ListNetworks(ctx, &swarmapi.ListNetworksRequest{
|
||||
Filters: &swarmapi.ListNetworksRequest_Filters{
|
||||
Names: []string{input},
|
||||
},
|
||||
})
|
||||
if err != nil || len(rl.Networks) == 0 {
|
||||
rl, err = c.ListNetworks(ctx, &swarmapi.ListNetworksRequest{
|
||||
Filters: &swarmapi.ListNetworksRequest_Filters{
|
||||
IDPrefixes: []string{input},
|
||||
},
|
||||
})
|
||||
}
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if len(rl.Networks) == 0 {
|
||||
return nil, fmt.Errorf("network %s not found", input)
|
||||
}
|
||||
|
||||
if l := len(rl.Networks); l > 1 {
|
||||
return nil, fmt.Errorf("network %s is ambiguous (%d matches found)", input, l)
|
||||
}
|
||||
|
||||
return rl.Networks[0], nil
|
||||
}
|
||||
|
|
281
daemon/cluster/networks.go
Normal file
281
daemon/cluster/networks.go
Normal file
|
@ -0,0 +1,281 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
|
||||
"github.com/Sirupsen/logrus"
|
||||
apierrors "github.com/docker/docker/api/errors"
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
"github.com/docker/docker/api/types/network"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
"github.com/docker/docker/runconfig"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
"github.com/pkg/errors"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
// GetNetworks returns all current cluster managed networks.
|
||||
func (c *Cluster) GetNetworks() ([]apitypes.NetworkResource, error) {
|
||||
return c.getNetworks(nil)
|
||||
}
|
||||
|
||||
func (c *Cluster) getNetworks(filters *swarmapi.ListNetworksRequest_Filters) ([]apitypes.NetworkResource, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
r, err := state.controlClient.ListNetworks(ctx, &swarmapi.ListNetworksRequest{Filters: filters})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var networks []apitypes.NetworkResource
|
||||
|
||||
for _, network := range r.Networks {
|
||||
networks = append(networks, convert.BasicNetworkFromGRPC(*network))
|
||||
}
|
||||
|
||||
return networks, nil
|
||||
}
|
||||
|
||||
// GetNetwork returns a cluster network by an ID.
|
||||
func (c *Cluster) GetNetwork(input string) (apitypes.NetworkResource, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return apitypes.NetworkResource{}, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
network, err := getNetwork(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return apitypes.NetworkResource{}, err
|
||||
}
|
||||
return convert.BasicNetworkFromGRPC(*network), nil
|
||||
}
|
||||
|
||||
// GetNetworksByName returns cluster managed networks by name.
|
||||
// It is ok to have multiple networks here. #18864
|
||||
func (c *Cluster) GetNetworksByName(name string) ([]apitypes.NetworkResource, error) {
|
||||
// Note that swarmapi.GetNetworkRequest.Name is not functional.
|
||||
// So we cannot just use that with c.GetNetwork.
|
||||
return c.getNetworks(&swarmapi.ListNetworksRequest_Filters{
|
||||
Names: []string{name},
|
||||
})
|
||||
}
|
||||
|
||||
func attacherKey(target, containerID string) string {
|
||||
return containerID + ":" + target
|
||||
}
|
||||
|
||||
// UpdateAttachment signals the attachment config to the attachment
|
||||
// waiter who is trying to start or attach the container to the
|
||||
// network.
|
||||
func (c *Cluster) UpdateAttachment(target, containerID string, config *network.NetworkingConfig) error {
|
||||
c.mu.RLock()
|
||||
attacher, ok := c.attachers[attacherKey(target, containerID)]
|
||||
c.mu.RUnlock()
|
||||
if !ok || attacher == nil {
|
||||
return fmt.Errorf("could not find attacher for container %s to network %s", containerID, target)
|
||||
}
|
||||
|
||||
attacher.attachWaitCh <- config
|
||||
close(attacher.attachWaitCh)
|
||||
return nil
|
||||
}
|
||||
|
||||
// WaitForDetachment waits for the container to stop or detach from
|
||||
// the network.
|
||||
func (c *Cluster) WaitForDetachment(ctx context.Context, networkName, networkID, taskID, containerID string) error {
|
||||
c.mu.RLock()
|
||||
attacher, ok := c.attachers[attacherKey(networkName, containerID)]
|
||||
if !ok {
|
||||
attacher, ok = c.attachers[attacherKey(networkID, containerID)]
|
||||
}
|
||||
state := c.currentNodeState()
|
||||
if state.swarmNode == nil || state.swarmNode.Agent() == nil {
|
||||
c.mu.RUnlock()
|
||||
return errors.New("invalid cluster node while waiting for detachment")
|
||||
}
|
||||
|
||||
c.mu.RUnlock()
|
||||
agent := state.swarmNode.Agent()
|
||||
if ok && attacher != nil &&
|
||||
attacher.detachWaitCh != nil &&
|
||||
attacher.attachCompleteCh != nil {
|
||||
// Attachment may be in progress still so wait for
|
||||
// attachment to complete.
|
||||
select {
|
||||
case <-attacher.attachCompleteCh:
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
}
|
||||
|
||||
if attacher.taskID == taskID {
|
||||
select {
|
||||
case <-attacher.detachWaitCh:
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return agent.ResourceAllocator().DetachNetwork(ctx, taskID)
|
||||
}
|
||||
|
||||
// AttachNetwork generates an attachment request towards the manager.
|
||||
func (c *Cluster) AttachNetwork(target string, containerID string, addresses []string) (*network.NetworkingConfig, error) {
|
||||
aKey := attacherKey(target, containerID)
|
||||
c.mu.Lock()
|
||||
state := c.currentNodeState()
|
||||
if state.swarmNode == nil || state.swarmNode.Agent() == nil {
|
||||
c.mu.Unlock()
|
||||
return nil, errors.New("invalid cluster node while attaching to network")
|
||||
}
|
||||
if attacher, ok := c.attachers[aKey]; ok {
|
||||
c.mu.Unlock()
|
||||
return attacher.config, nil
|
||||
}
|
||||
|
||||
agent := state.swarmNode.Agent()
|
||||
attachWaitCh := make(chan *network.NetworkingConfig)
|
||||
detachWaitCh := make(chan struct{})
|
||||
attachCompleteCh := make(chan struct{})
|
||||
c.attachers[aKey] = &attacher{
|
||||
attachWaitCh: attachWaitCh,
|
||||
attachCompleteCh: attachCompleteCh,
|
||||
detachWaitCh: detachWaitCh,
|
||||
}
|
||||
c.mu.Unlock()
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
taskID, err := agent.ResourceAllocator().AttachNetwork(ctx, containerID, target, addresses)
|
||||
if err != nil {
|
||||
c.mu.Lock()
|
||||
delete(c.attachers, aKey)
|
||||
c.mu.Unlock()
|
||||
return nil, fmt.Errorf("Could not attach to network %s: %v", target, err)
|
||||
}
|
||||
|
||||
c.mu.Lock()
|
||||
c.attachers[aKey].taskID = taskID
|
||||
close(attachCompleteCh)
|
||||
c.mu.Unlock()
|
||||
|
||||
logrus.Debugf("Successfully attached to network %s with tid %s", target, taskID)
|
||||
|
||||
var config *network.NetworkingConfig
|
||||
select {
|
||||
case config = <-attachWaitCh:
|
||||
case <-ctx.Done():
|
||||
return nil, fmt.Errorf("attaching to network failed, make sure your network options are correct and check manager logs: %v", ctx.Err())
|
||||
}
|
||||
|
||||
c.mu.Lock()
|
||||
c.attachers[aKey].config = config
|
||||
c.mu.Unlock()
|
||||
return config, nil
|
||||
}
|
||||
|
||||
// DetachNetwork unblocks the waiters waiting on WaitForDetachment so
|
||||
// that a request to detach can be generated towards the manager.
|
||||
func (c *Cluster) DetachNetwork(target string, containerID string) error {
|
||||
aKey := attacherKey(target, containerID)
|
||||
|
||||
c.mu.Lock()
|
||||
attacher, ok := c.attachers[aKey]
|
||||
delete(c.attachers, aKey)
|
||||
c.mu.Unlock()
|
||||
|
||||
if !ok {
|
||||
return fmt.Errorf("could not find network attachment for container %s to network %s", containerID, target)
|
||||
}
|
||||
|
||||
close(attacher.detachWaitCh)
|
||||
return nil
|
||||
}
|
||||
|
||||
// CreateNetwork creates a new cluster managed network.
|
||||
func (c *Cluster) CreateNetwork(s apitypes.NetworkCreateRequest) (string, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return "", c.errNoManager(state)
|
||||
}
|
||||
|
||||
if runconfig.IsPreDefinedNetwork(s.Name) {
|
||||
err := fmt.Errorf("%s is a pre-defined network and cannot be created", s.Name)
|
||||
return "", apierrors.NewRequestForbiddenError(err)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
networkSpec := convert.BasicNetworkCreateToGRPC(s)
|
||||
r, err := state.controlClient.CreateNetwork(ctx, &swarmapi.CreateNetworkRequest{Spec: &networkSpec})
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
|
||||
return r.Network.ID, nil
|
||||
}
|
||||
|
||||
// RemoveNetwork removes a cluster network.
|
||||
func (c *Cluster) RemoveNetwork(input string) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
network, err := getNetwork(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
_, err = state.controlClient.RemoveNetwork(ctx, &swarmapi.RemoveNetworkRequest{NetworkID: network.ID})
|
||||
return err
|
||||
}
|
||||
|
||||
func (c *Cluster) populateNetworkID(ctx context.Context, client swarmapi.ControlClient, s *types.ServiceSpec) error {
|
||||
// Always prefer NetworkAttachmentConfigs from TaskTemplate
|
||||
// but fallback to service spec for backward compatibility
|
||||
networks := s.TaskTemplate.Networks
|
||||
if len(networks) == 0 {
|
||||
networks = s.Networks
|
||||
}
|
||||
|
||||
for i, n := range networks {
|
||||
apiNetwork, err := getNetwork(ctx, client, n.Target)
|
||||
if err != nil {
|
||||
if ln, _ := c.config.Backend.FindNetwork(n.Target); ln != nil && !ln.Info().Dynamic() {
|
||||
err = fmt.Errorf("The network %s cannot be used with services. Only networks scoped to the swarm can be used, such as those created with the overlay driver.", ln.Name())
|
||||
return apierrors.NewRequestForbiddenError(err)
|
||||
}
|
||||
return err
|
||||
}
|
||||
networks[i].Target = apiNetwork.ID
|
||||
}
|
||||
return nil
|
||||
}
|
120
daemon/cluster/nodes.go
Normal file
120
daemon/cluster/nodes.go
Normal file
|
@ -0,0 +1,120 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
apierrors "github.com/docker/docker/api/errors"
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
)
|
||||
|
||||
// GetNodes returns a list of all nodes known to a cluster.
|
||||
func (c *Cluster) GetNodes(options apitypes.NodeListOptions) ([]types.Node, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
filters, err := newListNodesFilters(options.Filters)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
r, err := state.controlClient.ListNodes(
|
||||
ctx,
|
||||
&swarmapi.ListNodesRequest{Filters: filters})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
nodes := []types.Node{}
|
||||
|
||||
for _, node := range r.Nodes {
|
||||
nodes = append(nodes, convert.NodeFromGRPC(*node))
|
||||
}
|
||||
return nodes, nil
|
||||
}
|
||||
|
||||
// GetNode returns a node based on an ID.
|
||||
func (c *Cluster) GetNode(input string) (types.Node, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return types.Node{}, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
node, err := getNode(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return types.Node{}, err
|
||||
}
|
||||
return convert.NodeFromGRPC(*node), nil
|
||||
}
|
||||
|
||||
// UpdateNode updates existing nodes properties.
|
||||
func (c *Cluster) UpdateNode(input string, version uint64, spec types.NodeSpec) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
nodeSpec, err := convert.NodeSpecToGRPC(spec)
|
||||
if err != nil {
|
||||
return apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
currentNode, err := getNode(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
_, err = state.controlClient.UpdateNode(
|
||||
ctx,
|
||||
&swarmapi.UpdateNodeRequest{
|
||||
NodeID: currentNode.ID,
|
||||
Spec: &nodeSpec,
|
||||
NodeVersion: &swarmapi.Version{
|
||||
Index: version,
|
||||
},
|
||||
},
|
||||
)
|
||||
return err
|
||||
}
|
||||
|
||||
// RemoveNode removes a node from a cluster
|
||||
func (c *Cluster) RemoveNode(input string, force bool) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
node, err := getNode(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
_, err = state.controlClient.RemoveNode(ctx, &swarmapi.RemoveNodeRequest{NodeID: node.ID, Force: force})
|
||||
return err
|
||||
}
|
|
@ -1,63 +1,14 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"strings"
|
||||
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
func getSecretByNameOrIDPrefix(ctx context.Context, state *nodeState, nameOrIDPrefix string) (*swarmapi.Secret, error) {
|
||||
// attempt to lookup secret by full ID
|
||||
if r, err := state.controlClient.GetSecret(ctx, &swarmapi.GetSecretRequest{
|
||||
SecretID: nameOrIDPrefix,
|
||||
}); err == nil {
|
||||
return r.Secret, nil
|
||||
}
|
||||
|
||||
// attempt to lookup secret by full name and partial ID
|
||||
// Note here ListSecretRequest_Filters operate with `or`
|
||||
r, err := state.controlClient.ListSecrets(ctx, &swarmapi.ListSecretsRequest{
|
||||
Filters: &swarmapi.ListSecretsRequest_Filters{
|
||||
Names: []string{nameOrIDPrefix},
|
||||
IDPrefixes: []string{nameOrIDPrefix},
|
||||
},
|
||||
})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
// attempt to lookup secret by full name
|
||||
for _, s := range r.Secrets {
|
||||
if s.Spec.Annotations.Name == nameOrIDPrefix {
|
||||
return s, nil
|
||||
}
|
||||
}
|
||||
// attempt to lookup secret by partial ID (prefix)
|
||||
// return error if more than one matches found (ambiguous)
|
||||
n := 0
|
||||
var found *swarmapi.Secret
|
||||
for _, s := range r.Secrets {
|
||||
if strings.HasPrefix(s.ID, nameOrIDPrefix) {
|
||||
found = s
|
||||
n++
|
||||
}
|
||||
}
|
||||
if n > 1 {
|
||||
return nil, fmt.Errorf("secret %s is ambiguous (%d matches found)", nameOrIDPrefix, n)
|
||||
}
|
||||
if found == nil {
|
||||
return nil, fmt.Errorf("no such secret: %s", nameOrIDPrefix)
|
||||
}
|
||||
return found, nil
|
||||
}
|
||||
|
||||
// GetSecret returns a secret from a managed swarm cluster
|
||||
func (c *Cluster) GetSecret(nameOrIDPrefix string) (types.Secret, error) {
|
||||
func (c *Cluster) GetSecret(input string) (types.Secret, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
|
@ -69,7 +20,7 @@ func (c *Cluster) GetSecret(nameOrIDPrefix string) (types.Secret, error) {
|
|||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
secret, err := getSecretByNameOrIDPrefix(ctx, &state, nameOrIDPrefix)
|
||||
secret, err := getSecret(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return types.Secret{}, err
|
||||
}
|
||||
|
@ -133,7 +84,7 @@ func (c *Cluster) CreateSecret(s types.SecretSpec) (string, error) {
|
|||
}
|
||||
|
||||
// RemoveSecret removes a secret from a managed swarm cluster.
|
||||
func (c *Cluster) RemoveSecret(nameOrIDPrefix string) error {
|
||||
func (c *Cluster) RemoveSecret(input string) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
|
@ -145,7 +96,7 @@ func (c *Cluster) RemoveSecret(nameOrIDPrefix string) error {
|
|||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
secret, err := getSecretByNameOrIDPrefix(ctx, &state, nameOrIDPrefix)
|
||||
secret, err := getSecret(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
|
389
daemon/cluster/services.go
Normal file
389
daemon/cluster/services.go
Normal file
|
@ -0,0 +1,389 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
"encoding/base64"
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io"
|
||||
"os"
|
||||
"strings"
|
||||
|
||||
"github.com/Sirupsen/logrus"
|
||||
"github.com/docker/distribution/reference"
|
||||
apierrors "github.com/docker/docker/api/errors"
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
"github.com/docker/docker/api/types/backend"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
"github.com/docker/docker/daemon/logger"
|
||||
"github.com/docker/docker/pkg/ioutils"
|
||||
"github.com/docker/docker/pkg/stdcopy"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
gogotypes "github.com/gogo/protobuf/types"
|
||||
"github.com/pkg/errors"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
// GetServices returns all services of a managed swarm cluster.
|
||||
func (c *Cluster) GetServices(options apitypes.ServiceListOptions) ([]types.Service, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
filters, err := newListServicesFilters(options.Filters)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
r, err := state.controlClient.ListServices(
|
||||
ctx,
|
||||
&swarmapi.ListServicesRequest{Filters: filters})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
services := []types.Service{}
|
||||
|
||||
for _, service := range r.Services {
|
||||
services = append(services, convert.ServiceFromGRPC(*service))
|
||||
}
|
||||
|
||||
return services, nil
|
||||
}
|
||||
|
||||
// GetService returns a service based on an ID or name.
|
||||
func (c *Cluster) GetService(input string) (types.Service, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return types.Service{}, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
service, err := getService(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return types.Service{}, err
|
||||
}
|
||||
return convert.ServiceFromGRPC(*service), nil
|
||||
}
|
||||
|
||||
// CreateService creates a new service in a managed swarm cluster.
|
||||
func (c *Cluster) CreateService(s types.ServiceSpec, encodedAuth string) (*apitypes.ServiceCreateResponse, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
err := c.populateNetworkID(ctx, state.controlClient, &s)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
serviceSpec, err := convert.ServiceSpecToGRPC(s)
|
||||
if err != nil {
|
||||
return nil, apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
ctnr := serviceSpec.Task.GetContainer()
|
||||
if ctnr == nil {
|
||||
return nil, errors.New("service does not use container tasks")
|
||||
}
|
||||
|
||||
if encodedAuth != "" {
|
||||
ctnr.PullOptions = &swarmapi.ContainerSpec_PullOptions{RegistryAuth: encodedAuth}
|
||||
}
|
||||
|
||||
// retrieve auth config from encoded auth
|
||||
authConfig := &apitypes.AuthConfig{}
|
||||
if encodedAuth != "" {
|
||||
if err := json.NewDecoder(base64.NewDecoder(base64.URLEncoding, strings.NewReader(encodedAuth))).Decode(authConfig); err != nil {
|
||||
logrus.Warnf("invalid authconfig: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
resp := &apitypes.ServiceCreateResponse{}
|
||||
|
||||
// pin image by digest
|
||||
if os.Getenv("DOCKER_SERVICE_PREFER_OFFLINE_IMAGE") != "1" {
|
||||
digestImage, err := c.imageWithDigestString(ctx, ctnr.Image, authConfig)
|
||||
if err != nil {
|
||||
logrus.Warnf("unable to pin image %s to digest: %s", ctnr.Image, err.Error())
|
||||
resp.Warnings = append(resp.Warnings, fmt.Sprintf("unable to pin image %s to digest: %s", ctnr.Image, err.Error()))
|
||||
} else if ctnr.Image != digestImage {
|
||||
logrus.Debugf("pinning image %s by digest: %s", ctnr.Image, digestImage)
|
||||
ctnr.Image = digestImage
|
||||
} else {
|
||||
logrus.Debugf("creating service using supplied digest reference %s", ctnr.Image)
|
||||
}
|
||||
}
|
||||
|
||||
r, err := state.controlClient.CreateService(ctx, &swarmapi.CreateServiceRequest{Spec: &serviceSpec})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
resp.ID = r.Service.ID
|
||||
return resp, nil
|
||||
}
|
||||
|
||||
// UpdateService updates existing service to match new properties.
|
||||
func (c *Cluster) UpdateService(serviceIDOrName string, version uint64, spec types.ServiceSpec, encodedAuth string, registryAuthFrom string) (*apitypes.ServiceUpdateResponse, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
err := c.populateNetworkID(ctx, state.controlClient, &spec)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
serviceSpec, err := convert.ServiceSpecToGRPC(spec)
|
||||
if err != nil {
|
||||
return nil, apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
currentService, err := getService(ctx, state.controlClient, serviceIDOrName)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
newCtnr := serviceSpec.Task.GetContainer()
|
||||
if newCtnr == nil {
|
||||
return nil, errors.New("service does not use container tasks")
|
||||
}
|
||||
|
||||
if encodedAuth != "" {
|
||||
newCtnr.PullOptions = &swarmapi.ContainerSpec_PullOptions{RegistryAuth: encodedAuth}
|
||||
} else {
|
||||
// this is needed because if the encodedAuth isn't being updated then we
|
||||
// shouldn't lose it, and continue to use the one that was already present
|
||||
var ctnr *swarmapi.ContainerSpec
|
||||
switch registryAuthFrom {
|
||||
case apitypes.RegistryAuthFromSpec, "":
|
||||
ctnr = currentService.Spec.Task.GetContainer()
|
||||
case apitypes.RegistryAuthFromPreviousSpec:
|
||||
if currentService.PreviousSpec == nil {
|
||||
return nil, errors.New("service does not have a previous spec")
|
||||
}
|
||||
ctnr = currentService.PreviousSpec.Task.GetContainer()
|
||||
default:
|
||||
return nil, errors.New("unsupported registryAuthFrom value")
|
||||
}
|
||||
if ctnr == nil {
|
||||
return nil, errors.New("service does not use container tasks")
|
||||
}
|
||||
newCtnr.PullOptions = ctnr.PullOptions
|
||||
// update encodedAuth so it can be used to pin image by digest
|
||||
if ctnr.PullOptions != nil {
|
||||
encodedAuth = ctnr.PullOptions.RegistryAuth
|
||||
}
|
||||
}
|
||||
|
||||
// retrieve auth config from encoded auth
|
||||
authConfig := &apitypes.AuthConfig{}
|
||||
if encodedAuth != "" {
|
||||
if err := json.NewDecoder(base64.NewDecoder(base64.URLEncoding, strings.NewReader(encodedAuth))).Decode(authConfig); err != nil {
|
||||
logrus.Warnf("invalid authconfig: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
resp := &apitypes.ServiceUpdateResponse{}
|
||||
|
||||
// pin image by digest
|
||||
if os.Getenv("DOCKER_SERVICE_PREFER_OFFLINE_IMAGE") != "1" {
|
||||
digestImage, err := c.imageWithDigestString(ctx, newCtnr.Image, authConfig)
|
||||
if err != nil {
|
||||
logrus.Warnf("unable to pin image %s to digest: %s", newCtnr.Image, err.Error())
|
||||
resp.Warnings = append(resp.Warnings, fmt.Sprintf("unable to pin image %s to digest: %s", newCtnr.Image, err.Error()))
|
||||
} else if newCtnr.Image != digestImage {
|
||||
logrus.Debugf("pinning image %s by digest: %s", newCtnr.Image, digestImage)
|
||||
newCtnr.Image = digestImage
|
||||
} else {
|
||||
logrus.Debugf("updating service using supplied digest reference %s", newCtnr.Image)
|
||||
}
|
||||
}
|
||||
|
||||
_, err = state.controlClient.UpdateService(
|
||||
ctx,
|
||||
&swarmapi.UpdateServiceRequest{
|
||||
ServiceID: currentService.ID,
|
||||
Spec: &serviceSpec,
|
||||
ServiceVersion: &swarmapi.Version{
|
||||
Index: version,
|
||||
},
|
||||
},
|
||||
)
|
||||
|
||||
return resp, err
|
||||
}
|
||||
|
||||
// RemoveService removes a service from a managed swarm cluster.
|
||||
func (c *Cluster) RemoveService(input string) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
service, err := getService(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
_, err = state.controlClient.RemoveService(ctx, &swarmapi.RemoveServiceRequest{ServiceID: service.ID})
|
||||
return err
|
||||
}
|
||||
|
||||
// ServiceLogs collects service logs and writes them back to `config.OutStream`
|
||||
func (c *Cluster) ServiceLogs(ctx context.Context, input string, config *backend.ContainerLogsConfig, started chan struct{}) error {
|
||||
c.mu.RLock()
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
c.mu.RUnlock()
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
service, err := getService(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
c.mu.RUnlock()
|
||||
return err
|
||||
}
|
||||
|
||||
stream, err := state.logsClient.SubscribeLogs(ctx, &swarmapi.SubscribeLogsRequest{
|
||||
Selector: &swarmapi.LogSelector{
|
||||
ServiceIDs: []string{service.ID},
|
||||
},
|
||||
Options: &swarmapi.LogSubscriptionOptions{
|
||||
Follow: config.Follow,
|
||||
},
|
||||
})
|
||||
if err != nil {
|
||||
c.mu.RUnlock()
|
||||
return err
|
||||
}
|
||||
|
||||
wf := ioutils.NewWriteFlusher(config.OutStream)
|
||||
defer wf.Close()
|
||||
close(started)
|
||||
wf.Flush()
|
||||
|
||||
outStream := stdcopy.NewStdWriter(wf, stdcopy.Stdout)
|
||||
errStream := stdcopy.NewStdWriter(wf, stdcopy.Stderr)
|
||||
|
||||
// Release the lock before starting the stream.
|
||||
c.mu.RUnlock()
|
||||
for {
|
||||
// Check the context before doing anything.
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
return ctx.Err()
|
||||
default:
|
||||
}
|
||||
|
||||
subscribeMsg, err := stream.Recv()
|
||||
if err == io.EOF {
|
||||
return nil
|
||||
}
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
for _, msg := range subscribeMsg.Messages {
|
||||
data := []byte{}
|
||||
|
||||
if config.Timestamps {
|
||||
ts, err := gogotypes.TimestampFromProto(msg.Timestamp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
data = append(data, []byte(ts.Format(logger.TimeFormat)+" ")...)
|
||||
}
|
||||
|
||||
data = append(data, []byte(fmt.Sprintf("%s.node.id=%s,%s.service.id=%s,%s.task.id=%s ",
|
||||
contextPrefix, msg.Context.NodeID,
|
||||
contextPrefix, msg.Context.ServiceID,
|
||||
contextPrefix, msg.Context.TaskID,
|
||||
))...)
|
||||
|
||||
data = append(data, msg.Data...)
|
||||
|
||||
switch msg.Stream {
|
||||
case swarmapi.LogStreamStdout:
|
||||
outStream.Write(data)
|
||||
case swarmapi.LogStreamStderr:
|
||||
errStream.Write(data)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// imageWithDigestString takes an image such as name or name:tag
|
||||
// and returns the image pinned to a digest, such as name@sha256:34234
|
||||
func (c *Cluster) imageWithDigestString(ctx context.Context, image string, authConfig *apitypes.AuthConfig) (string, error) {
|
||||
ref, err := reference.ParseAnyReference(image)
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
namedRef, ok := ref.(reference.Named)
|
||||
if !ok {
|
||||
if _, ok := ref.(reference.Digested); ok {
|
||||
return "", errors.New("image reference is an image ID")
|
||||
}
|
||||
return "", errors.Errorf("unknown image reference format: %s", image)
|
||||
}
|
||||
// only query registry if not a canonical reference (i.e. with digest)
|
||||
if _, ok := namedRef.(reference.Canonical); !ok {
|
||||
namedRef = reference.TagNameOnly(namedRef)
|
||||
|
||||
taggedRef, ok := namedRef.(reference.NamedTagged)
|
||||
if !ok {
|
||||
return "", errors.Errorf("image reference not tagged: %s", image)
|
||||
}
|
||||
|
||||
repo, _, err := c.config.Backend.GetRepository(ctx, taggedRef, authConfig)
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
dscrptr, err := repo.Tags(ctx).Get(ctx, taggedRef.Tag())
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
|
||||
namedDigestedRef, err := reference.WithDigest(taggedRef, dscrptr.Digest)
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
// return familiar form until interface updated to return type
|
||||
return reference.FamiliarString(namedDigestedRef), nil
|
||||
}
|
||||
// reference already contains a digest, so just return it
|
||||
return reference.FamiliarString(ref), nil
|
||||
}
|
562
daemon/cluster/swarm.go
Normal file
562
daemon/cluster/swarm.go
Normal file
|
@ -0,0 +1,562 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"net"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/Sirupsen/logrus"
|
||||
apierrors "github.com/docker/docker/api/errors"
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
"github.com/docker/docker/api/types/filters"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
"github.com/docker/docker/opts"
|
||||
"github.com/docker/docker/pkg/signal"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
"github.com/docker/swarmkit/manager/encryption"
|
||||
swarmnode "github.com/docker/swarmkit/node"
|
||||
"github.com/pkg/errors"
|
||||
"golang.org/x/net/context"
|
||||
)
|
||||
|
||||
// Init initializes new cluster from user provided request.
|
||||
func (c *Cluster) Init(req types.InitRequest) (string, error) {
|
||||
c.controlMutex.Lock()
|
||||
defer c.controlMutex.Unlock()
|
||||
c.mu.Lock()
|
||||
if c.nr != nil {
|
||||
if req.ForceNewCluster {
|
||||
if err := c.nr.Stop(); err != nil {
|
||||
c.mu.Unlock()
|
||||
return "", err
|
||||
}
|
||||
} else {
|
||||
c.mu.Unlock()
|
||||
return "", errSwarmExists
|
||||
}
|
||||
}
|
||||
c.mu.Unlock()
|
||||
|
||||
if err := validateAndSanitizeInitRequest(&req); err != nil {
|
||||
return "", apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
|
||||
advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
|
||||
localAddr := listenHost
|
||||
|
||||
// If the local address is undetermined, the advertise address
|
||||
// will be used as local address, if it belongs to this system.
|
||||
// If the advertise address is not local, then we try to find
|
||||
// a system address to use as local address. If this fails,
|
||||
// we give up and ask the user to pass the listen address.
|
||||
if net.ParseIP(localAddr).IsUnspecified() {
|
||||
advertiseIP := net.ParseIP(advertiseHost)
|
||||
|
||||
found := false
|
||||
for _, systemIP := range listSystemIPs() {
|
||||
if systemIP.Equal(advertiseIP) {
|
||||
localAddr = advertiseIP.String()
|
||||
found = true
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if !found {
|
||||
ip, err := c.resolveSystemAddr()
|
||||
if err != nil {
|
||||
logrus.Warnf("Could not find a local address: %v", err)
|
||||
return "", errMustSpecifyListenAddr
|
||||
}
|
||||
localAddr = ip.String()
|
||||
}
|
||||
}
|
||||
|
||||
if !req.ForceNewCluster {
|
||||
clearPersistentState(c.root)
|
||||
}
|
||||
|
||||
nr, err := c.newNodeRunner(nodeStartConfig{
|
||||
forceNewCluster: req.ForceNewCluster,
|
||||
autolock: req.AutoLockManagers,
|
||||
LocalAddr: localAddr,
|
||||
ListenAddr: net.JoinHostPort(listenHost, listenPort),
|
||||
AdvertiseAddr: net.JoinHostPort(advertiseHost, advertisePort),
|
||||
availability: req.Availability,
|
||||
})
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
c.mu.Lock()
|
||||
c.nr = nr
|
||||
c.mu.Unlock()
|
||||
|
||||
if err := <-nr.Ready(); err != nil {
|
||||
if !req.ForceNewCluster { // if failure on first attempt don't keep state
|
||||
if err := clearPersistentState(c.root); err != nil {
|
||||
return "", err
|
||||
}
|
||||
}
|
||||
if err != nil {
|
||||
c.mu.Lock()
|
||||
c.nr = nil
|
||||
c.mu.Unlock()
|
||||
}
|
||||
return "", err
|
||||
}
|
||||
state := nr.State()
|
||||
if state.swarmNode == nil { // should never happen but protect from panic
|
||||
return "", errors.New("invalid cluster state for spec initialization")
|
||||
}
|
||||
if err := initClusterSpec(state.swarmNode, req.Spec); err != nil {
|
||||
return "", err
|
||||
}
|
||||
return state.NodeID(), nil
|
||||
}
|
||||
|
||||
// Join makes current Cluster part of an existing swarm cluster.
|
||||
func (c *Cluster) Join(req types.JoinRequest) error {
|
||||
c.controlMutex.Lock()
|
||||
defer c.controlMutex.Unlock()
|
||||
c.mu.Lock()
|
||||
if c.nr != nil {
|
||||
c.mu.Unlock()
|
||||
return errSwarmExists
|
||||
}
|
||||
c.mu.Unlock()
|
||||
|
||||
if err := validateAndSanitizeJoinRequest(&req); err != nil {
|
||||
return apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
listenHost, listenPort, err := resolveListenAddr(req.ListenAddr)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
var advertiseAddr string
|
||||
if req.AdvertiseAddr != "" {
|
||||
advertiseHost, advertisePort, err := c.resolveAdvertiseAddr(req.AdvertiseAddr, listenPort)
|
||||
// For joining, we don't need to provide an advertise address,
|
||||
// since the remote side can detect it.
|
||||
if err == nil {
|
||||
advertiseAddr = net.JoinHostPort(advertiseHost, advertisePort)
|
||||
}
|
||||
}
|
||||
|
||||
clearPersistentState(c.root)
|
||||
|
||||
nr, err := c.newNodeRunner(nodeStartConfig{
|
||||
RemoteAddr: req.RemoteAddrs[0],
|
||||
ListenAddr: net.JoinHostPort(listenHost, listenPort),
|
||||
AdvertiseAddr: advertiseAddr,
|
||||
joinAddr: req.RemoteAddrs[0],
|
||||
joinToken: req.JoinToken,
|
||||
availability: req.Availability,
|
||||
})
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
c.mu.Lock()
|
||||
c.nr = nr
|
||||
c.mu.Unlock()
|
||||
|
||||
select {
|
||||
case <-time.After(swarmConnectTimeout):
|
||||
return errSwarmJoinTimeoutReached
|
||||
case err := <-nr.Ready():
|
||||
if err != nil {
|
||||
c.mu.Lock()
|
||||
c.nr = nil
|
||||
c.mu.Unlock()
|
||||
}
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
// Inspect retrieves the configuration properties of a managed swarm cluster.
|
||||
func (c *Cluster) Inspect() (types.Swarm, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return types.Swarm{}, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
swarm, err := getSwarm(ctx, state.controlClient)
|
||||
if err != nil {
|
||||
return types.Swarm{}, err
|
||||
}
|
||||
|
||||
return convert.SwarmFromGRPC(*swarm), nil
|
||||
}
|
||||
|
||||
// Update updates configuration of a managed swarm cluster.
|
||||
func (c *Cluster) Update(version uint64, spec types.Spec, flags types.UpdateFlags) error {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
swarm, err := getSwarm(ctx, state.controlClient)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// In update, client should provide the complete spec of the swarm, including
|
||||
// Name and Labels. If a field is specified with 0 or nil, then the default value
|
||||
// will be used to swarmkit.
|
||||
clusterSpec, err := convert.SwarmSpecToGRPC(spec)
|
||||
if err != nil {
|
||||
return apierrors.NewBadRequestError(err)
|
||||
}
|
||||
|
||||
_, err = state.controlClient.UpdateCluster(
|
||||
ctx,
|
||||
&swarmapi.UpdateClusterRequest{
|
||||
ClusterID: swarm.ID,
|
||||
Spec: &clusterSpec,
|
||||
ClusterVersion: &swarmapi.Version{
|
||||
Index: version,
|
||||
},
|
||||
Rotation: swarmapi.KeyRotation{
|
||||
WorkerJoinToken: flags.RotateWorkerToken,
|
||||
ManagerJoinToken: flags.RotateManagerToken,
|
||||
ManagerUnlockKey: flags.RotateManagerUnlockKey,
|
||||
},
|
||||
},
|
||||
)
|
||||
return err
|
||||
}
|
||||
|
||||
// GetUnlockKey returns the unlock key for the swarm.
|
||||
func (c *Cluster) GetUnlockKey() (string, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return "", c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
client := swarmapi.NewCAClient(state.grpcConn)
|
||||
|
||||
r, err := client.GetUnlockKey(ctx, &swarmapi.GetUnlockKeyRequest{})
|
||||
if err != nil {
|
||||
return "", err
|
||||
}
|
||||
|
||||
if len(r.UnlockKey) == 0 {
|
||||
// no key
|
||||
return "", nil
|
||||
}
|
||||
|
||||
return encryption.HumanReadableKey(r.UnlockKey), nil
|
||||
}
|
||||
|
||||
// UnlockSwarm provides a key to decrypt data that is encrypted at rest.
|
||||
func (c *Cluster) UnlockSwarm(req types.UnlockRequest) error {
|
||||
c.controlMutex.Lock()
|
||||
defer c.controlMutex.Unlock()
|
||||
|
||||
c.mu.RLock()
|
||||
state := c.currentNodeState()
|
||||
|
||||
if !state.IsActiveManager() {
|
||||
// when manager is not active,
|
||||
// unless it is locked, otherwise return error.
|
||||
if err := c.errNoManager(state); err != errSwarmLocked {
|
||||
c.mu.RUnlock()
|
||||
return err
|
||||
}
|
||||
} else {
|
||||
// when manager is active, return an error of "not locked"
|
||||
c.mu.RUnlock()
|
||||
return errors.New("swarm is not locked")
|
||||
}
|
||||
|
||||
// only when swarm is locked, code running reaches here
|
||||
nr := c.nr
|
||||
c.mu.RUnlock()
|
||||
|
||||
key, err := encryption.ParseHumanReadableKey(req.UnlockKey)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
config := nr.config
|
||||
config.lockKey = key
|
||||
if err := nr.Stop(); err != nil {
|
||||
return err
|
||||
}
|
||||
nr, err = c.newNodeRunner(config)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
c.mu.Lock()
|
||||
c.nr = nr
|
||||
c.mu.Unlock()
|
||||
|
||||
if err := <-nr.Ready(); err != nil {
|
||||
if errors.Cause(err) == errSwarmLocked {
|
||||
return errors.New("swarm could not be unlocked: invalid key provided")
|
||||
}
|
||||
return fmt.Errorf("swarm component could not be started: %v", err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Leave shuts down Cluster and removes current state.
|
||||
func (c *Cluster) Leave(force bool) error {
|
||||
c.controlMutex.Lock()
|
||||
defer c.controlMutex.Unlock()
|
||||
|
||||
c.mu.Lock()
|
||||
nr := c.nr
|
||||
if nr == nil {
|
||||
c.mu.Unlock()
|
||||
return errNoSwarm
|
||||
}
|
||||
|
||||
state := c.currentNodeState()
|
||||
|
||||
if errors.Cause(state.err) == errSwarmLocked && !force {
|
||||
// leave a locked swarm without --force is not allowed
|
||||
c.mu.Unlock()
|
||||
return errors.New("Swarm is encrypted and locked. Please unlock it first or use `--force` to ignore this message.")
|
||||
}
|
||||
|
||||
if state.IsManager() && !force {
|
||||
msg := "You are attempting to leave the swarm on a node that is participating as a manager. "
|
||||
if state.IsActiveManager() {
|
||||
active, reachable, unreachable, err := managerStats(state.controlClient, state.NodeID())
|
||||
if err == nil {
|
||||
if active && removingManagerCausesLossOfQuorum(reachable, unreachable) {
|
||||
if isLastManager(reachable, unreachable) {
|
||||
msg += "Removing the last manager erases all current state of the swarm. Use `--force` to ignore this message. "
|
||||
c.mu.Unlock()
|
||||
return errors.New(msg)
|
||||
}
|
||||
msg += fmt.Sprintf("Removing this node leaves %v managers out of %v. Without a Raft quorum your swarm will be inaccessible. ", reachable-1, reachable+unreachable)
|
||||
}
|
||||
}
|
||||
} else {
|
||||
msg += "Doing so may lose the consensus of your cluster. "
|
||||
}
|
||||
|
||||
msg += "The only way to restore a swarm that has lost consensus is to reinitialize it with `--force-new-cluster`. Use `--force` to suppress this message."
|
||||
c.mu.Unlock()
|
||||
return errors.New(msg)
|
||||
}
|
||||
// release readers in here
|
||||
if err := nr.Stop(); err != nil {
|
||||
logrus.Errorf("failed to shut down cluster node: %v", err)
|
||||
signal.DumpStacks("")
|
||||
c.mu.Unlock()
|
||||
return err
|
||||
}
|
||||
c.nr = nil
|
||||
c.mu.Unlock()
|
||||
if nodeID := state.NodeID(); nodeID != "" {
|
||||
nodeContainers, err := c.listContainerForNode(nodeID)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
for _, id := range nodeContainers {
|
||||
if err := c.config.Backend.ContainerRm(id, &apitypes.ContainerRmConfig{ForceRemove: true}); err != nil {
|
||||
logrus.Errorf("error removing %v: %v", id, err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
c.configEvent <- struct{}{}
|
||||
// todo: cleanup optional?
|
||||
if err := clearPersistentState(c.root); err != nil {
|
||||
return err
|
||||
}
|
||||
c.config.Backend.DaemonLeavesCluster()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Info returns information about the current cluster state.
|
||||
func (c *Cluster) Info() types.Info {
|
||||
info := types.Info{
|
||||
NodeAddr: c.GetAdvertiseAddress(),
|
||||
}
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
info.LocalNodeState = state.status
|
||||
if state.err != nil {
|
||||
info.Error = state.err.Error()
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
if state.IsActiveManager() {
|
||||
info.ControlAvailable = true
|
||||
swarm, err := c.Inspect()
|
||||
if err != nil {
|
||||
info.Error = err.Error()
|
||||
}
|
||||
|
||||
// Strip JoinTokens
|
||||
info.Cluster = swarm.ClusterInfo
|
||||
|
||||
if r, err := state.controlClient.ListNodes(ctx, &swarmapi.ListNodesRequest{}); err != nil {
|
||||
info.Error = err.Error()
|
||||
} else {
|
||||
info.Nodes = len(r.Nodes)
|
||||
for _, n := range r.Nodes {
|
||||
if n.ManagerStatus != nil {
|
||||
info.Managers = info.Managers + 1
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if state.swarmNode != nil {
|
||||
for _, r := range state.swarmNode.Remotes() {
|
||||
info.RemoteManagers = append(info.RemoteManagers, types.Peer{NodeID: r.NodeID, Addr: r.Addr})
|
||||
}
|
||||
info.NodeID = state.swarmNode.NodeID()
|
||||
}
|
||||
|
||||
return info
|
||||
}
|
||||
|
||||
func validateAndSanitizeInitRequest(req *types.InitRequest) error {
|
||||
var err error
|
||||
req.ListenAddr, err = validateAddr(req.ListenAddr)
|
||||
if err != nil {
|
||||
return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
|
||||
}
|
||||
|
||||
if req.Spec.Annotations.Name == "" {
|
||||
req.Spec.Annotations.Name = "default"
|
||||
} else if req.Spec.Annotations.Name != "default" {
|
||||
return errors.New(`swarm spec must be named "default"`)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func validateAndSanitizeJoinRequest(req *types.JoinRequest) error {
|
||||
var err error
|
||||
req.ListenAddr, err = validateAddr(req.ListenAddr)
|
||||
if err != nil {
|
||||
return fmt.Errorf("invalid ListenAddr %q: %v", req.ListenAddr, err)
|
||||
}
|
||||
if len(req.RemoteAddrs) == 0 {
|
||||
return errors.New("at least 1 RemoteAddr is required to join")
|
||||
}
|
||||
for i := range req.RemoteAddrs {
|
||||
req.RemoteAddrs[i], err = validateAddr(req.RemoteAddrs[i])
|
||||
if err != nil {
|
||||
return fmt.Errorf("invalid remoteAddr %q: %v", req.RemoteAddrs[i], err)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func validateAddr(addr string) (string, error) {
|
||||
if addr == "" {
|
||||
return addr, errors.New("invalid empty address")
|
||||
}
|
||||
newaddr, err := opts.ParseTCPAddr(addr, defaultAddr)
|
||||
if err != nil {
|
||||
return addr, nil
|
||||
}
|
||||
return strings.TrimPrefix(newaddr, "tcp://"), nil
|
||||
}
|
||||
|
||||
func initClusterSpec(node *swarmnode.Node, spec types.Spec) error {
|
||||
ctx, _ := context.WithTimeout(context.Background(), 5*time.Second)
|
||||
for conn := range node.ListenControlSocket(ctx) {
|
||||
if ctx.Err() != nil {
|
||||
return ctx.Err()
|
||||
}
|
||||
if conn != nil {
|
||||
client := swarmapi.NewControlClient(conn)
|
||||
var cluster *swarmapi.Cluster
|
||||
for i := 0; ; i++ {
|
||||
lcr, err := client.ListClusters(ctx, &swarmapi.ListClustersRequest{})
|
||||
if err != nil {
|
||||
return fmt.Errorf("error on listing clusters: %v", err)
|
||||
}
|
||||
if len(lcr.Clusters) == 0 {
|
||||
if i < 10 {
|
||||
time.Sleep(200 * time.Millisecond)
|
||||
continue
|
||||
}
|
||||
return errors.New("empty list of clusters was returned")
|
||||
}
|
||||
cluster = lcr.Clusters[0]
|
||||
break
|
||||
}
|
||||
// In init, we take the initial default values from swarmkit, and merge
|
||||
// any non nil or 0 value from spec to GRPC spec. This will leave the
|
||||
// default value alone.
|
||||
// Note that this is different from Update(), as in Update() we expect
|
||||
// user to specify the complete spec of the cluster (as they already know
|
||||
// the existing one and knows which field to update)
|
||||
clusterSpec, err := convert.MergeSwarmSpecToGRPC(spec, cluster.Spec)
|
||||
if err != nil {
|
||||
return fmt.Errorf("error updating cluster settings: %v", err)
|
||||
}
|
||||
_, err = client.UpdateCluster(ctx, &swarmapi.UpdateClusterRequest{
|
||||
ClusterID: cluster.ID,
|
||||
ClusterVersion: &cluster.Meta.Version,
|
||||
Spec: &clusterSpec,
|
||||
})
|
||||
if err != nil {
|
||||
return fmt.Errorf("error updating cluster settings: %v", err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
return ctx.Err()
|
||||
}
|
||||
|
||||
func (c *Cluster) listContainerForNode(nodeID string) ([]string, error) {
|
||||
var ids []string
|
||||
filters := filters.NewArgs()
|
||||
filters.Add("label", fmt.Sprintf("com.docker.swarm.node.id=%s", nodeID))
|
||||
containers, err := c.config.Backend.Containers(&apitypes.ContainerListOptions{
|
||||
Filters: filters,
|
||||
})
|
||||
if err != nil {
|
||||
return []string{}, err
|
||||
}
|
||||
for _, c := range containers {
|
||||
ids = append(ids, c.ID)
|
||||
}
|
||||
return ids, nil
|
||||
}
|
90
daemon/cluster/tasks.go
Normal file
90
daemon/cluster/tasks.go
Normal file
|
@ -0,0 +1,90 @@
|
|||
package cluster
|
||||
|
||||
import (
|
||||
apitypes "github.com/docker/docker/api/types"
|
||||
"github.com/docker/docker/api/types/filters"
|
||||
types "github.com/docker/docker/api/types/swarm"
|
||||
"github.com/docker/docker/daemon/cluster/convert"
|
||||
swarmapi "github.com/docker/swarmkit/api"
|
||||
)
|
||||
|
||||
// GetTasks returns a list of tasks matching the filter options.
|
||||
func (c *Cluster) GetTasks(options apitypes.TaskListOptions) ([]types.Task, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return nil, c.errNoManager(state)
|
||||
}
|
||||
|
||||
byName := func(filter filters.Args) error {
|
||||
if filter.Include("service") {
|
||||
serviceFilters := filter.Get("service")
|
||||
for _, serviceFilter := range serviceFilters {
|
||||
service, err := c.GetService(serviceFilter)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
filter.Del("service", serviceFilter)
|
||||
filter.Add("service", service.ID)
|
||||
}
|
||||
}
|
||||
if filter.Include("node") {
|
||||
nodeFilters := filter.Get("node")
|
||||
for _, nodeFilter := range nodeFilters {
|
||||
node, err := c.GetNode(nodeFilter)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
filter.Del("node", nodeFilter)
|
||||
filter.Add("node", node.ID)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
filters, err := newListTasksFilters(options.Filters, byName)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
r, err := state.controlClient.ListTasks(
|
||||
ctx,
|
||||
&swarmapi.ListTasksRequest{Filters: filters})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
tasks := []types.Task{}
|
||||
|
||||
for _, task := range r.Tasks {
|
||||
if task.Spec.GetContainer() != nil {
|
||||
tasks = append(tasks, convert.TaskFromGRPC(*task))
|
||||
}
|
||||
}
|
||||
return tasks, nil
|
||||
}
|
||||
|
||||
// GetTask returns a task by an ID.
|
||||
func (c *Cluster) GetTask(input string) (types.Task, error) {
|
||||
c.mu.RLock()
|
||||
defer c.mu.RUnlock()
|
||||
|
||||
state := c.currentNodeState()
|
||||
if !state.IsActiveManager() {
|
||||
return types.Task{}, c.errNoManager(state)
|
||||
}
|
||||
|
||||
ctx, cancel := c.getRequestContext()
|
||||
defer cancel()
|
||||
|
||||
task, err := getTask(ctx, state.controlClient, input)
|
||||
if err != nil {
|
||||
return types.Task{}, err
|
||||
}
|
||||
return convert.TaskFromGRPC(*task), nil
|
||||
}
|
Loading…
Add table
Reference in a new issue