5130fe5d |
package daemon
import (
"fmt"
"io"
"io/ioutil" |
ab30e19b |
"strings" |
bfebdfde |
"sync" |
5130fe5d |
|
ee7dd44c |
log "github.com/Sirupsen/logrus" |
5130fe5d |
"github.com/docker/docker/daemon/execdriver" |
ab30e19b |
"github.com/docker/docker/daemon/execdriver/lxc" |
5130fe5d |
"github.com/docker/docker/engine"
"github.com/docker/docker/pkg/broadcastwriter" |
7a9c944b |
"github.com/docker/docker/pkg/common" |
5130fe5d |
"github.com/docker/docker/pkg/ioutils" |
b845a621 |
"github.com/docker/docker/pkg/promise" |
5130fe5d |
"github.com/docker/docker/runconfig"
)
|
bfebdfde |
type execConfig struct { |
39030382 |
sync.Mutex |
bfebdfde |
ID string |
c786a8ee |
Running bool |
90928eb1 |
ExitCode int |
5130fe5d |
ProcessConfig execdriver.ProcessConfig |
e1cf95b5 |
StreamConfig |
bfebdfde |
OpenStdin bool
OpenStderr bool
OpenStdout bool
Container *Container |
5130fe5d |
}
|
bfebdfde |
type execStore struct {
s map[string]*execConfig |
4b43a6df |
sync.RWMutex |
bfebdfde |
}
func newExecStore() *execStore {
return &execStore{s: make(map[string]*execConfig, 0)}
}
func (e *execStore) Add(id string, execConfig *execConfig) {
e.Lock()
e.s[id] = execConfig
e.Unlock()
}
func (e *execStore) Get(id string) *execConfig { |
4b43a6df |
e.RLock() |
bfebdfde |
res := e.s[id] |
4b43a6df |
e.RUnlock() |
bfebdfde |
return res
}
func (e *execStore) Delete(id string) {
e.Lock()
delete(e.s, id)
e.Unlock()
}
|
4b43a6df |
func (e *execStore) List() []string {
var IDs []string
e.RLock() |
bbb92e14 |
for id := range e.s { |
4b43a6df |
IDs = append(IDs, id)
}
e.RUnlock()
return IDs
}
|
bfebdfde |
func (execConfig *execConfig) Resize(h, w int) error {
return execConfig.ProcessConfig.Terminal.Resize(h, w)
}
func (d *Daemon) registerExecCommand(execConfig *execConfig) { |
5b794c41 |
// Storing execs in container in order to kill them gracefully whenever the container is stopped or removed. |
bfebdfde |
execConfig.Container.execCommands.Add(execConfig.ID, execConfig)
// Storing execs in daemon for easy access via remote API.
d.execCommands.Add(execConfig.ID, execConfig)
}
func (d *Daemon) getExecConfig(name string) (*execConfig, error) {
if execConfig := d.execCommands.Get(name); execConfig != nil {
if !execConfig.Container.IsRunning() { |
d19d8008 |
return nil, fmt.Errorf("Container %s is not running", execConfig.Container.ID) |
bfebdfde |
}
return execConfig, nil |
5130fe5d |
}
|
021ecb1d |
return nil, fmt.Errorf("No such exec instance '%s' found in daemon", name) |
bfebdfde |
}
func (d *Daemon) unregisterExecCommand(execConfig *execConfig) {
execConfig.Container.execCommands.Delete(execConfig.ID)
d.execCommands.Delete(execConfig.ID)
} |
5130fe5d |
|
bfebdfde |
func (d *Daemon) getActiveContainer(name string) (*Container, error) { |
d25a6537 |
container, err := d.Get(name)
if err != nil {
return nil, err |
5130fe5d |
}
|
669561c2 |
if !container.IsRunning() { |
d19d8008 |
return nil, fmt.Errorf("Container %s is not running", name) |
5130fe5d |
} |
1bb02117 |
if container.IsPaused() {
return nil, fmt.Errorf("Container %s is paused, unpause the container before exec", name)
} |
bfebdfde |
return container, nil
} |
5130fe5d |
|
bfebdfde |
func (d *Daemon) ContainerExecCreate(job *engine.Job) engine.Status {
if len(job.Args) != 1 {
return job.Errorf("Usage: %s [options] container command [args]", job.Name) |
5130fe5d |
} |
bfebdfde |
|
ab30e19b |
if strings.HasPrefix(d.execDriver.Name(), lxc.DriverName) {
return job.Error(lxc.ErrExec)
}
|
bfebdfde |
var name = job.Args[0]
container, err := d.getActiveContainer(name)
if err != nil {
return job.Error(err) |
5130fe5d |
}
|
c8a3d313 |
config, err := runconfig.ExecConfigFromJob(job)
if err != nil {
return job.Error(err)
} |
bfebdfde |
|
5130fe5d |
entrypoint, args := d.getEntrypointAndArgs(nil, config.Cmd)
processConfig := execdriver.ProcessConfig{
Tty: config.Tty,
Entrypoint: entrypoint,
Arguments: args,
}
|
bfebdfde |
execConfig := &execConfig{ |
7a9c944b |
ID: common.GenerateRandomID(), |
5130fe5d |
OpenStdin: config.AttachStdin, |
bfebdfde |
OpenStdout: config.AttachStdout,
OpenStderr: config.AttachStderr, |
5130fe5d |
StreamConfig: StreamConfig{},
ProcessConfig: processConfig, |
bfebdfde |
Container: container, |
c786a8ee |
Running: false, |
bfebdfde |
}
|
e3d813f3 |
container.LogEvent("exec_create: " + execConfig.ProcessConfig.Entrypoint + " " + strings.Join(execConfig.ProcessConfig.Arguments, " "))
|
bfebdfde |
d.registerExecCommand(execConfig)
job.Printf("%s\n", execConfig.ID)
return engine.StatusOK
}
func (d *Daemon) ContainerExecStart(job *engine.Job) engine.Status { |
c786a8ee |
if len(job.Args) != 1 {
return job.Errorf("Usage: %s [options] exec", job.Name) |
bfebdfde |
}
var (
cStdin io.ReadCloser
cStdout, cStderr io.Writer
execName = job.Args[0]
)
|
39030382 |
execConfig, err := d.getExecConfig(execName)
if err != nil {
return job.Error(err) |
bfebdfde |
}
|
39030382 |
func() {
execConfig.Lock()
defer execConfig.Unlock()
if execConfig.Running {
err = fmt.Errorf("Error: Exec command %s is already running", execName)
}
execConfig.Running = true
}() |
bfebdfde |
if err != nil {
return job.Error(err)
}
|
c786a8ee |
log.Debugf("starting exec command %s in container %s", execConfig.ID, execConfig.Container.ID) |
bfebdfde |
container := execConfig.Container
|
e3d813f3 |
container.LogEvent("exec_start: " + execConfig.ProcessConfig.Entrypoint + " " + strings.Join(execConfig.ProcessConfig.Arguments, " "))
|
bfebdfde |
if execConfig.OpenStdin {
r, w := io.Pipe()
go func() {
defer w.Close() |
5572dbb7 |
defer log.Debugf("Closing buffered stdin pipe") |
bfebdfde |
io.Copy(w, job.Stdin)
}()
cStdin = r
}
if execConfig.OpenStdout {
cStdout = job.Stdout
}
if execConfig.OpenStderr {
cStderr = job.Stderr |
5130fe5d |
}
execConfig.StreamConfig.stderr = broadcastwriter.New()
execConfig.StreamConfig.stdout = broadcastwriter.New()
// Attach to stdin
if execConfig.OpenStdin {
execConfig.StreamConfig.stdin, execConfig.StreamConfig.stdinPipe = io.Pipe()
} else {
execConfig.StreamConfig.stdinPipe = ioutils.NopWriteCloser(ioutil.Discard) // Silently drop stdin
}
|
1095d5e5 |
attachErr := d.Attach(&execConfig.StreamConfig, execConfig.OpenStdin, true, execConfig.ProcessConfig.Tty, cStdin, cStdout, cStderr) |
5130fe5d |
|
d130c10a |
execErr := make(chan error) |
bfebdfde |
|
90928eb1 |
// Note, the execConfig data will be removed when the container
// itself is deleted. This allows us to query it (for things like
// the exitStatus) even after the cmd is done running. |
bfebdfde |
|
5130fe5d |
go func() {
err := container.Exec(execConfig)
if err != nil { |
bfebdfde |
execErr <- fmt.Errorf("Cannot run exec command %s in container %s: %s", execName, container.ID, err) |
5130fe5d |
}
}()
select {
case err := <-attachErr: |
d130c10a |
if err != nil {
return job.Errorf("attach failed with error: %s", err)
}
break |
5130fe5d |
case err := <-execErr:
return job.Error(err)
}
return engine.StatusOK
}
|
bfebdfde |
func (d *Daemon) Exec(c *Container, execConfig *execConfig, pipes *execdriver.Pipes, startCallback execdriver.StartCallback) (int, error) { |
90928eb1 |
exitStatus, err := d.execDriver.Exec(c.command, &execConfig.ProcessConfig, pipes, startCallback)
// On err, make sure we don't leave ExitCode at zero
if err != nil && exitStatus == 0 {
exitStatus = 128
}
execConfig.ExitCode = exitStatus
execConfig.Running = false
return exitStatus, err |
5130fe5d |
}
|
4b43a6df |
func (container *Container) GetExecIDs() []string {
return container.execCommands.List()
}
|
bfebdfde |
func (container *Container) Exec(execConfig *execConfig) error { |
5130fe5d |
container.Lock()
defer container.Unlock()
waitStart := make(chan struct{})
callback := func(processConfig *execdriver.ProcessConfig, pid int) {
if processConfig.Tty {
// The callback is called after the process Start() |
bfc9d8bb |
// so we are in the parent process. In TTY mode, stdin/out/err is the PtySlave |
5130fe5d |
// which we close here.
if c, ok := processConfig.Stdout.(io.Closer); ok {
c.Close()
}
}
close(waitStart)
}
// We use a callback here instead of a goroutine and an chan for
// syncronization purposes |
b845a621 |
cErr := promise.Go(func() error { return container.monitorExec(execConfig, callback) }) |
5130fe5d |
// Exec should not return until the process is actually running
select {
case <-waitStart:
case err := <-cErr:
return err
}
return nil
}
|
bfebdfde |
func (container *Container) monitorExec(execConfig *execConfig, callback execdriver.StartCallback) error { |
5130fe5d |
var (
err error
exitCode int
)
pipes := execdriver.NewPipes(execConfig.StreamConfig.stdin, execConfig.StreamConfig.stdout, execConfig.StreamConfig.stderr, execConfig.OpenStdin)
exitCode, err = container.daemon.Exec(container, execConfig, pipes, callback)
if err != nil {
log.Errorf("Error running command in existing container %s: %s", container.ID, err)
}
log.Debugf("Exec task in container %s exited with code %d", container.ID, exitCode)
if execConfig.OpenStdin {
if err := execConfig.StreamConfig.stdin.Close(); err != nil {
log.Errorf("Error closing stdin while running in %s: %s", container.ID, err)
}
}
if err := execConfig.StreamConfig.stdout.Clean(); err != nil {
log.Errorf("Error closing stdout while running in %s: %s", container.ID, err)
}
if err := execConfig.StreamConfig.stderr.Clean(); err != nil {
log.Errorf("Error closing stderr while running in %s: %s", container.ID, err)
}
if execConfig.ProcessConfig.Terminal != nil {
if err := execConfig.ProcessConfig.Terminal.Close(); err != nil {
log.Errorf("Error closing terminal while running in container %s: %s", container.ID, err)
}
}
return err
} |