blob: cc878a0c7a01ad5f2177e4fb5efe01da2360e2d1 [file] [log] [blame]
// Copyright 2015 The Vanadium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
// Command dmrun runs a binary on a remote VM instance using device manager.
//
// dmrun creates the VM instance, installs and starts device manager on it, and
// then installs and starts an app from the specified binary.
//
// dmrun uses the credentials it is running with in order to claim the device
// manager and provide the app with blessings. To specify credentials for
// dmrun, use the V23_CREDENTIALS environment variable instead of the
// --v23.credentials flag.
//
// Usage:
// dmrun [ENV=VAL ...] path/to/binary [--flag=val ...]
//
// All flags and environment variable settings are passed to the app.
//
// Usage examples:
// vcloud backend -- vcloud must be properly set up to talk to GCE:
// dmrun ${JIRI_ROOT}/release/go/bin/principal dump
//
// aws backend -- "aws" command must be installed and access+secret keys configured:
// dmrun --aws ${JIRI_ROOT}/release/go/bin/principal dump
//
// (For help setting up aws, see
// https://docs.google.com/document/d/1hRo6if4IF4EMKJNFu8HB1kbnk-D7UMe47iJuPH0CRAg/view )
//
// ssh backend can also be invoked directly to talk to an existing VM:
// (e.g. AWS ubuntu VM set up so you can ssh with "ssh -i /tmp/key" w/o typing a password.)
// dmrun --ssh "ubuntu@54.152.51.129" --sshoptions "-i /tmp/key" ${JIRI_ROOT}/release/go/bin/principal dump
//
package main
import (
"archive/zip"
"bytes"
"flag"
"fmt"
"io"
"io/ioutil"
"net"
"os"
"os/exec"
"os/user"
"path"
"path/filepath"
"regexp"
"strings"
"time"
"v.io/jiri"
"v.io/jiri/project"
"v.io/x/lib/cmdline"
"v.io/x/ref"
"v.io/x/ref/services/device/dmrun/backend"
)
var (
workDir string
vcloud string
device string
vm backend.CloudVM
cleanupOnDeath func()
)
var dmBins = [...]string{
"v.io/x/ref/services/device/deviced",
"v.io/x/ref/services/device/restarter",
"v.io/x/ref/services/agent/v23agentd",
"v.io/x/ref/services/device/inithelper",
"v.io/x/ref/services/device/suidhelper",
}
const (
vcloudBin = "v.io/x/devtools/vcloud"
deviceBin = "v.io/x/ref/services/device/device"
devicexRemote = "https://vanadium.googlesource.com/release.go.x.ref"
devicex = "services/device/devicex"
)
func die(format string, args ...interface{}) {
fmt.Fprintf(os.Stderr, format, args...)
fmt.Fprintln(os.Stderr)
if cleanupOnDeath != nil {
savedCleanupFn := cleanupOnDeath
cleanupOnDeath = func() {
fmt.Fprintf(os.Stderr, "Avoided recursive call to cleanup in die()\n")
}
savedCleanupFn()
}
os.Exit(1)
}
func dieIfErr(err error, format string, args ...interface{}) {
if err != nil {
fmt.Fprintf(os.Stderr, "Encountered error: %v\n", err)
die(format, args...)
}
}
// getPath returns the filesystem path to a file specified by a remote and a
// repository-relative path. If multiple projects have the same remote, the
// function will die with an error.
func getPath(remote, file string) string {
env := cmdline.EnvFromOS()
jirix, err := jiri.NewX(env)
dieIfErr(err, "jiri.NewX(%v) failed: %v", env, err)
projects, err := project.LocalProjects(jirix, project.FastScan)
dieIfErr(err, "project.LocalProjects(%v, %v) failed: %v", jirix, project.FastScan, err)
path := ""
for _, p := range projects {
if p.Remote == remote {
if path != "" {
die("Projects at paths %q and %q both have remote %s", path, p.Path, remote)
}
path = p.Path
}
}
if path == "" {
die("Couldn't find project with remote %s", remote)
}
return filepath.Join(path, filepath.FromSlash(file))
}
// setupWorkDir creates a directory for all the local files created by this
// tool.
func setupWorkDir() {
var err error
workDir, err = ioutil.TempDir("", filepath.Base(os.Args[0]))
dieIfErr(err, "Couldn't set up work dir")
dieIfErr(os.Chmod(workDir, 0777), "Couldn't chmod work dir")
fmt.Printf("Working dir: %s\n", workDir)
}
// buildV23Binaries builds the specified binaries and returns the paths to the
// executables.
func buildV23Binaries(pkg ...string) []string {
fmt.Print("Building ", pkg, " ...")
defer fmt.Println("Done.")
args := append([]string{"go", "install", "-x"}, pkg...)
cmd := exec.Command("jiri", args...)
cmd.Env = append(os.Environ(), "GOBIN="+workDir)
output, err := cmd.CombinedOutput()
dieIfErr(err, "Running build command %v failed. Output:\n%v", strings.Join(cmd.Args, " "), string(output))
dest := make([]string, len(pkg))
for i, p := range pkg {
dest[i] = filepath.Join(workDir, path.Base(p))
}
return dest
}
// createArchive creates a zip archive from the given files.
func createArchive(files []string) string {
zipFile := filepath.Join(workDir, "dm.zip")
z, err := os.OpenFile(zipFile, os.O_CREATE|os.O_WRONLY, os.FileMode(0644))
dieIfErr(err, "Couldn't create zip archive file")
defer z.Close()
w := zip.NewWriter(z)
for _, file := range files {
info, err := os.Stat(file)
dieIfErr(err, "Couldn't stat %v", file)
fh, err := zip.FileInfoHeader(info)
dieIfErr(err, "Couldn't set up file info header")
fh.Method = zip.Deflate
fwrite, err := w.CreateHeader(fh)
dieIfErr(err, "Couldn't create writer")
fread, err := os.Open(file)
dieIfErr(err, "Couldn't creater reader")
_, err = io.Copy(fwrite, fread)
dieIfErr(err, "Couldn't write to archive")
dieIfErr(fread.Close(), "Couldn't close reader")
}
dieIfErr(w.Close(), "Couldn't close zip archive")
return zipFile
}
// setupInstance creates a new VM instance and returns its name and IP address.
func setupInstance(vmOptions interface{}) (backend.CloudVM, string, string) {
fmt.Println("Setting up instance ...")
currUser, err := user.Current()
dieIfErr(err, "Couldn't obtain current user")
instanceName := fmt.Sprintf("%s-%s", currUser.Username, time.Now().UTC().Format("20060102-150405"))
vm, err = backend.CreateCloudVM(instanceName, vmOptions)
dieIfErr(err, "VM Instance Creation Failed: %v", err)
// Make sure nothing is using the ports we plan to give to deviced.
// TODO(caprita): Don't hardcode the ports and all that.
if output, err := vm.RunCommand("!", "netstat", "-tulpn", "2>/dev/null", "|", "grep", "'LISTEN'", "|", "grep", "-E", "':8150 |:8160 |:8151 '"); err != nil {
die("device manager ports are already in use:\n" + string(output))
}
instanceIP := vm.IP()
// Install unzip so we can unpack the archive.
// TODO(caprita): Use tar instead.
output, err := vm.RunCommand("sudo", "apt-get", "install", "unzip")
dieIfErr(err, "Installing unzip failed. Output:\n%v", string(output))
fmt.Println("Created VM instance", instanceName, "with IP", instanceIP)
return vm, instanceName, instanceIP
}
// installArchive ships the archive to the VM instance and unpacks it.
func installArchive(archive, instance string) {
err := vm.CopyFile(archive, "")
dieIfErr(err, "Copying archive failed: %v", err)
output, err := vm.RunCommand("unzip", path.Join("./", filepath.Base(archive)), "-d", "./unpacked")
dieIfErr(err, "Extracting archive failed. Output:\n%v", string(output))
}
// installDevice installs and starts device manager, and returns the public key
// and pairing token needed for claiming.
func installDevice(instance string) (string, string) {
fmt.Println("Installing device manager...")
defer fmt.Println("Done installing device manager.")
output, err := vm.RunCommand("V23_DEVICE_DIR=`pwd`/dm", "./unpacked/devicex", "install", "./unpacked", "--single_user", "--", "--v23.tcp.address=:8151", "--deviced-port=8150", "--use-pairing-token")
dieIfErr(err, "Installing device manager failed. Output:\n%v", string(output))
output, err = vm.RunCommand("V23_DEVICE_DIR=`pwd`/dm", "./unpacked/devicex", "start")
dieIfErr(err, "Starting device manager failed. Output:\n%v", string(output))
// Grab the token and public key from the device manager log.
dieAfter := time.After(5 * time.Second)
firstIteration := true
for {
if !firstIteration {
select {
case <-dieAfter:
die("Failed to find token and public key in log: %v", string(output))
case <-time.After(100 * time.Millisecond):
}
} else {
firstIteration = false
}
output, err = vm.RunCommand("cat", "./dm/dmroot/device-manager/logs/deviced.INFO")
dieIfErr(err, "Reading device manager log failed. Output:\n%v", string(output))
pairingTokenRE := regexp.MustCompile("Device manager pairing token: (.*)")
matches := pairingTokenRE.FindSubmatch(output)
if matches == nil {
continue
}
pairingToken := string(matches[1])
publicKeyRE := regexp.MustCompile("public_key: (.*)")
matches = publicKeyRE.FindSubmatch(output)
if matches == nil {
continue
}
publicKey := string(matches[1])
return publicKey, pairingToken
}
}
// setCredentialsEnv sets the command's environment to share the principal of
// dmrun.
func setCredentialsEnv(cmd *exec.Cmd) {
// TODO(caprita): This doesn't work with --v23.credentials.
if creds := os.Getenv(ref.EnvCredentials); len(creds) > 0 {
cmd.Env = append(cmd.Env, fmt.Sprintf("%s=%s", ref.EnvCredentials, creds))
} else if agentCreds := os.Getenv(ref.EnvAgentPath); len(agentCreds) > 0 {
cmd.Env = append(cmd.Env, fmt.Sprintf("%s=%s", ref.EnvAgentPath, agentCreds))
} else {
fmt.Fprintf(os.Stderr, "WARNING: no credentials found. You'll probably have authorization issues later on.\n")
}
}
// claimDevice claims the device manager, blessing it with extension.
func claimDevice(deviceName, ip, publicKey, pairingToken, extension string) {
fmt.Println("claiming device manager ...")
cmd := exec.Command(device, "claim", deviceName, extension, pairingToken, publicKey)
setCredentialsEnv(cmd)
output, err := cmd.CombinedOutput()
dieIfErr(err, "Claiming device manager (%v) failed. Output:\n%v", strings.Join(cmd.Args, " "), string(output))
cmd = exec.Command(device, "acl", "get", fmt.Sprintf("/%s:8150/device", ip))
setCredentialsEnv(cmd)
output, err = cmd.CombinedOutput()
dieIfErr(err, "Getting device manager acls (%v) failed. Output:\n%v", strings.Join(cmd.Args, " "), string(output))
fmt.Printf("Done claiming device manager. Device manager ACLs:\n%s", string(output))
}
// installApp installs the binary specified on the command-line and returns the
// Vanadium name for the installation object.
func installApp(deviceName, ip string) string {
args := []string{fmt.Sprintf("--v23.proxy=/%s:8160", ip), "install-local", deviceName + "/apps", "app"}
args = append(args, flag.Args()...)
cmd := exec.Command(device, args...)
setCredentialsEnv(cmd)
cmd.Env = append(cmd.Env, fmt.Sprintf("V23_NAMESPACE=/%s:8151", ip))
// During installation, there are sometimes timeout messages on stderr even when the
// installation succeeds -- so we need to capture stderr separately from stdout
buf := new(bytes.Buffer)
cmd.Stderr = buf
output, err := cmd.Output()
dieIfErr(err, "Installing app (%v) failed. Output:\n%v\nStderr:\n%v", strings.Join(cmd.Args, " "), string(output), buf.String())
installationName := strings.TrimSpace(string(output))
fmt.Println("Installed", installationName)
return installationName
}
// startApp creates and launches an instance of the given installation, blessing
// it with extension. It returns the Vanadium name for the instance object.
func startApp(installationName, extension string) string {
cmd := exec.Command(device, "instantiate", installationName, extension)
setCredentialsEnv(cmd)
output, err := cmd.CombinedOutput()
dieIfErr(err, "Instantiating app (%v) failed. Output:\n%v", strings.Join(cmd.Args, " "), string(output))
instanceName := strings.TrimSpace(string(output))
fmt.Println("Instantiated", instanceName)
cmd = exec.Command(device, "run", instanceName)
setCredentialsEnv(cmd)
output, err = cmd.CombinedOutput()
dieIfErr(err, "Starting app (%v) failed. Output:\n%v", strings.Join(cmd.Args, " "), string(output))
return instanceName
}
// check flags and produce the options used to build the backend VM
func handleFlags() (vmOpts interface{}, suppressCleanup bool) {
var keyDir string
flag.StringVar(&keyDir, "keydir", "/tmp", "where (on the local filesystem) to store ssh keys")
var debug bool
flag.BoolVar(&debug, "debug", false, "print debug messages")
flag.BoolVar(&suppressCleanup, "suppress-cleanup", false, "suppress cleanup on death to facilitate diagnosis")
// flags for ssh backend
var sshTarget, sshOptions string
flag.StringVar(&sshTarget, "ssh", "", "specify [user@]ip target for ssh")
flag.StringVar(&sshOptions, "sshoptions", "", "flags to pass to ssh, e.g. \"-i <keyfile>\"")
// flags for AWS backend
var aws bool
var awsRegion, awsImageID string
flag.BoolVar(&aws, "aws", false, "use AWS backend")
flag.StringVar(&awsRegion, "aws-region", "", "AWS region -- When specified, you also need to specify aws-image-id")
flag.StringVar(&awsImageID, "aws-image-id", "", "ID of AWS vm image to use")
flag.Parse()
if len(flag.Args()) == 0 {
die("Usage: %s [--ssh [user@]ip] [--sshoptions \"<options>\"] <app> <arguments ... >\n", os.Args[0])
}
binPath := flag.Args()[0]
if fi, err := os.Stat(binPath); err != nil {
die("failed to stat %v: %v", binPath, err)
} else if fi.IsDir() {
die("%v is a directory", binPath)
}
var dbg backend.DebugPrinter = backend.NoopDebugPrinter{}
if debug {
dbg = backend.StderrDebugPrinter{}
}
// Pick backend based on flags
switch {
default:
// Vcloud backend
vcloud = buildV23Binaries(vcloudBin)[0]
vmOpts = backend.VcloudVMOptions{VcloudBinary: vcloud}
case sshTarget != "":
// Ssh backend
opts := backend.SSHVMOptions{
Dbg: dbg,
}
targetComponents := strings.Split(sshTarget, "@")
switch len(targetComponents) {
case 1:
opts.SSHHostIP = targetComponents[0]
case 2:
opts.SSHUser = targetComponents[0]
opts.SSHHostIP = targetComponents[1]
default:
die("Unable to parse sshTarget: %s\n", sshTarget)
}
if sshOptions = strings.TrimSpace(sshOptions); sshOptions != "" {
opts.SSHOptions = strings.Split(sshOptions, " ")
}
vmOpts = opts
case aws:
vmOpts = backend.AWSVMOptions{
AWSBinary: "", // Assume "aws" command is on the path
Region: awsRegion,
ImageID: awsImageID,
KeyDir: keyDir,
Dbg: dbg,
}
}
return
}
func main() {
setupWorkDir()
cleanupOnDeath = func() {
os.RemoveAll(workDir)
}
defer os.RemoveAll(workDir)
vmOpts, suppressCleanup := handleFlags()
dmBinaries := buildV23Binaries(append([]string{deviceBin}, dmBins[:]...)...)
device, dmBinaries = dmBinaries[0], dmBinaries[1:]
archive := createArchive(append(dmBinaries, getPath(devicexRemote, devicex)))
vm, vmInstanceName, vmInstanceIP := setupInstance(vmOpts)
if suppressCleanup {
cleanupOnDeath = nil
} else {
cleanupOnDeath = func() {
fmt.Fprintf(os.Stderr, "Attempting to stop restarter/deviced ...\n")
vm.RunCommand("sudo", "killall", "-9", "restarter", "deviced") // errors are ignored
fmt.Fprintf(os.Stderr, "Cleaning up VM instance ...\n")
err := vm.Delete()
fmt.Fprintf(os.Stderr, "Removing local tmp files ...\n")
os.RemoveAll(workDir)
dieIfErr(err, "Cleaning up VM instance failed")
fmt.Fprintf(os.Stderr, "To keep device manager alive for debugging, run again with --suppress-cleanup.\n")
}
}
installArchive(archive, vmInstanceName)
publicKey, pairingToken := installDevice(vmInstanceName)
deviceAddr := net.JoinHostPort(vmInstanceIP, "8150")
deviceName := "/" + deviceAddr
claimDevice(deviceName, vmInstanceIP, publicKey, pairingToken, vmInstanceName)
installationName := installApp(deviceName, vmInstanceIP)
instanceName := startApp(installationName, "app")
fmt.Println("Launched app.")
fmt.Println("-------------")
fmt.Println("See its status:")
fmt.Printf("\t${JIRI_ROOT}/release/go/bin/device status %s\n", instanceName)
fmt.Println("See the logs:")
fmt.Printf("\t${JIRI_ROOT}/release/go/bin/debug glob %s/logs/*\n", instanceName)
fmt.Println("Dump e.g. the INFO log:")
fmt.Printf("\t${JIRI_ROOT}/release/go/bin/debug logs read %s/logs/app.INFO\n", instanceName)
fmt.Println("Clean up the VM instance:")
fmt.Printf("\t${JIRI_ROOT}/release/go/bin/device kill %s\n", instanceName)
fmt.Printf("\t%s\n", vm.RunCommandForUser("sudo", "killall", "-9", "restarter", "deviced"))
fmt.Printf("\t%s\n", vm.DeleteCommandForUser())
}