Merge "services/device/devicex: mkdir -p"
diff --git a/runtime/flow/model.go b/runtime/flow/model.go
new file mode 100644
index 0000000..df250b4
--- /dev/null
+++ b/runtime/flow/model.go
@@ -0,0 +1,38 @@
+// Copyright 2015 The Vanadium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+// Package Flow contains a super-interface of the Flow defined in v23/flow/model.go
+// to expose information that rpc implementors need to correctly implement the protocol
+// (i.e. VOM TypeEncoder, RPC BlessingCache).
+//
+// TODO(suharshs): We currently do not want to expose those in the flow API
+// because the flow API should be independent of the RPC code, so this is
+// our current way of punting on this.
+package flow
+
+import (
+ "v.io/v23/flow"
+)
+
+// Flow is the interface for a flow-controlled channel multiplexed over underlying network connections.
+type Flow interface {
+ flow.Flow
+
+ // SharedData returns a SharedData cache used for all flows on the underlying authenticated
+ // connection.
+ SharedData() SharedData
+}
+
+// SharedData is a thread-safe store that allows data to be shared across the underlying
+// authenticated connection that a flow is multiplexed over.
+type SharedData interface {
+ // Get returns the 'value' associated with 'key'.
+ Get(key interface{}) interface{}
+
+ // GetOrInsert returns the 'value' associated with 'key'. If an entry already exists in the
+ // cache with the 'key', the 'value' is returned, otherwise 'create' is called to create a new
+ // value N, the cache is updated, and N is returned. GetOrInsert may be called from
+ // multiple goroutines concurrently.
+ GetOrInsert(key interface{}, create func() interface{}) interface{}
+}
diff --git a/runtime/internal/flow/model.go b/runtime/internal/flow/model.go
deleted file mode 100644
index 53de3e1..0000000
--- a/runtime/internal/flow/model.go
+++ /dev/null
@@ -1,94 +0,0 @@
-// Copyright 2015 The Vanadium Authors. All rights reserved.
-// Use of this source code is governed by a BSD-style
-// license that can be found in the LICENSE file.
-
-package flow
-
-import (
- "io"
- "v.io/v23/context"
- "v.io/v23/naming"
- "v.io/v23/security"
-)
-
-// Manager is the interface for managing the creation of Flows.
-type Manager interface {
- // Acceptor creates a Acceptor that can be used to accept new flows from addresses
- // that the Manager is listening on.
- //
- // For example:
- // err := m.Listen(ctx, "tcp", ":0")
- // a, err := m.Acceptor(ctx, blessings)
- // for {
- // flow, err := a.Accept(ctx)
- // // process flow
- // }
- //
- // can be used to accept Flows initiated by remote processes to the endpoints returned
- // by Acceptor.Endpoints. All created Acceptors accept Flows from all addresses the
- // Manager is listening on, but are differentiated by a RoutingID unique to each
- // Acceptor.
- //
- // 'blessings' are the Blessings presented to the Client during authentication.
- Acceptor(ctx *context.T, blessings security.Blessings) (Acceptor, error)
-
- // Listen causes the Manager to accept flows from the provided protocol and address.
- // Listen may be called muliple times.
- Listen(ctx *context.T, protocol, address string) error
-
- // Dial creates a Flow to the provided remote endpoint. 'auth' is used to authorize
- // the remote end.
- //
- // To maximize re-use of connections, the Manager will also Listen on Dialed
- // connections for the lifetime of the connection.
- //
- // TODO(suharshs): Revisit passing in an authorizer here. Perhaps restrict server
- // authorization to a smaller set of policies, or use a different mechanism to
- // allow the user to specify a policy.
- Dial(ctx *context.T, remote naming.Endpoint, auth security.Authorizer) (Flow, error)
-
- // Closed returns a channel that remains open for the lifetime of the Manager
- // object. Once the channel is closed any operations on the Manager will
- // necessarily fail.
- Closed() <-chan struct{}
-}
-
-// Flow is the interface for a flow-controlled channel multiplexed on connection.
-type Flow interface {
- io.ReadWriter
-
- // LocalEndpoint returns the local vanadium Endpoint
- LocalEndpoint() naming.Endpoint
- // RemoteEndpoint returns the remote vanadium Endpoint
- RemoteEndpoint() naming.Endpoint
- // LocalBlessings returns the blessings presented by the local end of the flow during authentication.
- LocalBlessings() security.Blessings
- // RemoteBlessings returns the blessings presented by the remote end of the flow during authentication.
- RemoteBlessings() security.Blessings
- // LocalDischarges returns the discharges presented by the local end of the flow during authentication.
- //
- // Discharges are organized in a map keyed by the discharge-identifier.
- LocalDischarges() map[string]security.Discharge
- // RemoteDischarges returns the discharges presented by the remote end of the flow during authentication.
- //
- // Discharges are organized in a map keyed by the discharge-identifier.
- RemoteDischarges() map[string]security.Discharge
-
- // Closed returns a channel that remains open until the flow has been closed or
- // the ctx to the Dial or Accept call used to create the flow has been cancelled.
- Closed() <-chan struct{}
-}
-
-// Acceptor is the interface for accepting Flows created by a remote process.
-type Acceptor interface {
- // ListeningEndpoints returns the endpoints that the Manager has explicitly
- // listened on. The Acceptor will accept new flows on these endpoints.
- // Returned endpoints all have a RoutingID unique to the Acceptor.
- ListeningEndpoints() []naming.Endpoint
- // Accept blocks until a new Flow has been initiated by a remote process.
- Accept(ctx *context.T) (Flow, error)
- // Closed returns a channel that remains open until the Acceptor has been closed.
- // i.e. the context provided to Manager.Acceptor() has been cancelled. Once the
- // channel is closed, all calls to Accept will result in an error.
- Closed() <-chan struct{}
-}
diff --git a/services/application/applicationd/impl_test.go b/services/application/applicationd/impl_test.go
index 68b07f9..b045bbf 100644
--- a/services/application/applicationd/impl_test.go
+++ b/services/application/applicationd/impl_test.go
@@ -5,6 +5,7 @@
package main_test
import (
+ "fmt"
"io/ioutil"
"os"
"reflect"
@@ -303,3 +304,238 @@
t.Fatalf("Incorrect output: expected %v, got %v", envelopeV1, output)
}
}
+
+// TestTidyNow tests that TidyNow operates correctly.
+func TestTidyNow(t *testing.T) {
+ ctx, shutdown := test.V23Init()
+ defer shutdown()
+
+ dir, prefix := "", ""
+ store, err := ioutil.TempDir(dir, prefix)
+ if err != nil {
+ t.Fatalf("TempDir(%q, %q) failed: %v", dir, prefix, err)
+ }
+ defer os.RemoveAll(store)
+ dispatcher, err := appd.NewDispatcher(store)
+ if err != nil {
+ t.Fatalf("NewDispatcher() failed: %v", err)
+ }
+
+ server, err := xrpc.NewDispatchingServer(ctx, "", dispatcher)
+ if err != nil {
+ t.Fatalf("NewServer(%v) failed: %v", dispatcher, err)
+ }
+
+ defer func() {
+ if err := server.Stop(); err != nil {
+ t.Fatalf("Stop() failed: %v", err)
+ }
+ }()
+ endpoint := server.Status().Endpoints[0].String()
+
+ // Create client stubs for talking to the server.
+ stub := repository.ApplicationClient(naming.JoinAddressName(endpoint, "search"))
+ stubs := make([]repository.ApplicationClientStub, 0)
+ for _, vn := range []string{"v0", "v1", "v2", "v3"} {
+ s := repository.ApplicationClient(naming.JoinAddressName(endpoint, fmt.Sprintf("search/%s", vn)))
+ stubs = append(stubs, s)
+ }
+ blessings, sig := newPublisherSignature(t, ctx, []byte("binarycontents"))
+
+ // Create example envelopes.
+ envelopeV1 := application.Envelope{
+ Args: []string{"--help"},
+ Env: []string{"DEBUG=1"},
+ Binary: application.SignedFile{
+ File: "/v23/name/of/binary",
+ Signature: sig,
+ },
+ Publisher: blessings,
+ }
+ envelopeV2 := application.Envelope{
+ Args: []string{"--verbose"},
+ Env: []string{"DEBUG=0"},
+ Binary: application.SignedFile{
+ File: "/v23/name/of/binary",
+ Signature: sig,
+ },
+ Publisher: blessings,
+ }
+ envelopeV3 := application.Envelope{
+ Args: []string{"--verbose", "--spiffynewflag"},
+ Env: []string{"DEBUG=0"},
+ Binary: application.SignedFile{
+ File: "/v23/name/of/binary",
+ Signature: sig,
+ },
+ Publisher: blessings,
+ }
+
+ stuffEnvelopes(t, ctx, stubs, []profEnvTuple{
+ {
+ &envelopeV1,
+ []string{"base", "media"},
+ },
+ })
+
+ // Verify that we have one
+ testGlob(t, ctx, endpoint, []string{
+ "",
+ "search",
+ "search/v0",
+ })
+
+ // Tidy when already tidy does not alter.
+ if err := stubs[0].TidyNow(ctx); err != nil {
+ t.Errorf("TidyNow failed: %v", err)
+ }
+ testGlob(t, ctx, endpoint, []string{
+ "",
+ "search",
+ "search/v0",
+ })
+
+ stuffEnvelopes(t, ctx, stubs, []profEnvTuple{
+ {
+ &envelopeV1,
+ []string{"base", "media"},
+ },
+ {
+ &envelopeV2,
+ []string{"media"},
+ },
+ {
+ &envelopeV3,
+ []string{"base"},
+ },
+ })
+
+ // Now there are three envelopes which is one more than the
+ // numberOfVersionsToKeep set for the test. However
+ // we need both envelopes v0 and v2 to keep two versions for
+ // profile media and envelopes v0 and v3 to keep two versions
+ // for profile base so we continue to have three versions.
+ if err := stubs[0].TidyNow(ctx); err != nil {
+ t.Errorf("TidyNow failed: %v", err)
+ }
+ testGlob(t, ctx, endpoint, []string{
+ "",
+ "search",
+ "search/v0",
+ "search/v1",
+ "search/v2",
+ })
+
+ // And the newest version for each profile differs because
+ // not every version supports all profiles.
+ output1, err := stub.Match(ctx, []string{"media"})
+ if err != nil {
+ t.Fatalf("Match(%v) failed: %v", "media", err)
+ }
+ if !reflect.DeepEqual(envelopeV2, output1) {
+ t.Fatalf("Incorrect output: expected %v, got %v", envelopeV2, output1)
+ }
+
+ output2, err := stub.Match(ctx, []string{"base"})
+ if err != nil {
+ t.Fatalf("Match(%v) failed: %v", "base", err)
+ }
+ if !reflect.DeepEqual(envelopeV3, output2) {
+ t.Fatalf("Incorrect output: expected %v, got %v", envelopeV3, output2)
+ }
+
+ // Test that we can add an envelope for v3 with profile media and after calling
+ // TidyNow(), there will be all versions still in glob but v0 will only match profile
+ // base and not have an envelope for profile media.
+ if err := stubs[3].Put(ctx, []string{"media"}, envelopeV3); err != nil {
+ t.Fatalf("Put() failed: %v", err)
+ }
+
+ if err := stubs[0].TidyNow(ctx); err != nil {
+ t.Errorf("TidyNow failed: %v", err)
+ }
+ testGlob(t, ctx, endpoint, []string{
+ "",
+ "search",
+ "search/v0",
+ "search/v1",
+ "search/v2",
+ "search/v3",
+ })
+
+ output3, err := stubs[0].Match(ctx, []string{"base"})
+ if err != nil {
+ t.Fatalf("Match(%v) failed: %v", "base", err)
+ }
+ if !reflect.DeepEqual(envelopeV3, output2) {
+ t.Fatalf("Incorrect output: expected %v, got %v", envelopeV3, output3)
+ }
+
+ output4, err := stubs[0].Match(ctx, []string{"base"})
+ if err != nil {
+ t.Fatalf("Match(%v) failed: %v", "base", err)
+ }
+ if !reflect.DeepEqual(envelopeV3, output2) {
+ t.Fatalf("Incorrect output: expected %v, got %v", envelopeV3, output4)
+ }
+
+ _, err = stubs[0].Match(ctx, []string{"media"})
+ if verror.ErrorID(err) != verror.ErrNoExist.ID {
+ t.Fatalf("got error %v, expected %v", err, verror.ErrNoExist)
+ }
+
+ stuffEnvelopes(t, ctx, stubs, []profEnvTuple{
+ {
+ &envelopeV1,
+ []string{"base", "media"},
+ },
+ {
+ &envelopeV2,
+ []string{"base", "media"},
+ },
+ {
+ &envelopeV3,
+ []string{"base", "media"},
+ },
+ {
+ &envelopeV3,
+ []string{"base", "media"},
+ },
+ })
+
+ // Now there are four versions for all profiles so tidying
+ // will remove the older versions.
+ if err := stubs[0].TidyNow(ctx); err != nil {
+ t.Errorf("TidyNow failed: %v", err)
+ }
+
+ testGlob(t, ctx, endpoint, []string{
+ "",
+ "search",
+ "search/v2",
+ "search/v3",
+ })
+}
+
+type profEnvTuple struct {
+ e *application.Envelope
+ p []string
+}
+
+func testGlob(t *testing.T, ctx *context.T, endpoint string, expected []string) {
+ matches, _, err := testutil.GlobName(ctx, naming.JoinAddressName(endpoint, ""), "...")
+ if err != nil {
+ t.Errorf("Unexpected Glob error: %v", err)
+ }
+ if !reflect.DeepEqual(matches, expected) {
+ t.Errorf("unexpected Glob results. Got %q, want %q", matches, expected)
+ }
+}
+
+func stuffEnvelopes(t *testing.T, ctx *context.T, stubs []repository.ApplicationClientStub, pets []profEnvTuple) {
+ for i, pet := range pets {
+ if err := stubs[i].Put(ctx, pet.p, *pet.e); err != nil {
+ t.Fatalf("%d: Put(%v) failed: %v", i, pet, err)
+ }
+ }
+}
diff --git a/services/application/applicationd/only_for_test.go b/services/application/applicationd/only_for_test.go
new file mode 100644
index 0000000..0b3724e
--- /dev/null
+++ b/services/application/applicationd/only_for_test.go
@@ -0,0 +1,9 @@
+// Copyright 2015 The Vanadium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+package main
+
+func init() {
+ numberOfVersionsToKeep = 2
+}
diff --git a/services/application/applicationd/service.go b/services/application/applicationd/service.go
index 2fc2a5f..e11f26d 100644
--- a/services/application/applicationd/service.go
+++ b/services/application/applicationd/service.go
@@ -5,7 +5,6 @@
package main
import (
- "fmt"
"sort"
"strings"
@@ -72,7 +71,7 @@
defer i.store.Unlock()
if version == "" {
- versions, err := i.allAppVersions(name)
+ versions, err := i.allAppVersionsForProfiles(name, profiles)
if err != nil {
return empty, err
}
@@ -187,15 +186,13 @@
return apps, nil
}
-func (i *appRepoService) allAppVersions(appName string) ([]string, error) {
- profiles, err := i.store.BindObject(naming.Join("/applications", appName)).Children()
- if err != nil {
- return nil, err
- }
+func (i *appRepoService) allAppVersionsForProfiles(appName string, profiles []string) ([]string, error) {
uniqueVersions := make(map[string]struct{})
for _, profile := range profiles {
versions, err := i.store.BindObject(naming.Join("/applications", appName, profile)).Children()
- if err != nil {
+ if verror.ErrorID(err) == verror.ErrNoExist.ID {
+ continue
+ } else if err != nil {
return nil, err
}
set.String.Union(uniqueVersions, set.String.FromSlice(versions))
@@ -203,6 +200,14 @@
return set.String.ToSlice(uniqueVersions), nil
}
+func (i *appRepoService) allAppVersions(appName string) ([]string, error) {
+ profiles, err := i.store.BindObject(naming.Join("/applications", appName)).Children()
+ if err != nil {
+ return nil, err
+ }
+ return i.allAppVersionsForProfiles(appName, profiles)
+}
+
func (i *appRepoService) GlobChildren__(ctx *context.T, _ rpc.ServerCall) (<-chan string, error) {
ctx.VI(0).Infof("%v.GlobChildren__()", i.suffix)
i.store.Lock()
@@ -334,6 +339,64 @@
return nil
}
+func (i *appRepoService) tidyRemoveVersions(call rpc.ServerCall, tname, appName, profile string, versions []string) error {
+ for _, v := range versions {
+ path := naming.Join(tname, "/applications", appName, profile, v)
+ object := i.store.BindObject(path)
+ if err := object.Remove(call); err != nil {
+ return err
+ }
+ }
+ return nil
+}
+
+// numberOfVersionsToKeep can be set for tests.
+var numberOfVersionsToKeep = 5
+
func (i *appRepoService) TidyNow(ctx *context.T, call rpc.ServerCall) error {
- return fmt.Errorf("method not implemented")
+ ctx.VI(2).Infof("%v.TidyNow()", i.suffix)
+ i.store.Lock()
+ defer i.store.Unlock()
+
+ tname, err := i.store.BindTransactionRoot("").CreateTransaction(call)
+ if err != nil {
+ return err
+ }
+
+ apps, err := i.allApplications()
+ if err != nil {
+ return err
+ }
+
+ for _, app := range apps {
+ profiles, err := i.store.BindObject(naming.Join("/applications", app)).Children()
+ if err != nil {
+ return err
+ }
+
+ for _, profile := range profiles {
+ versions, err := i.store.BindObject(naming.Join("/applications", app, profile)).Children()
+ if err != nil {
+ return err
+ }
+
+ lv := len(versions)
+ if lv <= numberOfVersionsToKeep {
+ continue
+ }
+
+ // Per assumption in Match, version names should ascend.
+ sort.Strings(versions)
+ versionsToRemove := versions[0 : lv-numberOfVersionsToKeep]
+ if err := i.tidyRemoveVersions(call, tname, app, profile, versionsToRemove); err != nil {
+ return err
+ }
+ }
+ }
+
+ if err := i.store.BindTransaction(tname).Commit(call); err != nil {
+ return verror.New(ErrOperationFailed, ctx)
+ }
+ return nil
+
}
diff --git a/services/device/deviced/internal/impl/dispatcher.go b/services/device/deviced/internal/impl/dispatcher.go
index 5cd449e..d1824ea 100644
--- a/services/device/deviced/internal/impl/dispatcher.go
+++ b/services/device/deviced/internal/impl/dispatcher.go
@@ -40,8 +40,8 @@
securityAgent *securityAgentState
restartHandler func()
testMode bool
- // reap is the app process monitoring subsystem.
- reap *reaper
+ // runner is responsible for running app instances.
+ runner *appRunner
// tidying is the automatic state tidying subsystem.
tidying chan<- tidyRequests
}
@@ -107,14 +107,16 @@
return &claimable{token: pairingToken, permsStore: permsStore, permsDir: permsDir, notify: notify}, notify
}
-// NewDispatcher is the device manager dispatcher factory.
-func NewDispatcher(ctx *context.T, config *config.State, mtAddress string, testMode bool, restartHandler func(), permStore *pathperms.PathStore) (rpc.Dispatcher, error) {
+// NewDispatcher is the device manager dispatcher factory. It returns a new
+// dispatcher as well as a shutdown function, to be called when the dispatcher
+// is no longer needed.
+func NewDispatcher(ctx *context.T, config *config.State, mtAddress string, testMode bool, restartHandler func(), permStore *pathperms.PathStore) (rpc.Dispatcher, func(), error) {
if err := config.Validate(); err != nil {
- return nil, verror.New(errInvalidConfig, ctx, config, err)
+ return nil, nil, verror.New(errInvalidConfig, ctx, config, err)
}
uat, err := NewBlessingSystemAssociationStore(config.Root)
if err != nil {
- return nil, verror.New(errCantCreateAccountStore, ctx, err)
+ return nil, nil, verror.New(errCantCreateAccountStore, ctx, err)
}
InitSuidHelper(ctx, config.Helper)
d := &dispatcher{
@@ -135,39 +137,30 @@
// If we're in 'security agent mode', set up the key manager agent.
if len(os.Getenv(ref.EnvAgentEndpoint)) > 0 {
if keyMgrAgent, err := keymgr.NewAgent(); err != nil {
- return nil, verror.New(errNewAgentFailed, ctx, err)
+ return nil, nil, verror.New(errNewAgentFailed, ctx, err)
} else {
d.internal.securityAgent = &securityAgentState{
keyMgrAgent: keyMgrAgent,
}
}
}
- reap, err := newReaper(ctx, config.Root, &appRunner{
+ runner := &appRunner{
callback: d.internal.callback,
securityAgent: d.internal.securityAgent,
appServiceName: naming.Join(d.config.Name, appsSuffix),
- })
- if err != nil {
- return nil, verror.New(errCantCreateAppWatcher, ctx, err)
+ mtAddress: d.mtAddress,
}
- d.internal.reap = reap
+ d.internal.runner = runner
+ reap, err := newReaper(ctx, config.Root, runner)
+ if err != nil {
+ return nil, nil, verror.New(errCantCreateAppWatcher, ctx, err)
+ }
+ runner.reap = reap
if testMode {
- return &testModeDispatcher{d}, nil
+ return &testModeDispatcher{d}, reap.shutdown, nil
}
- return d, nil
-}
-
-// Shutdown the dispatcher.
-func Shutdown(ctx *context.T, rpcd rpc.Dispatcher) {
- switch d := rpcd.(type) {
- case *dispatcher:
- d.internal.reap.shutdown()
- case *testModeDispatcher:
- Shutdown(ctx, d.realDispatcher)
- default:
- ctx.Panicf("%v not a supported dispatcher type.", rpcd)
- }
+ return d, reap.shutdown, nil
}
// Logging invoker that logs any error messages before returning.
@@ -333,13 +326,7 @@
suffix: components[1:],
uat: d.uat,
permsStore: d.permsStore,
- runner: &appRunner{
- reap: d.internal.reap,
- callback: d.internal.callback,
- securityAgent: d.internal.securityAgent,
- mtAddress: d.mtAddress,
- appServiceName: naming.Join(d.config.Name, appsSuffix),
- },
+ runner: d.internal.runner,
})
appSpecificAuthorizer, err := newAppSpecificAuthorizer(auth, d.config, components[1:], d.permsStore)
if err != nil {
diff --git a/services/device/deviced/internal/impl/instance_reaping.go b/services/device/deviced/internal/impl/instance_reaping.go
index 320f097..9b2f4ce 100644
--- a/services/device/deviced/internal/impl/instance_reaping.go
+++ b/services/device/deviced/internal/impl/instance_reaping.go
@@ -43,14 +43,13 @@
}
type reaper struct {
- c chan pidInstanceDirPair
- startState *appRunner
- ctx *context.T
+ c chan pidInstanceDirPair
+ ctx *context.T
}
var stashedPidMap map[string]int
-func newReaper(ctx *context.T, root string, startState *appRunner) (*reaper, error) {
+func newReaper(ctx *context.T, root string, appRunner *appRunner) (*reaper, error) {
pidMap, restartCandidates, err := findAllTheInstances(ctx, root)
// Used only by the testing code that verifies that all processes
@@ -61,16 +60,14 @@
}
r := &reaper{
- c: make(chan pidInstanceDirPair),
- startState: startState,
- ctx: ctx,
+ c: make(chan pidInstanceDirPair),
+ ctx: ctx,
}
- r.startState.reap = r
- go r.processStatusPolling(ctx, pidMap)
+ go r.processStatusPolling(ctx, pidMap, appRunner)
// Restart daemon jobs if they're not running (say because the machine crashed.)
for _, idir := range restartCandidates {
- go r.startState.restartAppIfNecessary(ctx, idir)
+ go appRunner.restartAppIfNecessary(ctx, idir)
}
return r, nil
}
@@ -92,16 +89,16 @@
// functionality. For example, use the kevent facility in darwin or
// replace init. See http://www.incenp.org/dvlpt/wait4.html for
// inspiration.
-func (r *reaper) processStatusPolling(ctx *context.T, trackedPids map[string]int) {
+func (r *reaper) processStatusPolling(ctx *context.T, trackedPids map[string]int, appRunner *appRunner) {
poll := func(ctx *context.T) {
for idir, pid := range trackedPids {
switch err := syscall.Kill(pid, 0); err {
case syscall.ESRCH:
// No such PID.
- go r.startState.restartAppIfNecessary(ctx, idir)
+ go appRunner.restartAppIfNecessary(ctx, idir)
ctx.VI(2).Infof("processStatusPolling discovered pid %d ended", pid)
markNotRunning(ctx, idir)
- go r.startState.restartAppIfNecessary(ctx, idir)
+ go appRunner.restartAppIfNecessary(ctx, idir)
delete(trackedPids, idir)
case nil, syscall.EPERM:
ctx.VI(2).Infof("processStatusPolling saw live pid: %d", pid)
diff --git a/services/device/deviced/internal/starter/starter.go b/services/device/deviced/internal/starter/starter.go
index 37b7dc3..848eccc 100644
--- a/services/device/deviced/internal/starter/starter.go
+++ b/services/device/deviced/internal/starter/starter.go
@@ -330,7 +330,7 @@
}
args.ConfigState.Name = endpoints[0].Name()
- dispatcher, err := impl.NewDispatcher(ctx, args.ConfigState, mt, args.TestMode, args.RestartCallback, permStore)
+ dispatcher, dShutdown, err := impl.NewDispatcher(ctx, args.ConfigState, mt, args.TestMode, args.RestartCallback, permStore)
if err != nil {
shutdown()
return nil, err
@@ -341,7 +341,7 @@
// the dispatcher and exposing it in Status.
ctx.Infof("Stopping device server...")
server.Stop()
- impl.Shutdown(ctx, dispatcher)
+ dShutdown()
ctx.Infof("Stopped device.")
}
if err := server.ServeDispatcher(args.name(mt), dispatcher); err != nil {
diff --git a/services/groups/groupsd/groupsd_v23_test.go b/services/groups/groupsd/groupsd_v23_test.go
index e304cf4..9334799 100644
--- a/services/groups/groupsd/groupsd_v23_test.go
+++ b/services/groups/groupsd/groupsd_v23_test.go
@@ -67,11 +67,11 @@
// Test simple group resolution.
{
- var buffer bytes.Buffer
- clientBin.Start("relate", groupB, "a/b/c/d").WaitOrDie(&buffer, &buffer)
+ var buffer, stderrBuf bytes.Buffer
+ clientBin.Start("relate", groupB, "a/b/c/d").WaitOrDie(&buffer, &stderrBuf)
var got relateResult
if err := json.Unmarshal(buffer.Bytes(), &got); err != nil {
- t.Fatalf("Unmarshal(%v) failed: %v", buffer.Bytes(), err)
+ t.Fatalf("Unmarshal(%v) failed: %v", buffer.String(), err)
}
want := relateResult{
Remainder: set.String.FromSlice([]string{"c/d", "d"}),
@@ -81,15 +81,18 @@
if !reflect.DeepEqual(got, want) {
t.Errorf("got %v, want %v", got, want)
}
+ if got, want := stderrBuf.Len(), 0; got != want {
+ t.Errorf("got %v, want %v", got, want)
+ }
}
// Test recursive group resolution.
{
- var buffer bytes.Buffer
- clientBin.Start("relate", groupA, "a/b/c/d").WaitOrDie(&buffer, &buffer)
+ var buffer, stderrBuf bytes.Buffer
+ clientBin.Start("relate", groupA, "a/b/c/d").WaitOrDie(&buffer, &stderrBuf)
var got relateResult
if err := json.Unmarshal(buffer.Bytes(), &got); err != nil {
- t.Fatalf("Unmarshal(%v) failed: %v", buffer.Bytes(), err)
+ t.Fatalf("Unmarshal(%v) failed: %v", buffer.String(), err)
}
want := relateResult{
Remainder: set.String.FromSlice([]string{"b/c/d", "c/d", "d"}),
@@ -99,17 +102,20 @@
if !reflect.DeepEqual(got, want) {
t.Errorf("got %v, want %v", got, want)
}
+ if got, want := stderrBuf.Len(), 0; got != want {
+ t.Errorf("got %v, want %v", got, want)
+ }
}
// Test group resolution failure. Note that under-approximation is
// used as the default to handle resolution failures.
{
clientBin.Start("add", groupB, "<grp:groups-server/groupC>").WaitOrDie(os.Stdout, os.Stderr)
- var buffer bytes.Buffer
- clientBin.Start("relate", groupB, "a/b/c/d").WaitOrDie(&buffer, &buffer)
+ var buffer, stderrBuf bytes.Buffer
+ clientBin.Start("relate", groupB, "a/b/c/d").WaitOrDie(&buffer, &stderrBuf)
var got relateResult
if err := json.Unmarshal(buffer.Bytes(), &got); err != nil {
- t.Fatalf("Unmarshal(%v) failed: %v", buffer.Bytes(), err)
+ t.Fatalf("Unmarshal(%v) failed: %v", buffer.String(), err)
}
want := relateResult{
Remainder: set.String.FromSlice([]string{"c/d", "d"}),
@@ -124,6 +130,9 @@
if !reflect.DeepEqual(got, want) {
t.Errorf("got %v, want %v", got, want)
}
+ if got, want := stderrBuf.Len(), 0; got != want {
+ t.Errorf("got %v, want %v", got, want)
+ }
}
// Delete the groups.
diff --git a/services/groups/internal/store/leveldb/store.go b/services/groups/internal/store/leveldb/store.go
index 5f38f52..23d0733 100644
--- a/services/groups/internal/store/leveldb/store.go
+++ b/services/groups/internal/store/leveldb/store.go
@@ -31,7 +31,7 @@
// Open opens a groups server store located at the given path,
// creating it if it doesn't exist.
func Open(path string) (store.Store, error) {
- db, err := leveldb.Open(path)
+ db, err := leveldb.Open(path, leveldb.OpenOptions{CreateIfMissing: true, ErrorIfExists: false})
if err != nil {
return nil, convertError(err)
}
diff --git a/services/wspr/browsprd/main_nacl.go b/services/wspr/browsprd/main_nacl.go
index 7e0b2bd..08940f7 100644
--- a/services/wspr/browsprd/main_nacl.go
+++ b/services/wspr/browsprd/main_nacl.go
@@ -26,6 +26,7 @@
"v.io/x/ref/services/wspr/internal/app"
"v.io/x/ref/services/wspr/internal/browspr"
"v.io/x/ref/services/wspr/internal/channel/channel_nacl"
+ "v.io/x/ref/services/wspr/internal/principal"
"v.io/x/ref/services/wspr/internal/rpc/server"
)
@@ -138,11 +139,11 @@
}
func (inst *browsprInstance) newPrincipal(ecdsaKey *ecdsa.PrivateKey, blessingRootsData, blessingRootsSig, blessingStoreData, blessingStoreSig string) (security.Principal, error) {
- roots, err := browspr.NewFileSerializer(blessingRootsData, blessingRootsSig, inst.fs)
+ roots, err := principal.NewFileSerializer(blessingRootsData, blessingRootsSig, inst.fs)
if err != nil {
return nil, fmt.Errorf("failed to create blessing roots serializer:%s", err)
}
- store, err := browspr.NewFileSerializer(blessingStoreData, blessingStoreSig, inst.fs)
+ store, err := principal.NewFileSerializer(blessingStoreData, blessingStoreSig, inst.fs)
if err != nil {
return nil, fmt.Errorf("failed to create blessing store serializer:%s", err)
}
@@ -193,15 +194,15 @@
return nil, fmt.Errorf("HandleStartMessage did not receive StartMessage, received: %v, %v", val, err)
}
- principal, err := inst.newPersistantPrincipal(msg.IdentitydBlessingRoot.Names)
+ p, err := inst.newPersistantPrincipal(msg.IdentitydBlessingRoot.Names)
if err != nil {
return nil, err
}
blessingName := "browspr-default-blessing"
- blessing, err := principal.BlessSelf(blessingName)
+ blessing, err := p.BlessSelf(blessingName)
if err != nil {
- return nil, fmt.Errorf("principal.BlessSelf(%v) failed: %v", blessingName, err)
+ return nil, fmt.Errorf("p.BlessSelf(%v) failed: %v", blessingName, err)
}
// If msg.IdentitydBlessingRoot has a public key and names, then add
@@ -222,18 +223,18 @@
pattern := security.BlessingPattern(name)
// Trust the identity servers blessing root.
- principal.Roots().Add(key, pattern)
+ p.Roots().Add(key, pattern)
// Use our blessing to only talk to the identity server.
- if _, err := principal.BlessingStore().Set(blessing, pattern); err != nil {
- return nil, fmt.Errorf("principal.BlessingStore().Set(%v, %v) failed: %v", blessing, pattern, err)
+ if _, err := p.BlessingStore().Set(blessing, pattern); err != nil {
+ return nil, fmt.Errorf("p.BlessingStore().Set(%v, %v) failed: %v", blessing, pattern, err)
}
}
} else {
inst.logger.VI(1).Infof("IdentitydBlessingRoot.PublicKey is empty. Will allow browspr blessing to be shareable with all principals.")
// Set our blessing as shareable with all peers.
- if _, err := principal.BlessingStore().Set(blessing, security.AllPrincipals); err != nil {
- return nil, fmt.Errorf("principal.BlessingStore().Set(%v, %v) failed: %v", blessing, security.AllPrincipals, err)
+ if _, err := p.BlessingStore().Set(blessing, security.AllPrincipals); err != nil {
+ return nil, fmt.Errorf("p.BlessingStore().Set(%v, %v) failed: %v", blessing, security.AllPrincipals, err)
}
}
@@ -241,7 +242,7 @@
// TODO(suharshs,mattr): Should we worried about not shutting down here?
ctx, _ := v23.Init()
- ctx, err = v23.WithPrincipal(ctx, principal)
+ ctx, err = v23.WithPrincipal(ctx, p)
if err != nil {
return nil, err
}
@@ -265,12 +266,18 @@
listenSpec := v23.GetListenSpec(ctx)
listenSpec.Proxy = msg.Proxy
+ principalSerializer, err := principal.NewFileSerializer(browsprDir+"/principalData", browsprDir+"/principalSignature", inst.fs)
+ if err != nil {
+ return nil, fmt.Errorf("principal.NewFileSerializer() failed: %v", err)
+ }
+
inst.logger.VI(1).Infof("Starting browspr with config: proxy=%q mounttable=%q identityd=%q identitydBlessingRoot=%q ", msg.Proxy, msg.NamespaceRoot, msg.Identityd, msg.IdentitydBlessingRoot)
inst.browspr = browspr.NewBrowspr(ctx,
inst.BrowsprOutgoingPostMessage,
&listenSpec,
msg.Identityd,
- []string{msg.NamespaceRoot})
+ []string{msg.NamespaceRoot},
+ principalSerializer)
// Add the rpc handlers that depend on inst.browspr.
inst.channel.RegisterRequestHandler("auth:create-account", inst.browspr.HandleAuthCreateAccountRpc)
diff --git a/services/wspr/internal/account/account.go b/services/wspr/internal/account/account.go
index 7d977d6..213404e 100644
--- a/services/wspr/internal/account/account.go
+++ b/services/wspr/internal/account/account.go
@@ -47,7 +47,6 @@
ctx *context.T
blesser BlesserService
principalManager *principal.PrincipalManager
- accounts []string
}
func NewAccountManager(identitydEndpoint string, principalManager *principal.PrincipalManager) *AccountManager {
@@ -72,13 +71,11 @@
return "", fmt.Errorf("Error adding account: %v", err)
}
- am.accounts = append(am.accounts, account)
-
return account, nil
}
func (am *AccountManager) GetAccounts() []string {
- return am.accounts
+ return am.principalManager.GetAccounts()
}
func (am *AccountManager) AssociateAccount(ctx *context.T, origin, account string, cavs []Caveat) error {
diff --git a/services/wspr/internal/browspr/browspr.go b/services/wspr/internal/browspr/browspr.go
index 409ea62..e0cc5ec 100644
--- a/services/wspr/internal/browspr/browspr.go
+++ b/services/wspr/internal/browspr/browspr.go
@@ -15,6 +15,7 @@
"v.io/v23/rpc"
"v.io/v23/vdl"
"v.io/v23/vtrace"
+ "v.io/x/ref/lib/security"
"v.io/x/ref/services/wspr/internal/account"
"v.io/x/ref/services/wspr/internal/app"
"v.io/x/ref/services/wspr/internal/principal"
@@ -39,7 +40,8 @@
postMessage func(instanceId int32, ty, msg string),
listenSpec *rpc.ListenSpec,
identd string,
- wsNamespaceRoots []string) *Browspr {
+ wsNamespaceRoots []string,
+ principalSerializer security.SerializerReaderWriter) *Browspr {
if listenSpec.Proxy == "" {
ctx.Fatalf("a vanadium proxy must be set")
}
@@ -58,7 +60,12 @@
// TODO(nlacasse, bjornick) use a serializer that can actually persist.
var err error
p := v23.GetPrincipal(ctx)
- if browspr.principalManager, err = principal.NewPrincipalManager(p, &principal.InMemorySerializer{}); err != nil {
+
+ if principalSerializer == nil {
+ ctx.Fatalf("principalSerializer must not be nil")
+ }
+
+ if browspr.principalManager, err = principal.NewPrincipalManager(p, principalSerializer); err != nil {
ctx.Fatalf("principal.NewPrincipalManager failed: %s", err)
}
@@ -167,7 +174,8 @@
// HandleAuthGetAccountsRpc gets the root account name from the account manager.
func (b *Browspr) HandleAuthGetAccountsRpc(*vdl.Value) (*vdl.Value, error) {
- return vdl.ValueFromReflect(reflect.ValueOf(b.accountManager.GetAccounts()))
+ accounts := b.accountManager.GetAccounts()
+ return vdl.ValueFromReflect(reflect.ValueOf(accounts))
}
// HandleAuthOriginHasAccountRpc returns true iff the origin has an associated
diff --git a/services/wspr/internal/browspr/browspr_account_test.go b/services/wspr/internal/browspr/browspr_account_test.go
index ff35bbf..7c0ae4f 100644
--- a/services/wspr/internal/browspr/browspr_account_test.go
+++ b/services/wspr/internal/browspr/browspr_account_test.go
@@ -6,6 +6,8 @@
import (
"fmt"
+ "reflect"
+ "sort"
"testing"
"v.io/v23"
@@ -15,6 +17,7 @@
"v.io/v23/vdl"
_ "v.io/x/ref/runtime/factories/generic"
+ "v.io/x/ref/services/wspr/internal/principal"
"v.io/x/ref/test"
)
@@ -48,7 +51,7 @@
spec := v23.GetListenSpec(ctx)
spec.Proxy = "/mock/proxy"
mockPostMessage := func(_ int32, _, _ string) {}
- browspr := NewBrowspr(ctx, mockPostMessage, &spec, "/mock:1234/identd", nil)
+ browspr := NewBrowspr(ctx, mockPostMessage, &spec, "/mock:1234/identd", nil, principal.NewInMemorySerializer())
principal := v23.GetPrincipal(browspr.ctx)
browspr.accountManager.SetMockBlesser(newMockBlesserService(principal))
@@ -109,8 +112,13 @@
if err != nil {
t.Fatalf("browspr.HandleAuthGetAccountsRpc(%v) failed: %v", nilValue, err)
}
- if want := vdl.ValueOf([]string{account1.RawString(), account2.RawString()}); !vdl.EqualValue(want, gotAccounts2) {
- t.Fatalf("Expected account to be %v but got empty but got %v", want, gotAccounts2)
+ var got []string
+ if err := vdl.Convert(&got, gotAccounts2); err != nil {
+ t.Fatalf("vdl.Convert failed: %v", err)
+ }
+ sort.Strings(got)
+ if want := []string{account1.RawString(), account2.RawString()}; !reflect.DeepEqual(want, got) {
+ t.Fatalf("Expected account to be %v but got %v", want, got)
}
// Verify that principalManager has both accounts
diff --git a/services/wspr/internal/browspr/browspr_test.go b/services/wspr/internal/browspr/browspr_test.go
index 20b1466..4ca50c8 100644
--- a/services/wspr/internal/browspr/browspr_test.go
+++ b/services/wspr/internal/browspr/browspr_test.go
@@ -25,6 +25,7 @@
"v.io/x/ref/services/mounttable/mounttablelib"
"v.io/x/ref/services/wspr/internal/app"
"v.io/x/ref/services/wspr/internal/lib"
+ "v.io/x/ref/services/wspr/internal/principal"
"v.io/x/ref/test"
)
@@ -161,7 +162,7 @@
}
v23.GetNamespace(ctx).SetRoots(root)
- browspr := NewBrowspr(ctx, postMessageHandler, &spec, "/mock:1234/identd", []string{root})
+ browspr := NewBrowspr(ctx, postMessageHandler, &spec, "/mock:1234/identd", []string{root}, principal.NewInMemorySerializer())
// browspr sets its namespace root to use the "ws" protocol, but we want to force "tcp" here.
browspr.namespaceRoots = []string{root}
diff --git a/services/wspr/internal/browspr/file_serializer_nacl.go b/services/wspr/internal/principal/file_serializer_nacl.go
similarity index 93%
rename from services/wspr/internal/browspr/file_serializer_nacl.go
rename to services/wspr/internal/principal/file_serializer_nacl.go
index ceed808..5a73758 100644
--- a/services/wspr/internal/browspr/file_serializer_nacl.go
+++ b/services/wspr/internal/principal/file_serializer_nacl.go
@@ -2,12 +2,14 @@
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
-package browspr
+package principal
import (
"io"
"os"
"runtime/ppapi"
+
+ "v.io/x/ref/lib/security"
)
// fileSerializer implements vsecurity.SerializerReaderWriter that persists state to
@@ -21,6 +23,8 @@
signatureFile string
}
+var _ security.SerializerReaderWriter = (*fileSerializer)(nil)
+
func (fs *fileSerializer) Readers() (data io.ReadCloser, sig io.ReadCloser, err error) {
if fs.data == nil || fs.signature == nil {
return nil, nil, nil
diff --git a/services/wspr/internal/principal/in_memory_serializer.go b/services/wspr/internal/principal/in_memory_serializer.go
new file mode 100644
index 0000000..5ae1dad
--- /dev/null
+++ b/services/wspr/internal/principal/in_memory_serializer.go
@@ -0,0 +1,39 @@
+// Copyright 2015 The Vanadium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+package principal
+
+import (
+ "io"
+
+ "v.io/x/ref/lib/security"
+)
+
+// inMemorySerializer implements SerializerReaderWriter. This Serializer should only
+// be used in tests.
+type inMemorySerializer struct {
+ data bufferCloser
+ signature bufferCloser
+ hasData bool
+}
+
+var _ security.SerializerReaderWriter = (*inMemorySerializer)(nil)
+
+func NewInMemorySerializer() *inMemorySerializer {
+ return &inMemorySerializer{}
+}
+
+func (s *inMemorySerializer) Readers() (io.ReadCloser, io.ReadCloser, error) {
+ if !s.hasData {
+ return nil, nil, nil
+ }
+ return &s.data, &s.signature, nil
+}
+
+func (s *inMemorySerializer) Writers() (io.WriteCloser, io.WriteCloser, error) {
+ s.hasData = true
+ s.data.Reset()
+ s.signature.Reset()
+ return &s.data, &s.signature, nil
+}
diff --git a/services/wspr/internal/principal/principal.go b/services/wspr/internal/principal/principal.go
index f13ac68..dc7dca2 100644
--- a/services/wspr/internal/principal/principal.go
+++ b/services/wspr/internal/principal/principal.go
@@ -33,7 +33,6 @@
import (
"bytes"
"fmt"
- "io"
"net/url"
"sync"
"time"
@@ -91,28 +90,6 @@
return nil
}
-// InMemorySerializer implements SerializerReaderWriter. This Serializer should only
-// be used in tests.
-type InMemorySerializer struct {
- data bufferCloser
- signature bufferCloser
- hasData bool
-}
-
-func (s *InMemorySerializer) Readers() (io.ReadCloser, io.ReadCloser, error) {
- if !s.hasData {
- return nil, nil, nil
- }
- return &s.data, &s.signature, nil
-}
-
-func (s *InMemorySerializer) Writers() (io.WriteCloser, io.WriteCloser, error) {
- s.hasData = true
- s.data.Reset()
- s.signature.Reset()
- return &s.data, &s.signature, nil
-}
-
// PrincipalManager manages app principals. We only serialize the accounts
// associated with this principal manager and the mapping of apps to permissions
// that they were given.
@@ -261,6 +238,19 @@
return nil
}
+// GetAccounts returns a list of account names known to the Principal Manager.
+func (i *PrincipalManager) GetAccounts() []string {
+ i.mu.Lock()
+ defer i.mu.Unlock()
+
+ accounts := make([]string, 0, len(i.state.Accounts))
+ for account := range i.state.Accounts {
+ accounts = append(accounts, account)
+ }
+
+ return accounts
+}
+
// AddOrigin adds an origin to the manager linked to the given account.
func (i *PrincipalManager) AddOrigin(origin string, account string, caveats []security.Caveat, expirations []time.Time) error {
i.mu.Lock()
diff --git a/services/wspr/internal/principal/principal_test.go b/services/wspr/internal/principal/principal_test.go
index 858d2cc..189c052 100644
--- a/services/wspr/internal/principal/principal_test.go
+++ b/services/wspr/internal/principal/principal_test.go
@@ -134,7 +134,7 @@
func TestPrincipalManager(t *testing.T) {
root := testutil.NewPrincipal()
- m, err := NewPrincipalManager(root, &InMemorySerializer{})
+ m, err := NewPrincipalManager(root, NewInMemorySerializer())
if err != nil {
t.Fatalf("NewPrincipalManager failed: %v", err)
}
@@ -150,7 +150,7 @@
func TestPrincipalManagerPersistence(t *testing.T) {
root := testutil.NewPrincipal()
- serializer := &InMemorySerializer{}
+ serializer := NewInMemorySerializer()
m, err := NewPrincipalManager(root, serializer)
if err != nil {
t.Fatalf("NewPrincipalManager failed: %v", err)
@@ -174,7 +174,7 @@
func TestOriginHasAccount(t *testing.T) {
root := testutil.NewPrincipal()
- m, err := NewPrincipalManager(root, &InMemorySerializer{})
+ m, err := NewPrincipalManager(root, NewInMemorySerializer())
if err != nil {
t.Fatalf("NewPrincipalManager failed: %v", err)
}
diff --git a/services/wspr/wsprlib/wspr.go b/services/wspr/wsprlib/wspr.go
index d626ed5..4622136 100644
--- a/services/wspr/wsprlib/wspr.go
+++ b/services/wspr/wsprlib/wspr.go
@@ -123,10 +123,10 @@
pipes: map[*http.Request]*pipe{},
}
- // TODO(nlacasse, bjornick) use a serializer that can actually persist.
p := v23.GetPrincipal(ctx)
var err error
- if wspr.principalManager, err = principal.NewPrincipalManager(p, &principal.InMemorySerializer{}); err != nil {
+ // TODO(nlacasse): Use a serializer that can actually persist, as we do in browspr.
+ if wspr.principalManager, err = principal.NewPrincipalManager(p, principal.NewInMemorySerializer()); err != nil {
ctx.Fatalf("principal.NewPrincipalManager failed: %s", err)
}