blob: 82111941cc085a62b4b0120e8ae3a9d4b9bc105e [file] [log] [blame]
// Copyright 2015 The Vanadium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package server
import (
"math/rand"
"path"
"sync"
"time"
"v.io/v23/context"
"v.io/v23/glob"
"v.io/v23/query/engine"
ds "v.io/v23/query/engine/datasource"
"v.io/v23/query/syncql"
"v.io/v23/rpc"
"v.io/v23/security"
"v.io/v23/security/access"
wire "v.io/v23/services/syncbase"
pubutil "v.io/v23/syncbase/util"
"v.io/v23/verror"
"v.io/v23/vom"
"v.io/x/lib/vlog"
"v.io/x/ref/services/syncbase/common"
"v.io/x/ref/services/syncbase/server/interfaces"
"v.io/x/ref/services/syncbase/server/util"
"v.io/x/ref/services/syncbase/store"
storeutil "v.io/x/ref/services/syncbase/store/util"
"v.io/x/ref/services/syncbase/store/watchable"
"v.io/x/ref/services/syncbase/vsync"
sbwatchable "v.io/x/ref/services/syncbase/watchable"
)
// database is a per-database singleton (i.e. not per-request) that handles
// Database RPCs.
// Note: If a database does not exist at the time of a database RPC, the
// dispatcher creates a short-lived database object to service that particular
// request.
type database struct {
id wire.Id
s *service
// The fields below are initialized iff this database exists.
exists bool
// TODO(sadovsky): Make st point to a store.Store wrapper that handles paging,
// and do not actually open the store in NewDatabase.
st *watchable.Store // stores all data for a single database
// Active snapshots and transactions corresponding to client batches.
// TODO(sadovsky): Add timeouts and GC.
mu sync.Mutex // protects the fields below
sns map[uint64]store.Snapshot
txs map[uint64]*transactionState
// Active ConflictResolver connection from the app to this database.
// NOTE: For now, we assume there's only one open conflict resolution stream
// per database (typically, from the app that owns the database).
crStream wire.ConflictManagerStartConflictResolverServerCall
// Mutex lock to protect concurrent read/write of crStream pointer
crMu sync.Mutex
}
var (
_ wire.DatabaseServerMethods = (*database)(nil)
_ interfaces.Database = (*database)(nil)
)
// DatabaseOptions configures a database.
type DatabaseOptions struct {
// Database-level permissions.
Perms access.Permissions
// Root dir for data storage.
RootDir string
// Storage engine to use.
Engine string
}
type permissionState struct {
dataChanged bool
permsChanged bool
initialPerms access.Permissions
finalPerms access.Permissions
}
type transactionState struct {
tx *watchable.Transaction
permsChanges map[wire.Id]*permissionState
}
// Keeps track that this collection had a mutation and the permissions at the time. If no
// permissions are yet known for the collection then remember the current permissions as the
// initial permissions of the collection.
// When the transaction is committed we will know to validate this collection's permissions.
func (ts *transactionState) MarkDataChanged(collectionId wire.Id, perms access.Permissions) {
state := ts.permsState(collectionId)
state.dataChanged = true
if state.initialPerms == nil {
state.initialPerms = perms
}
state.finalPerms = perms
}
// Keeps track that the permissions were changed on this collection and the before and after
// permissions. If no permissions are yet known for the collection then remember the current
// permissions as the initial permissions of the collection.
func (ts *transactionState) MarkPermsChanged(collectionId wire.Id, permsBefore access.Permissions, permsAfter access.Permissions) {
state := ts.permsState(collectionId)
state.permsChanged = true
if state.initialPerms == nil {
state.initialPerms = permsBefore
}
state.finalPerms = permsAfter
}
// Resets all tracked changes to the collection. Used on collection destroy. Since destroy
// cannot happen on a synced collection, the destroy and any updates before it will not be
// seen remotely, so validation must start from the implicit permissions in case the
// collection is created again. This also allows destroy to not require both write and
// admin permissions.
func (ts *transactionState) ResetCollectionChanges(collectionId wire.Id) {
delete(ts.permsChanges, collectionId)
}
// validatePermissionChanges performs an auth check on each collection that has a data change or
// permission change and returns false if any of the auth checks fail.
// TODO(ivanpi): This check should be done against signing blessings at signing time, in
// both batch and non-batch cases.
func (ts *transactionState) validatePermissionChanges(ctx *context.T, securityCall security.Call) bool {
for _, collectionState := range ts.permsChanges {
// This collection was modified, make sure that the write acl is either present at
// the end or that it had the write acl to begin with. This way we can be sure that
// a mutation didn't take place when it appeared that there was no write acl before
// and after the transaction.
if collectionState.dataChanged {
before := hasPermission(ctx, securityCall, collectionState.initialPerms, access.Write)
after := hasPermission(ctx, securityCall, collectionState.finalPerms, access.Write)
if !after && !before {
return false
}
}
// The permissions were changed on the collection, make sure that the admin acl is
// present at the beginning.
if collectionState.permsChanged {
if !hasPermission(ctx, securityCall, collectionState.initialPerms, access.Admin) {
return false
}
}
}
return true
}
func (ts *transactionState) permsState(collectionId wire.Id) *permissionState {
if ts.permsChanges == nil {
ts.permsChanges = make(map[wire.Id]*permissionState)
}
state, ok := ts.permsChanges[collectionId]
if !ok {
state = &permissionState{}
ts.permsChanges[collectionId] = state
}
return state
}
// hasPermission returns true if the caller is authorized for the specific tag based on the
// passed in perms.
func hasPermission(ctx *context.T, securityCall security.Call, perms access.Permissions, tag access.Tag) bool {
permForTag, ok := perms[string(tag)]
// Authorize returns either an error or nil, so nil means the caller is authorized.
return ok && permForTag.Authorize(ctx, securityCall) == nil
}
// openDatabase opens a database and returns a *database for it. Designed for
// use from within newDatabase and newService.
func openDatabase(ctx *context.T, s *service, id wire.Id, opts DatabaseOptions, openOpts storeutil.OpenOptions) (*database, error) {
st, err := storeutil.OpenStore(opts.Engine, path.Join(opts.RootDir, opts.Engine), openOpts)
if err != nil {
return nil, err
}
wst, err := watchable.Wrap(st, s.vclock, &watchable.Options{
// TODO(ivanpi): Since ManagedPrefixes control what gets synced, they
// should be moved to a more visible place (e.g. constants). Also consider
// decoupling managed and synced prefixes.
ManagedPrefixes: []string{common.CollectionPermsPrefix, common.RowPrefix},
})
if err != nil {
return nil, err
}
return &database{
id: id,
s: s,
exists: true,
st: wst,
sns: make(map[uint64]store.Snapshot),
txs: make(map[uint64]*transactionState),
}, nil
}
// newDatabase creates a new database instance and returns it.
// Designed for use from within service.createDatabase.
func newDatabase(ctx *context.T, s *service, id wire.Id, metadata *wire.SchemaMetadata, opts DatabaseOptions) (*database, error) {
if opts.Perms == nil {
return nil, verror.New(verror.ErrInternal, ctx, "perms must be specified")
}
d, err := openDatabase(ctx, s, id, opts, storeutil.OpenOptions{CreateIfMissing: true, ErrorIfExists: true})
if err != nil {
return nil, err
}
data := &DatabaseData{
Perms: opts.Perms,
SchemaMetadata: metadata,
}
if err := store.Put(ctx, d.st, d.stKey(), data); err != nil {
return nil, err
}
// Start a Sync watcher on this newly created database store.
vsync.NewSyncDatabase(d).StartStoreWatcher(ctx)
return d, nil
}
////////////////////////////////////////
// RPC methods
func (d *database) Create(ctx *context.T, call rpc.ServerCall, metadata *wire.SchemaMetadata, perms access.Permissions) error {
if d.exists {
return verror.New(verror.ErrExist, ctx, d.id)
}
// This database does not yet exist; d is just an ephemeral handle that holds
// {id wire.Id, s *service}. d.s.createDatabase will create a new database
// handle and store it in d.s.dbs[d.id].
return d.s.createDatabase(ctx, call, d.id, perms, metadata)
}
func (d *database) Destroy(ctx *context.T, call rpc.ServerCall) error {
return d.s.destroyDatabase(ctx, call, d.id)
}
func (d *database) Exists(ctx *context.T, call rpc.ServerCall) (bool, error) {
if !d.exists {
return false, nil
}
return util.ErrorToExists(util.GetWithAuth(ctx, call, d.st, d.stKey(), &DatabaseData{}))
}
var rng *rand.Rand = rand.New(rand.NewSource(time.Now().UTC().UnixNano()))
func (d *database) BeginBatch(ctx *context.T, call rpc.ServerCall, opts wire.BatchOptions) (wire.BatchHandle, error) {
if !d.exists {
return "", verror.New(verror.ErrNoExist, ctx, d.id)
}
d.mu.Lock()
defer d.mu.Unlock()
var id uint64
var batchType common.BatchType
for {
id = uint64(rng.Int63())
if opts.ReadOnly {
if _, ok := d.sns[id]; !ok {
d.sns[id] = d.st.NewSnapshot()
batchType = common.BatchTypeSn
break
}
} else {
if _, ok := d.txs[id]; !ok {
d.txs[id] = &transactionState{tx: d.st.NewWatchableTransaction()}
batchType = common.BatchTypeTx
break
}
}
}
return common.JoinBatchHandle(batchType, id), nil
}
func (d *database) Commit(ctx *context.T, call rpc.ServerCall, bh wire.BatchHandle) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
if bh == "" {
return wire.NewErrNotBoundToBatch(ctx)
}
_, ts, batchId, err := d.batchLookupInternal(ctx, bh)
if err != nil {
return err
}
if ts == nil {
return wire.NewErrReadOnlyBatch(ctx)
}
if !ts.validatePermissionChanges(ctx, call.Security()) {
return wire.NewErrInvalidPermissionsChange(ctx)
}
if err = ts.tx.Commit(); err == nil {
d.mu.Lock()
delete(d.txs, batchId)
d.mu.Unlock()
}
// TODO(ivanpi): Best effort abort if commit fails? Watchable Commit can fail
// before the underlying snapshot is aborted.
if verror.ErrorID(err) == store.ErrConcurrentTransaction.ID {
return verror.New(wire.ErrConcurrentBatch, ctx, err)
}
return err
}
func (d *database) Abort(ctx *context.T, call rpc.ServerCall, bh wire.BatchHandle) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
if bh == "" {
return wire.NewErrNotBoundToBatch(ctx)
}
sn, ts, batchId, err := d.batchLookupInternal(ctx, bh)
if err != nil {
return err
}
if ts != nil {
d.mu.Lock()
delete(d.txs, batchId)
d.mu.Unlock()
// TODO(ivanpi): If tx.Abort fails, retry later?
return ts.tx.Abort()
} else {
d.mu.Lock()
delete(d.sns, batchId)
d.mu.Unlock()
// TODO(ivanpi): If sn.Abort fails, retry later?
return sn.Abort()
}
}
func (d *database) Exec(ctx *context.T, call wire.DatabaseExecServerCall, bh wire.BatchHandle, q string, params []*vom.RawBytes) error {
// RunInTransaction() cannot be used here because we may or may not be
// creating a transaction. qe.Exec must be called and the statement must be
// parsed before we know if a snapshot or a transaction should be created. To
// duplicate the semantics of RunInTransaction, if we are not inside a batch,
// we attempt the Exec up to 100 times and retry on ErrConcurrentTransaction.
// TODO(ivanpi): Refactor query parsing into a separate step, simplify request
// handling. Consider separate Query and Exec methods.
maxAttempts := 100
attempt := 0
for {
err := d.execInternal(ctx, call, bh, q, params)
if bh != "" || attempt >= maxAttempts || verror.ErrorID(err) != store.ErrConcurrentTransaction.ID {
return err
}
attempt++
}
}
func (d *database) execInternal(ctx *context.T, call wire.DatabaseExecServerCall, bh wire.BatchHandle, q string, params []*vom.RawBytes) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
impl := func() error {
db := &queryDb{
ctx: ctx,
call: call,
d: d,
bh: bh,
sntx: nil, // Filled in later with existing or created sn/tx.
ts: nil, // Only filled in if a new batch was created.
}
st, err := engine.Create(db).PrepareStatement(q)
if err != nil {
return execCommitOrAbort(db, err)
}
headers, rs, err := st.Exec(params...)
if err != nil {
return execCommitOrAbort(db, err)
}
if rs.Err() != nil {
return execCommitOrAbort(db, err)
}
sender := call.SendStream()
// Push the headers first -- the client will retrieve them and return
// them separately from the results.
var resultHeaders []*vom.RawBytes
for _, header := range headers {
resultHeaders = append(resultHeaders, vom.RawBytesOf(header))
}
sender.Send(resultHeaders)
for rs.Advance() {
result := rs.Result()
if err := sender.Send(result); err != nil {
rs.Cancel()
return execCommitOrAbort(db, err)
}
}
return execCommitOrAbort(db, rs.Err())
}
return impl()
}
func execCommitOrAbort(qdb *queryDb, err error) error {
if qdb.bh != "" {
return err // part of an enclosing sn/tx
}
if err != nil {
if qdb.sntx != nil {
qdb.sntx.Abort()
}
return err
} else { // err is nil
if qdb.ts != nil {
return qdb.ts.tx.Commit()
} else if qdb.sntx != nil {
return qdb.sntx.Abort()
}
return nil
}
}
func (d *database) SetPermissions(ctx *context.T, call rpc.ServerCall, perms access.Permissions, version string) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
return d.s.setDatabasePerms(ctx, call, d.id, perms, version)
}
func (d *database) GetPermissions(ctx *context.T, call rpc.ServerCall) (perms access.Permissions, version string, err error) {
if !d.exists {
return nil, "", verror.New(verror.ErrNoExist, ctx, d.id)
}
data := &DatabaseData{}
if err := util.GetWithAuth(ctx, call, d.st, d.stKey(), data); err != nil {
return nil, "", err
}
return data.Perms, util.FormatVersion(data.Version), nil
}
func (d *database) GlobChildren__(ctx *context.T, call rpc.GlobChildrenServerCall, matcher *glob.Element) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
impl := func(sntx store.SnapshotOrTransaction) error {
// Check perms.
if err := util.GetWithAuth(ctx, call, sntx, d.stKey(), &DatabaseData{}); err != nil {
return err
}
return util.GlobChildren(ctx, call, matcher, sntx, common.CollectionPermsPrefix)
}
return store.RunWithSnapshot(d.st, impl)
}
// See comment in v.io/v23/services/syncbase/service.vdl for why we can't
// implement ListCollections using Glob.
func (d *database) ListCollections(ctx *context.T, call rpc.ServerCall, bh wire.BatchHandle) ([]wire.Id, error) {
if !d.exists {
return nil, verror.New(verror.ErrNoExist, ctx, d.id)
}
var res []wire.Id
impl := func(sntx store.SnapshotOrTransaction) error {
// Check perms.
if err := util.GetWithAuth(ctx, call, sntx, d.stKey(), &DatabaseData{}); err != nil {
return err
}
it := sntx.Scan(common.ScanPrefixArgs(common.CollectionPermsPrefix, ""))
keyBytes := []byte{}
for it.Advance() {
keyBytes = it.Key(keyBytes)
parts := common.SplitNKeyParts(string(keyBytes), 3)
id, err := pubutil.DecodeId(parts[1])
if err != nil {
it.Cancel()
return verror.New(verror.ErrInternal, ctx, err)
}
res = append(res, id)
}
if err := it.Err(); err != nil {
return err
}
return nil
}
if err := d.runWithExistingBatchOrNewSnapshot(ctx, bh, impl); err != nil {
return nil, err
}
return res, nil
}
func (d *database) PauseSync(ctx *context.T, call rpc.ServerCall) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
return d.runInTransaction(func(ts *transactionState) error {
return sbwatchable.AddDbStateChangeRequestOp(ctx, ts.tx, sbwatchable.StateChangePauseSync)
})
}
func (d *database) ResumeSync(ctx *context.T, call rpc.ServerCall) error {
if !d.exists {
return verror.New(verror.ErrNoExist, ctx, d.id)
}
return d.runInTransaction(func(ts *transactionState) error {
return sbwatchable.AddDbStateChangeRequestOp(ctx, ts.tx, sbwatchable.StateChangeResumeSync)
})
}
////////////////////////////////////////
// interfaces.Database methods
func (d *database) St() *watchable.Store {
if !d.exists {
vlog.Fatalf("database %v does not exist", d.id)
}
return d.st
}
func (d *database) Service() interfaces.Service {
return d.s
}
func (d *database) CheckPermsInternal(ctx *context.T, call rpc.ServerCall, st store.StoreReader) error {
if !d.exists {
vlog.Fatalf("database %v does not exist", d.id)
}
return util.GetWithAuth(ctx, call, st, d.stKey(), &DatabaseData{})
}
func (d *database) Id() wire.Id {
return d.id
}
func (d *database) CrConnectionStream() wire.ConflictManagerStartConflictResolverServerStream {
d.crMu.Lock()
defer d.crMu.Unlock()
return d.crStream
}
func (d *database) ResetCrConnectionStream() {
d.crMu.Lock()
defer d.crMu.Unlock()
// TODO(jlodhia): figure out a way for the connection to gracefully shutdown
// so that the client can get an appropriate error msg.
d.crStream = nil
}
////////////////////////////////////////
// query interface implementations
// queryDb implements ds.Database.
type queryDb struct {
ctx *context.T
call rpc.ServerCall
d *database
bh wire.BatchHandle
sntx store.SnapshotOrTransaction
ts *transactionState // Will only be set if in a transaction (else nil)
}
func (qdb *queryDb) GetContext() *context.T {
return qdb.ctx
}
func (qdb *queryDb) GetTable(name string, writeAccessReq bool) (ds.Table, error) {
// At this point, when the query package calls GetTable with the
// writeAccessReq arg, we know whether or not we need a [writable] transaction
// or a snapshot. If batchId is already set, there's nothing to do; but if
// not, the writeAccessReq arg dictates whether a snapshot or a transaction is
// should be created.
// TODO(ivanpi): Allow passing in non-default user blessings.
userBlessings, _ := security.RemoteBlessingNames(qdb.ctx, qdb.call.Security())
_, user, err := pubutil.AppAndUserPatternFromBlessings(userBlessings...)
if err != nil {
return nil, err
}
qt := &queryTable{
qdb: qdb,
cReq: &collectionReq{
id: wire.Id{Blessing: string(user), Name: name},
d: qdb.d,
},
}
if qt.qdb.bh != "" {
var err error
if writeAccessReq {
// We are in a batch (could be snapshot or transaction)
// and Write access is required. Attempt to get a
// transaction from the request.
qt.qdb.ts, err = qt.qdb.d.batchTransaction(qt.qdb.GetContext(), qt.qdb.bh)
if err != nil {
if verror.ErrorID(err) == wire.ErrReadOnlyBatch.ID {
// We are in a snapshot batch, write access cannot be provided.
// Return NotWritable.
return nil, syncql.NewErrNotWritable(qt.qdb.GetContext(), pubutil.EncodeId(qt.cReq.id))
}
return nil, err
}
qt.qdb.sntx = qt.qdb.ts.tx
} else {
qt.qdb.sntx, err = qt.qdb.d.batchReader(qt.qdb.GetContext(), qt.qdb.bh)
if err != nil {
return nil, err
}
}
} else {
// Now that we know if write access is required, create a snapshot
// or transaction.
if !writeAccessReq {
qt.qdb.sntx = qt.qdb.d.st.NewSnapshot()
} else { // writeAccessReq
qt.qdb.ts = &transactionState{tx: qt.qdb.d.st.NewWatchableTransaction()}
qt.qdb.sntx = qt.qdb.ts.tx
}
}
// Now that we have a collection, we need to check permissions.
collectionPerms, err := qt.cReq.checkAccess(qdb.ctx, qdb.call, qdb.sntx)
if err != nil {
return nil, err
}
if writeAccessReq {
qt.qdb.ts.MarkDataChanged(qt.cReq.id, collectionPerms)
}
return qt, nil
}
// queryTable implements ds.Table.
type queryTable struct {
qdb *queryDb
cReq *collectionReq
}
func (t *queryTable) GetIndexFields() []ds.Index {
// TODO(jkline): If and when secondary indexes are supported, they
// would be supplied here.
return []ds.Index{}
}
func (t *queryTable) Delete(k string) (bool, error) {
// Create a rowReq and call delete. Permissions will be checked.
rowReq := &rowReq{
key: k,
c: t.cReq,
}
if err := rowReq.delete(t.qdb.GetContext(), t.qdb.call, t.qdb.ts); err != nil {
return false, err
}
return true, nil
}
func (t *queryTable) Scan(indexRanges ...ds.IndexRanges) (ds.KeyValueStream, error) {
streams := []store.Stream{}
// Syncbase does not currently support secondary indexes. As such, indexRanges
// is guaranteed to be one in size as it will only specify the key ranges;
// hence, indexRanges[0] below.
for _, keyRange := range *indexRanges[0].StringRanges {
// TODO(jkline): For now, acquire all of the streams at once to minimize the
// race condition. Need a way to Scan multiple ranges at the same state of
// uncommitted changes.
streams = append(streams, t.qdb.sntx.Scan(common.ScanRangeArgs(common.JoinKeyParts(common.RowPrefix, t.cReq.stKeyPart()), keyRange.Start, keyRange.Limit)))
}
return &kvs{
t: t,
curr: 0,
validRow: false,
it: streams,
err: nil,
}, nil
}
// kvs implements ds.KeyValueStream.
type kvs struct {
t *queryTable
curr int
validRow bool
currKey string
currValue *vom.RawBytes
it []store.Stream // array of store.Streams
err error
}
func (s *kvs) Advance() bool {
if s.err != nil {
return false
}
for s.curr < len(s.it) {
if s.it[s.curr].Advance() {
// key
keyBytes := s.it[s.curr].Key(nil)
parts := common.SplitNKeyParts(string(keyBytes), 3)
// TODO(rogulenko): Check access for the key.
s.currKey = parts[2]
// value
valueBytes := s.it[s.curr].Value(nil)
var currValue *vom.RawBytes
if err := vom.Decode(valueBytes, &currValue); err != nil {
s.validRow = false
s.err = err
s.Cancel() // to cancel iterators after s.curr
return false
}
s.currValue = currValue
s.validRow = true
return true
}
// Advance returned false. It could be an err, or it could
// be we've reached the end.
if err := s.it[s.curr].Err(); err != nil {
s.validRow = false
s.err = err
s.Cancel() // to cancel iterators after s.curr
return false
}
// We've reached the end of the iterator for this keyRange.
// Jump to the next one.
s.it[s.curr] = nil
s.curr++
s.validRow = false
}
// There are no more prefixes to scan.
return false
}
func (s *kvs) KeyValue() (string, *vom.RawBytes) {
if !s.validRow {
return "", nil
}
return s.currKey, s.currValue
}
func (s *kvs) Err() error {
return s.err
}
func (s *kvs) Cancel() {
if s.it != nil {
for i := s.curr; i < len(s.it); i++ {
s.it[i].Cancel()
}
s.it = nil
}
// set curr to end of keyRanges so Advance will return false
s.curr = len(s.it)
}
////////////////////////////////////////
// Internal helpers
func (d *database) stKey() string {
return common.DatabasePrefix
}
func (d *database) runWithExistingBatchOrNewSnapshot(ctx *context.T, bh wire.BatchHandle, fn func(sntx store.SnapshotOrTransaction) error) error {
if bh != "" {
if sntx, err := d.batchReader(ctx, bh); err != nil {
// Batch does not exist.
return err
} else {
return fn(sntx)
}
} else {
return store.RunWithSnapshot(d.st, fn)
}
}
func (d *database) runInExistingBatchOrNewTransaction(ctx *context.T, bh wire.BatchHandle, fn func(ts *transactionState) error) error {
if bh != "" {
if batch, err := d.batchTransaction(ctx, bh); err != nil {
// Batch does not exist or is readonly (snapshot).
return err
} else {
return fn(batch)
}
} else {
return d.runInTransaction(fn)
}
}
func (d *database) batchReader(ctx *context.T, bh wire.BatchHandle) (store.SnapshotOrTransaction, error) {
sn, ts, _, err := d.batchLookupInternal(ctx, bh)
if err != nil {
return nil, err
}
if sn != nil {
return sn, nil
}
return ts.tx, nil
}
func (d *database) batchTransaction(ctx *context.T, bh wire.BatchHandle) (*transactionState, error) {
sn, ts, _, err := d.batchLookupInternal(ctx, bh)
if err != nil {
return nil, err
}
if sn != nil {
return nil, wire.NewErrReadOnlyBatch(ctx)
}
return ts, nil
}
// batchLookupInternal parses the batch handle and retrieves the corresponding
// snapshot or transaction. It returns an error if the handle is malformed or
// the batch does not exist. Otherwise, exactly one of sn and ts will be != nil.
func (d *database) batchLookupInternal(ctx *context.T, bh wire.BatchHandle) (sn store.Snapshot, ts *transactionState, batchId uint64, _ error) {
if bh == "" {
return nil, nil, 0, verror.New(verror.ErrInternal, ctx, "batch lookup for empty handle")
}
bType, bId, err := common.SplitBatchHandle(bh)
if err != nil {
return nil, nil, 0, err
}
d.mu.Lock()
defer d.mu.Unlock()
var found bool
switch bType {
case common.BatchTypeSn:
sn, found = d.sns[bId]
case common.BatchTypeTx:
ts, found = d.txs[bId]
}
if !found {
return nil, nil, bId, wire.NewErrUnknownBatch(ctx)
}
return sn, ts, bId, nil
}
func (d *database) setPermsInternal(ctx *context.T, call rpc.ServerCall, perms access.Permissions, version string) error {
if !d.exists {
vlog.Fatalf("database %v does not exist", d.id)
}
if err := common.ValidatePerms(ctx, perms, wire.AllDatabaseTags); err != nil {
return err
}
return store.RunInTransaction(d.st, func(tx store.Transaction) error {
data := &DatabaseData{}
return util.UpdateWithAuth(ctx, call, tx, d.stKey(), data, func() error {
if err := util.CheckVersion(ctx, version, data.Version); err != nil {
return err
}
data.Perms = perms
data.Version++
return nil
})
})
}
// runInTransaction runs the given fn in a transaction, managing retries and
// commit/abort.
func (d *database) runInTransaction(fn func(ts *transactionState) error) error {
// TODO(rogulenko): Change the default number of attempts to 3. Currently,
// some storage engine tests fail when the number of attempts is that low.
return d.runInTransactionWithOpts(&store.TransactionOptions{NumAttempts: 100}, fn)
}
// runInTransactionWithOpts runs the given fn in a transaction, managing retries
// and commit/abort.
func (d *database) runInTransactionWithOpts(opts *store.TransactionOptions, fn func(ts *transactionState) error) error {
var err error
for i := 0; i < opts.NumAttempts; i++ {
// TODO(sadovsky): Should NewTransaction return an error? If not, how will
// we deal with RPC errors when talking to remote storage engines? (Note,
// client-side BeginBatch returns an error.)
ts := &transactionState{tx: d.st.NewWatchableTransaction()}
if err = fn(ts); err != nil {
ts.tx.Abort()
return err
}
// TODO(sadovsky): Commit() can fail for a number of reasons, e.g. RPC
// failure or ErrConcurrentTransaction. Depending on the cause of failure,
// it may be desirable to retry the Commit() and/or to call Abort().
if err = ts.tx.Commit(); verror.ErrorID(err) != store.ErrConcurrentTransaction.ID {
return err
}
}
return err
}