bolt db works
This commit is contained in:
parent
d261e5d870
commit
82e4d122fc
5 changed files with 272 additions and 203 deletions
|
|
@ -41,6 +41,7 @@ import (
|
||||||
"github.com/google/cayley/quad/nquads"
|
"github.com/google/cayley/quad/nquads"
|
||||||
|
|
||||||
// Load all supported backends.
|
// Load all supported backends.
|
||||||
|
_ "github.com/google/cayley/graph/bolt"
|
||||||
_ "github.com/google/cayley/graph/leveldb"
|
_ "github.com/google/cayley/graph/leveldb"
|
||||||
_ "github.com/google/cayley/graph/memstore"
|
_ "github.com/google/cayley/graph/memstore"
|
||||||
_ "github.com/google/cayley/graph/mongo"
|
_ "github.com/google/cayley/graph/mongo"
|
||||||
|
|
|
||||||
9
db/db.go
9
db/db.go
|
|
@ -70,6 +70,7 @@ func OpenQuadWriter(qs graph.TripleStore, cfg *config.Config) (graph.QuadWriter,
|
||||||
|
|
||||||
func Load(qw graph.QuadWriter, cfg *config.Config, dec quad.Unmarshaler) error {
|
func Load(qw graph.QuadWriter, cfg *config.Config, dec quad.Unmarshaler) error {
|
||||||
block := make([]quad.Quad, 0, cfg.LoadSize)
|
block := make([]quad.Quad, 0, cfg.LoadSize)
|
||||||
|
count := 0
|
||||||
for {
|
for {
|
||||||
t, err := dec.Unmarshal()
|
t, err := dec.Unmarshal()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -80,11 +81,19 @@ func Load(qw graph.QuadWriter, cfg *config.Config, dec quad.Unmarshaler) error {
|
||||||
}
|
}
|
||||||
block = append(block, t)
|
block = append(block, t)
|
||||||
if len(block) == cap(block) {
|
if len(block) == cap(block) {
|
||||||
|
count += len(block)
|
||||||
qw.AddQuadSet(block)
|
qw.AddQuadSet(block)
|
||||||
|
if glog.V(2) {
|
||||||
|
glog.V(2).Infof("Wrote %d quads.", count)
|
||||||
|
}
|
||||||
block = block[:0]
|
block = block[:0]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
count += len(block)
|
||||||
qw.AddQuadSet(block)
|
qw.AddQuadSet(block)
|
||||||
|
if glog.V(2) {
|
||||||
|
glog.V(2).Infof("Wrote %d quads.", count)
|
||||||
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -19,6 +19,9 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/barakmich/glog"
|
||||||
|
"github.com/boltdb/bolt"
|
||||||
|
|
||||||
"github.com/google/cayley/graph"
|
"github.com/google/cayley/graph"
|
||||||
"github.com/google/cayley/graph/iterator"
|
"github.com/google/cayley/graph/iterator"
|
||||||
"github.com/google/cayley/quad"
|
"github.com/google/cayley/quad"
|
||||||
|
|
@ -27,34 +30,22 @@ import (
|
||||||
type AllIterator struct {
|
type AllIterator struct {
|
||||||
uid uint64
|
uid uint64
|
||||||
tags graph.Tagger
|
tags graph.Tagger
|
||||||
prefix []byte
|
bucket []byte
|
||||||
dir quad.Direction
|
dir quad.Direction
|
||||||
open bool
|
|
||||||
qs *QuadStore
|
qs *QuadStore
|
||||||
result graph.Value
|
result *Token
|
||||||
|
buffer [][]byte
|
||||||
|
offset int
|
||||||
|
done bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewAllIterator(prefix string, d quad.Direction, ts *QuadStore) *AllIterator {
|
func NewAllIterator(bucket []byte, d quad.Direction, qs *QuadStore) *AllIterator {
|
||||||
opts := &opt.ReadOptions{
|
|
||||||
DontFillCache: true,
|
|
||||||
}
|
|
||||||
|
|
||||||
it := AllIterator{
|
it := AllIterator{
|
||||||
uid: iterator.NextUID(),
|
uid: iterator.NextUID(),
|
||||||
ro: opts,
|
bucket: bucket,
|
||||||
iter: ts.db.NewIterator(nil, opts),
|
|
||||||
prefix: []byte(prefix),
|
|
||||||
dir: d,
|
dir: d,
|
||||||
open: true,
|
qs: qs,
|
||||||
ts: ts,
|
|
||||||
}
|
|
||||||
|
|
||||||
it.iter.Seek(it.prefix)
|
|
||||||
if !it.iter.Valid() {
|
|
||||||
// FIXME(kortschak) What are the semantics here? Is this iterator usable?
|
|
||||||
// If not, we should return nil *Iterator and an error.
|
|
||||||
it.open = false
|
|
||||||
it.iter.Release()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return &it
|
return &it
|
||||||
|
|
@ -65,15 +56,9 @@ func (it *AllIterator) UID() uint64 {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Reset() {
|
func (it *AllIterator) Reset() {
|
||||||
if !it.open {
|
it.buffer = nil
|
||||||
it.iter = it.ts.db.NewIterator(nil, it.ro)
|
it.offset = 0
|
||||||
it.open = true
|
it.done = false
|
||||||
}
|
|
||||||
it.iter.Seek(it.prefix)
|
|
||||||
if !it.iter.Valid() {
|
|
||||||
it.open = false
|
|
||||||
it.iter.Release()
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Tagger() *graph.Tagger {
|
func (it *AllIterator) Tagger() *graph.Tagger {
|
||||||
|
|
@ -91,28 +76,65 @@ func (it *AllIterator) TagResults(dst map[string]graph.Value) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Clone() graph.Iterator {
|
func (it *AllIterator) Clone() graph.Iterator {
|
||||||
out := NewAllIterator(string(it.prefix), it.dir, it.ts)
|
out := NewAllIterator(it.bucket, it.dir, it.qs)
|
||||||
out.tags.CopyFrom(it)
|
out.tags.CopyFrom(it)
|
||||||
return out
|
return out
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Next() bool {
|
func (it *AllIterator) Next() bool {
|
||||||
if !it.open {
|
if it.done {
|
||||||
it.result = nil
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
var out []byte
|
if len(it.buffer) <= it.offset+1 {
|
||||||
out = make([]byte, len(it.iter.Key()))
|
it.offset = 0
|
||||||
copy(out, it.iter.Key())
|
var last []byte
|
||||||
it.iter.Next()
|
if it.buffer != nil {
|
||||||
if !it.iter.Valid() {
|
last = it.buffer[len(it.buffer)-1]
|
||||||
it.Close()
|
}
|
||||||
|
it.buffer = make([][]byte, 0, bufferSize)
|
||||||
|
err := it.qs.db.View(func(tx *bolt.Tx) error {
|
||||||
|
i := 0
|
||||||
|
b := tx.Bucket(it.bucket)
|
||||||
|
cur := b.Cursor()
|
||||||
|
if last == nil {
|
||||||
|
k, _ := cur.First()
|
||||||
|
var out []byte
|
||||||
|
out = make([]byte, len(k))
|
||||||
|
copy(out, k)
|
||||||
|
it.buffer = append(it.buffer, out)
|
||||||
|
i++
|
||||||
|
} else {
|
||||||
|
k, _ := cur.Seek(last)
|
||||||
|
if !bytes.Equal(k, last) {
|
||||||
|
return fmt.Errorf("Couldn't pick up after", k)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i < bufferSize {
|
||||||
|
k, _ := cur.Next()
|
||||||
|
if k == nil {
|
||||||
|
it.buffer = append(it.buffer, k)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
var out []byte
|
||||||
|
out = make([]byte, len(k))
|
||||||
|
copy(out, k)
|
||||||
|
it.buffer = append(it.buffer, out)
|
||||||
|
i++
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
glog.Error("Error nexting in database: ", err)
|
||||||
|
it.done = true
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
it.offset++
|
||||||
}
|
}
|
||||||
if !bytes.HasPrefix(out, it.prefix) {
|
if it.Result() == nil {
|
||||||
it.Close()
|
it.done = true
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
it.result = Token(out)
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -121,7 +143,16 @@ func (it *AllIterator) ResultTree() *graph.ResultTree {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Result() graph.Value {
|
func (it *AllIterator) Result() graph.Value {
|
||||||
return it.result
|
if it.done {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if it.result != nil {
|
||||||
|
return it.result
|
||||||
|
}
|
||||||
|
if it.offset >= len(it.buffer) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return &Token{bucket: it.bucket, key: it.buffer[it.offset]}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) NextPath() bool {
|
func (it *AllIterator) NextPath() bool {
|
||||||
|
|
@ -134,29 +165,23 @@ func (it *AllIterator) SubIterators() []graph.Iterator {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Contains(v graph.Value) bool {
|
func (it *AllIterator) Contains(v graph.Value) bool {
|
||||||
it.result = v
|
it.result = v.(*Token)
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Close() {
|
func (it *AllIterator) Close() {
|
||||||
if it.open {
|
it.result = nil
|
||||||
it.iter.Release()
|
it.buffer = nil
|
||||||
it.open = false
|
it.done = true
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Size() (int64, bool) {
|
func (it *AllIterator) Size() (int64, bool) {
|
||||||
size, err := it.ts.SizeOfPrefix(it.prefix)
|
return it.qs.size, true
|
||||||
if err == nil {
|
|
||||||
return size, false
|
|
||||||
}
|
|
||||||
// INT64_MAX
|
|
||||||
return int64(^uint64(0) >> 1), false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) DebugString(indent int) string {
|
func (it *AllIterator) DebugString(indent int) string {
|
||||||
size, _ := it.Size()
|
size, _ := it.Size()
|
||||||
return fmt.Sprintf("%s(%s tags: %v leveldb size:%d %s %p)", strings.Repeat(" ", indent), it.Type(), it.tags.Tags(), size, it.dir, it)
|
return fmt.Sprintf("%s(%s tags: %v bolt size:%d %s %p)", strings.Repeat(" ", indent), it.Type(), it.tags.Tags(), size, it.dir, it)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *AllIterator) Type() graph.Type { return graph.All }
|
func (it *AllIterator) Type() graph.Type { return graph.All }
|
||||||
|
|
|
||||||
|
|
@ -17,12 +17,12 @@ package bolt
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/barakmich/glog"
|
"github.com/barakmich/glog"
|
||||||
ldbit "github.com/syndtr/goleveldb/leveldb/iterator"
|
"github.com/boltdb/bolt"
|
||||||
"github.com/syndtr/goleveldb/leveldb/opt"
|
|
||||||
|
|
||||||
"github.com/google/cayley/graph"
|
"github.com/google/cayley/graph"
|
||||||
"github.com/google/cayley/graph/iterator"
|
"github.com/google/cayley/graph/iterator"
|
||||||
|
|
@ -30,48 +30,38 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
type Iterator struct {
|
type Iterator struct {
|
||||||
uid uint64
|
uid uint64
|
||||||
tags graph.Tagger
|
tags graph.Tagger
|
||||||
nextPrefix []byte
|
bucket []byte
|
||||||
checkId []byte
|
checkId []byte
|
||||||
dir quad.Direction
|
dir quad.Direction
|
||||||
open bool
|
qs *QuadStore
|
||||||
iter ldbit.Iterator
|
result *Token
|
||||||
qs *QuadStore
|
buffer [][]byte
|
||||||
ro *opt.ReadOptions
|
offset int
|
||||||
originalPrefix string
|
done bool
|
||||||
result graph.Value
|
size int64
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewIterator(prefix string, d quad.Direction, value graph.Value, qs *QuadStore) graph.Iterator {
|
var bufferSize = 50
|
||||||
vb := value.(Token)
|
|
||||||
p := make([]byte, 0, 2+qs.hasher.Size())
|
|
||||||
p = append(p, []byte(prefix)...)
|
|
||||||
p = append(p, []byte(vb[1:])...)
|
|
||||||
|
|
||||||
opts := &opt.ReadOptions{
|
func NewIterator(bucket []byte, d quad.Direction, value graph.Value, qs *QuadStore) graph.Iterator {
|
||||||
DontFillCache: true,
|
tok := value.(*Token)
|
||||||
|
if !bytes.Equal(tok.bucket, nodeBucket) {
|
||||||
|
glog.Error("Creating an iterator from a non-node value.")
|
||||||
|
return &iterator.Null{}
|
||||||
}
|
}
|
||||||
|
|
||||||
it := Iterator{
|
it := Iterator{
|
||||||
uid: iterator.NextUID(),
|
uid: iterator.NextUID(),
|
||||||
nextPrefix: p,
|
bucket: bucket,
|
||||||
checkId: vb,
|
dir: d,
|
||||||
dir: d,
|
qs: qs,
|
||||||
originalPrefix: prefix,
|
size: qs.SizeOf(value),
|
||||||
ro: opts,
|
|
||||||
iter: qs.db.NewIterator(nil, opts),
|
|
||||||
open: true,
|
|
||||||
qs: qs,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
ok := it.iter.Seek(it.nextPrefix)
|
it.checkId = make([]byte, len(tok.key))
|
||||||
if !ok {
|
copy(it.checkId, tok.key)
|
||||||
it.open = false
|
|
||||||
it.iter.Release()
|
|
||||||
glog.Error("Opening LevelDB iterator couldn't seek to location ", it.nextPrefix)
|
|
||||||
return &iterator.Null{}
|
|
||||||
}
|
|
||||||
|
|
||||||
return &it
|
return &it
|
||||||
}
|
}
|
||||||
|
|
@ -81,15 +71,9 @@ func (it *Iterator) UID() uint64 {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Reset() {
|
func (it *Iterator) Reset() {
|
||||||
if !it.open {
|
it.buffer = nil
|
||||||
it.iter = it.qs.db.NewIterator(nil, it.ro)
|
it.offset = 0
|
||||||
it.open = true
|
it.done = false
|
||||||
}
|
|
||||||
ok := it.iter.Seek(it.nextPrefix)
|
|
||||||
if !ok {
|
|
||||||
it.open = false
|
|
||||||
it.iter.Release()
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Tagger() *graph.Tagger {
|
func (it *Iterator) Tagger() *graph.Tagger {
|
||||||
|
|
@ -107,16 +91,15 @@ func (it *Iterator) TagResults(dst map[string]graph.Value) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Clone() graph.Iterator {
|
func (it *Iterator) Clone() graph.Iterator {
|
||||||
out := NewIterator(it.originalPrefix, it.dir, Token(it.checkId), it.qs)
|
out := NewIterator(it.bucket, it.dir, &Token{nodeBucket, it.checkId}, it.qs)
|
||||||
out.Tagger().CopyFrom(it)
|
out.Tagger().CopyFrom(it)
|
||||||
return out
|
return out
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Close() {
|
func (it *Iterator) Close() {
|
||||||
if it.open {
|
it.result = nil
|
||||||
it.iter.Release()
|
it.buffer = nil
|
||||||
it.open = false
|
it.done = true
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) isLiveValue(val []byte) bool {
|
func (it *Iterator) isLiveValue(val []byte) bool {
|
||||||
|
|
@ -125,36 +108,73 @@ func (it *Iterator) isLiveValue(val []byte) bool {
|
||||||
return len(entry.History)%2 != 0
|
return len(entry.History)%2 != 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var errNotExist = errors.New("Triple doesn't exist")
|
||||||
|
|
||||||
func (it *Iterator) Next() bool {
|
func (it *Iterator) Next() bool {
|
||||||
if it.iter == nil {
|
if it.done {
|
||||||
it.result = nil
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
if !it.open {
|
if len(it.buffer) <= it.offset+1 {
|
||||||
it.result = nil
|
it.offset = 0
|
||||||
return false
|
var last []byte
|
||||||
}
|
if it.buffer != nil {
|
||||||
if !it.iter.Valid() {
|
last = it.buffer[len(it.buffer)-1]
|
||||||
it.result = nil
|
|
||||||
it.Close()
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if bytes.HasPrefix(it.iter.Key(), it.nextPrefix) {
|
|
||||||
if !it.isLiveValue(it.iter.Value()) {
|
|
||||||
return it.Next()
|
|
||||||
}
|
}
|
||||||
out := make([]byte, len(it.iter.Key()))
|
it.buffer = make([][]byte, 0, bufferSize)
|
||||||
copy(out, it.iter.Key())
|
err := it.qs.db.View(func(tx *bolt.Tx) error {
|
||||||
it.result = Token(out)
|
i := 0
|
||||||
ok := it.iter.Next()
|
b := tx.Bucket(it.bucket)
|
||||||
if !ok {
|
cur := b.Cursor()
|
||||||
it.Close()
|
if last == nil {
|
||||||
|
k, _ := cur.Seek(it.checkId)
|
||||||
|
if bytes.HasPrefix(k, it.checkId) {
|
||||||
|
var out []byte
|
||||||
|
out = make([]byte, len(k))
|
||||||
|
copy(out, k)
|
||||||
|
it.buffer = append(it.buffer, out)
|
||||||
|
i++
|
||||||
|
} else {
|
||||||
|
it.buffer = append(it.buffer, nil)
|
||||||
|
return errNotExist
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
k, _ := cur.Seek(last)
|
||||||
|
if !bytes.Equal(k, last) {
|
||||||
|
return fmt.Errorf("Couldn't pick up after", k)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i < bufferSize {
|
||||||
|
k, v := cur.Next()
|
||||||
|
if k == nil || !bytes.HasPrefix(k, it.checkId) {
|
||||||
|
it.buffer = append(it.buffer, nil)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if !it.isLiveValue(v) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var out []byte
|
||||||
|
out = make([]byte, len(k))
|
||||||
|
copy(out, k)
|
||||||
|
it.buffer = append(it.buffer, out)
|
||||||
|
i++
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
if err != errNotExist {
|
||||||
|
glog.Error("Error nexting in database: ", err)
|
||||||
|
}
|
||||||
|
it.done = true
|
||||||
|
return false
|
||||||
}
|
}
|
||||||
return true
|
} else {
|
||||||
|
it.offset++
|
||||||
}
|
}
|
||||||
it.Close()
|
if it.Result() == nil {
|
||||||
it.result = nil
|
it.done = true
|
||||||
return false
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) ResultTree() *graph.ResultTree {
|
func (it *Iterator) ResultTree() *graph.ResultTree {
|
||||||
|
|
@ -162,7 +182,19 @@ func (it *Iterator) ResultTree() *graph.ResultTree {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Result() graph.Value {
|
func (it *Iterator) Result() graph.Value {
|
||||||
return it.result
|
if it.done {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if it.result != nil {
|
||||||
|
return it.result
|
||||||
|
}
|
||||||
|
if it.offset >= len(it.buffer) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if it.buffer[it.offset] == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return &Token{bucket: it.bucket, key: it.buffer[it.offset]}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) NextPath() bool {
|
func (it *Iterator) NextPath() bool {
|
||||||
|
|
@ -174,65 +206,65 @@ func (it *Iterator) SubIterators() []graph.Iterator {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func PositionOf(prefix []byte, d quad.Direction, qs *QuadStore) int {
|
func PositionOf(tok *Token, d quad.Direction, qs *QuadStore) int {
|
||||||
if bytes.Equal(prefix, []byte("sp")) {
|
if bytes.Equal(tok.bucket, spoBucket) {
|
||||||
switch d {
|
switch d {
|
||||||
case quad.Subject:
|
case quad.Subject:
|
||||||
return 2
|
return 0
|
||||||
case quad.Predicate:
|
case quad.Predicate:
|
||||||
return qs.hasher.Size() + 2
|
return qs.hasher.Size()
|
||||||
case quad.Object:
|
case quad.Object:
|
||||||
return 2*qs.hasher.Size() + 2
|
return 2 * qs.hasher.Size()
|
||||||
case quad.Label:
|
case quad.Label:
|
||||||
return 3*qs.hasher.Size() + 2
|
return 3 * qs.hasher.Size()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if bytes.Equal(prefix, []byte("po")) {
|
if bytes.Equal(tok.bucket, posBucket) {
|
||||||
switch d {
|
switch d {
|
||||||
case quad.Subject:
|
case quad.Subject:
|
||||||
return 2*qs.hasher.Size() + 2
|
return 2 * qs.hasher.Size()
|
||||||
case quad.Predicate:
|
case quad.Predicate:
|
||||||
return 2
|
return 0
|
||||||
case quad.Object:
|
case quad.Object:
|
||||||
return qs.hasher.Size() + 2
|
return qs.hasher.Size()
|
||||||
case quad.Label:
|
case quad.Label:
|
||||||
return 3*qs.hasher.Size() + 2
|
return 3 * qs.hasher.Size()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if bytes.Equal(prefix, []byte("os")) {
|
if bytes.Equal(tok.bucket, ospBucket) {
|
||||||
switch d {
|
switch d {
|
||||||
case quad.Subject:
|
case quad.Subject:
|
||||||
return qs.hasher.Size() + 2
|
return qs.hasher.Size()
|
||||||
case quad.Predicate:
|
case quad.Predicate:
|
||||||
return 2*qs.hasher.Size() + 2
|
return 2 * qs.hasher.Size()
|
||||||
case quad.Object:
|
case quad.Object:
|
||||||
return 2
|
return 0
|
||||||
case quad.Label:
|
case quad.Label:
|
||||||
return 3*qs.hasher.Size() + 2
|
return 3 * qs.hasher.Size()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if bytes.Equal(prefix, []byte("cp")) {
|
if bytes.Equal(tok.bucket, cpsBucket) {
|
||||||
switch d {
|
switch d {
|
||||||
case quad.Subject:
|
case quad.Subject:
|
||||||
return 2*qs.hasher.Size() + 2
|
return 2 * qs.hasher.Size()
|
||||||
case quad.Predicate:
|
case quad.Predicate:
|
||||||
return qs.hasher.Size() + 2
|
return qs.hasher.Size()
|
||||||
case quad.Object:
|
case quad.Object:
|
||||||
return 3*qs.hasher.Size() + 2
|
return 3 * qs.hasher.Size()
|
||||||
case quad.Label:
|
case quad.Label:
|
||||||
return 2
|
return 0
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
panic("unreachable")
|
panic("unreachable")
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Contains(v graph.Value) bool {
|
func (it *Iterator) Contains(v graph.Value) bool {
|
||||||
val := v.(Token)
|
val := v.(*Token)
|
||||||
if val[0] == 'z' {
|
if bytes.Equal(val.bucket, nodeBucket) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
offset := PositionOf(val[0:2], it.dir, it.qs)
|
offset := PositionOf(val, it.dir, it.qs)
|
||||||
if bytes.HasPrefix(val[offset:], it.checkId[1:]) {
|
if bytes.HasPrefix(val.key[offset:], it.checkId) {
|
||||||
// You may ask, why don't we check to see if it's a valid (not deleted) triple
|
// You may ask, why don't we check to see if it's a valid (not deleted) triple
|
||||||
// again?
|
// again?
|
||||||
//
|
//
|
||||||
|
|
@ -248,31 +280,30 @@ func (it *Iterator) Contains(v graph.Value) bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) Size() (int64, bool) {
|
func (it *Iterator) Size() (int64, bool) {
|
||||||
return it.qs.SizeOf(Token(it.checkId)), true
|
return it.size, true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (it *Iterator) DebugString(indent int) string {
|
func (it *Iterator) DebugString(indent int) string {
|
||||||
size, _ := it.Size()
|
|
||||||
return fmt.Sprintf("%s(%s %d tags: %v dir: %s size:%d %s)",
|
return fmt.Sprintf("%s(%s %d tags: %v dir: %s size:%d %s)",
|
||||||
strings.Repeat(" ", indent),
|
strings.Repeat(" ", indent),
|
||||||
it.Type(),
|
it.Type(),
|
||||||
it.UID(),
|
it.UID(),
|
||||||
it.tags.Tags(),
|
it.tags.Tags(),
|
||||||
it.dir,
|
it.dir,
|
||||||
size,
|
it.size,
|
||||||
it.qs.NameOf(Token(it.checkId)),
|
it.qs.NameOf(&Token{it.bucket, it.checkId}),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
var levelDBType graph.Type
|
var boltType graph.Type
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
levelDBType = graph.RegisterIterator("leveldb")
|
boltType = graph.RegisterIterator("bolt")
|
||||||
}
|
}
|
||||||
|
|
||||||
func Type() graph.Type { return levelDBType }
|
func Type() graph.Type { return boltType }
|
||||||
|
|
||||||
func (it *Iterator) Type() graph.Type { return levelDBType }
|
func (it *Iterator) Type() graph.Type { return boltType }
|
||||||
func (it *Iterator) Sorted() bool { return false }
|
func (it *Iterator) Sorted() bool { return false }
|
||||||
|
|
||||||
func (it *Iterator) Optimize() (graph.Iterator, bool) {
|
func (it *Iterator) Optimize() (graph.Iterator, bool) {
|
||||||
|
|
|
||||||
|
|
@ -21,9 +21,11 @@ import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"hash"
|
"hash"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/barakmich/glog"
|
"github.com/barakmich/glog"
|
||||||
"github.com/boltdb/bolt"
|
"github.com/boltdb/bolt"
|
||||||
|
"github.com/boltdb/coalescer"
|
||||||
|
|
||||||
"github.com/google/cayley/graph"
|
"github.com/google/cayley/graph"
|
||||||
"github.com/google/cayley/graph/iterator"
|
"github.com/google/cayley/graph/iterator"
|
||||||
|
|
@ -31,7 +33,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
graph.RegisterTripleStore("bolt", true, newQuadStore, createNewLevelDB)
|
graph.RegisterTripleStore("bolt", true, newQuadStore, createNewBolt)
|
||||||
}
|
}
|
||||||
|
|
||||||
type Token struct {
|
type Token struct {
|
||||||
|
|
@ -40,7 +42,7 @@ type Token struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *Token) Key() interface{} {
|
func (t *Token) Key() interface{} {
|
||||||
return fmt.Sprint(t.bucket, t.data)
|
return fmt.Sprint(t.bucket, t.key)
|
||||||
}
|
}
|
||||||
|
|
||||||
type QuadStore struct {
|
type QuadStore struct {
|
||||||
|
|
@ -53,7 +55,6 @@ type QuadStore struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func createNewBolt(path string, _ graph.Options) error {
|
func createNewBolt(path string, _ graph.Options) error {
|
||||||
opts := &opt.Options{}
|
|
||||||
db, err := bolt.Open(path, 0600, nil)
|
db, err := bolt.Open(path, 0600, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.Errorf("Error: couldn't create Bolt database: %v", err)
|
glog.Errorf("Error: couldn't create Bolt database: %v", err)
|
||||||
|
|
@ -62,9 +63,6 @@ func createNewBolt(path string, _ graph.Options) error {
|
||||||
defer db.Close()
|
defer db.Close()
|
||||||
qs := &QuadStore{}
|
qs := &QuadStore{}
|
||||||
qs.db = db
|
qs.db = db
|
||||||
qs.writeopts = &opt.WriteOptions{
|
|
||||||
Sync: true,
|
|
||||||
}
|
|
||||||
err = qs.createBuckets()
|
err = qs.createBuckets()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
|
@ -91,10 +89,10 @@ func newQuadStore(path string, options graph.Options) (graph.TripleStore, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) createBuckets() error {
|
func (qs *QuadStore) createBuckets() error {
|
||||||
return db.Update(func(tx *bolt.Tx) error {
|
return qs.db.Update(func(tx *bolt.Tx) error {
|
||||||
var err error
|
var err error
|
||||||
for _, bucket := range [][]byte{spo, osp, pos, cps} {
|
for _, index := range [][4]quad.Direction{spo, osp, pos, cps} {
|
||||||
_, err = tx.CreateBucket(bucketFor(bucket))
|
_, err = tx.CreateBucket(bucketFor(index))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Couldn't create bucket: %s", err)
|
return fmt.Errorf("Couldn't create bucket: %s", err)
|
||||||
}
|
}
|
||||||
|
|
@ -111,6 +109,7 @@ func (qs *QuadStore) createBuckets() error {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("Couldn't create bucket: %s", err)
|
return fmt.Errorf("Couldn't create bucket: %s", err)
|
||||||
}
|
}
|
||||||
|
return nil
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -152,10 +151,14 @@ type IndexEntry struct {
|
||||||
|
|
||||||
// Short hand for direction permutations.
|
// Short hand for direction permutations.
|
||||||
var (
|
var (
|
||||||
spo = bucketFor([4]quad.Direction{quad.Subject, quad.Predicate, quad.Object, quad.Label})
|
spo = [4]quad.Direction{quad.Subject, quad.Predicate, quad.Object, quad.Label}
|
||||||
osp = bucketFor([4]quad.Direction{quad.Object, quad.Subject, quad.Predicate, quad.Label})
|
osp = [4]quad.Direction{quad.Object, quad.Subject, quad.Predicate, quad.Label}
|
||||||
pos = bucketFor([4]quad.Direction{quad.Predicate, quad.Object, quad.Subject, quad.Label})
|
pos = [4]quad.Direction{quad.Predicate, quad.Object, quad.Subject, quad.Label}
|
||||||
cps = bucketFor([4]quad.Direction{quad.Label, quad.Predicate, quad.Subject, quad.Object})
|
cps = [4]quad.Direction{quad.Label, quad.Predicate, quad.Subject, quad.Object}
|
||||||
|
spoBucket = bucketFor(spo)
|
||||||
|
ospBucket = bucketFor(osp)
|
||||||
|
posBucket = bucketFor(pos)
|
||||||
|
cpsBucket = bucketFor(cps)
|
||||||
)
|
)
|
||||||
|
|
||||||
var logBucket = []byte("log")
|
var logBucket = []byte("log")
|
||||||
|
|
@ -163,12 +166,14 @@ var nodeBucket = []byte("node")
|
||||||
var metaBucket = []byte("meta")
|
var metaBucket = []byte("meta")
|
||||||
|
|
||||||
func (qs *QuadStore) ApplyDeltas(deltas []*graph.Delta) error {
|
func (qs *QuadStore) ApplyDeltas(deltas []*graph.Delta) error {
|
||||||
batch := &leveldb.Batch{}
|
|
||||||
var size_change int64
|
var size_change int64
|
||||||
var new_horizon int64
|
var new_horizon int64
|
||||||
err := qs.db.Update(func(tx *bolt.Tx) error {
|
c, err := coalescer.New(qs.db, 200, 100*time.Millisecond)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
err = c.Update(func(tx *bolt.Tx) error {
|
||||||
var b *bolt.Bucket
|
var b *bolt.Bucket
|
||||||
var err error
|
|
||||||
resizeMap := make(map[string]int64)
|
resizeMap := make(map[string]int64)
|
||||||
size_change = int64(0)
|
size_change = int64(0)
|
||||||
for _, d := range deltas {
|
for _, d := range deltas {
|
||||||
|
|
@ -220,12 +225,11 @@ func (qs *QuadStore) ApplyDeltas(deltas []*graph.Delta) error {
|
||||||
|
|
||||||
func (qs *QuadStore) buildQuadWrite(tx *bolt.Tx, q quad.Quad, id int64, isAdd bool) error {
|
func (qs *QuadStore) buildQuadWrite(tx *bolt.Tx, q quad.Quad, id int64, isAdd bool) error {
|
||||||
var entry IndexEntry
|
var entry IndexEntry
|
||||||
b := tx.Bucket(bucketFor(spo))
|
b := tx.Bucket(spoBucket)
|
||||||
|
|
||||||
data := b.Get(qs.createKeyFor(spo, q))
|
data := b.Get(qs.createKeyFor(spo, q))
|
||||||
if data != nil {
|
if data != nil {
|
||||||
// We got something.
|
// We got something.
|
||||||
err = json.Unmarshal(data, &entry)
|
err := json.Unmarshal(data, &entry)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
@ -244,17 +248,17 @@ func (qs *QuadStore) buildQuadWrite(tx *bolt.Tx, q quad.Quad, id int64, isAdd bo
|
||||||
|
|
||||||
entry.History = append(entry.History, id)
|
entry.History = append(entry.History, id)
|
||||||
|
|
||||||
bytes, err := json.Marshal(entry)
|
jsonbytes, err := json.Marshal(entry)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.Errorf("Couldn't write to buffer for entry %#v: %s", entry, err)
|
glog.Errorf("Couldn't write to buffer for entry %#v: %s", entry, err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
for _, bucket := range [][4]quad.Direction{spo, osp, pos, cps} {
|
for _, index := range [][4]quad.Direction{spo, osp, pos, cps} {
|
||||||
if bucket == cps && q.Get(quad.Label) == "" {
|
if index == cps && q.Get(quad.Label) == "" {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
b := tx.Bucket(bucketFor(bucket))
|
b := tx.Bucket(bucketFor(index))
|
||||||
err = b.Put(qs.createKeyFor(bucket, q), bytes)
|
err = b.Put(qs.createKeyFor(index, q), jsonbytes)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
@ -268,14 +272,14 @@ type ValueData struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) UpdateValueKeyBy(name string, amount int64, tx *bolt.Tx) error {
|
func (qs *QuadStore) UpdateValueKeyBy(name string, amount int64, tx *bolt.Tx) error {
|
||||||
value := &ValueData{name, amount}
|
value := ValueData{name, amount}
|
||||||
b := tx.Bucket(nodeBucket)
|
b := tx.Bucket(nodeBucket)
|
||||||
key := qs.createValueKeyFor(name)
|
key := qs.createValueKeyFor(name)
|
||||||
data := b.Get(key)
|
data := b.Get(key)
|
||||||
|
|
||||||
if data != nil {
|
if data != nil {
|
||||||
// Node exists in the database -- unmarshal and update.
|
// Node exists in the database -- unmarshal and update.
|
||||||
err = json.Unmarshal(b, value)
|
err := json.Unmarshal(data, &value)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.Errorf("Error: couldn't reconstruct value: %v", err)
|
glog.Errorf("Error: couldn't reconstruct value: %v", err)
|
||||||
return err
|
return err
|
||||||
|
|
@ -336,13 +340,12 @@ func (qs *QuadStore) Quad(k graph.Value) quad.Quad {
|
||||||
tok := k.(*Token)
|
tok := k.(*Token)
|
||||||
err := qs.db.View(func(tx *bolt.Tx) error {
|
err := qs.db.View(func(tx *bolt.Tx) error {
|
||||||
b := tx.Bucket(tok.bucket)
|
b := tx.Bucket(tok.bucket)
|
||||||
data := qs.db.Get(tok.key, qs.readopts)
|
data := b.Get(tok.key)
|
||||||
if data == nil {
|
if data == nil {
|
||||||
// No harm, no foul.
|
// No harm, no foul.
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
err = json.Unmarshal(data, &q)
|
return json.Unmarshal(data, &q)
|
||||||
return err
|
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
glog.Error("Error getting triple: ", err)
|
glog.Error("Error getting triple: ", err)
|
||||||
|
|
@ -409,7 +412,7 @@ func (qs *QuadStore) getInt64ForKey(tx *bolt.Tx, key string, empty int64) (int64
|
||||||
return empty, nil
|
return empty, nil
|
||||||
}
|
}
|
||||||
buf := bytes.NewBuffer(data)
|
buf := bytes.NewBuffer(data)
|
||||||
err = binary.Read(buf, binary.LittleEndian, &out)
|
err := binary.Read(buf, binary.LittleEndian, &out)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
|
@ -419,31 +422,31 @@ func (qs *QuadStore) getInt64ForKey(tx *bolt.Tx, key string, empty int64) (int64
|
||||||
func (qs *QuadStore) getMetadata() error {
|
func (qs *QuadStore) getMetadata() error {
|
||||||
err := qs.db.View(func(tx *bolt.Tx) error {
|
err := qs.db.View(func(tx *bolt.Tx) error {
|
||||||
var err error
|
var err error
|
||||||
qs.size, err = qs.getInt64ForKey("size", 0)
|
qs.size, err = qs.getInt64ForKey(tx, "size", 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
qs.horizon, err = qs.getInt64ForKey("horizon", 0)
|
qs.horizon, err = qs.getInt64ForKey(tx, "horizon", 0)
|
||||||
return err
|
return err
|
||||||
})
|
})
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) TripleIterator(d quad.Direction, val graph.Value) graph.Iterator {
|
func (qs *QuadStore) TripleIterator(d quad.Direction, val graph.Value) graph.Iterator {
|
||||||
var prefix []byte
|
var bucket []byte
|
||||||
switch d {
|
switch d {
|
||||||
case quad.Subject:
|
case quad.Subject:
|
||||||
prefix = spo
|
bucket = spoBucket
|
||||||
case quad.Predicate:
|
case quad.Predicate:
|
||||||
prefix = pos
|
bucket = posBucket
|
||||||
case quad.Object:
|
case quad.Object:
|
||||||
prefix = osp
|
bucket = ospBucket
|
||||||
case quad.Label:
|
case quad.Label:
|
||||||
prefix = cps
|
bucket = cpsBucket
|
||||||
default:
|
default:
|
||||||
panic("unreachable " + d.String())
|
panic("unreachable " + d.String())
|
||||||
}
|
}
|
||||||
return NewIterator(prefix, d, val, qs)
|
return NewIterator(bucket, d, val, qs)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) NodesAllIterator() graph.Iterator {
|
func (qs *QuadStore) NodesAllIterator() graph.Iterator {
|
||||||
|
|
@ -451,7 +454,7 @@ func (qs *QuadStore) NodesAllIterator() graph.Iterator {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) TriplesAllIterator() graph.Iterator {
|
func (qs *QuadStore) TriplesAllIterator() graph.Iterator {
|
||||||
return NewAllIterator(pos, quad.Predicate, qs)
|
return NewAllIterator(posBucket, quad.Predicate, qs)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) TripleDirection(val graph.Value, d quad.Direction) graph.Value {
|
func (qs *QuadStore) TripleDirection(val graph.Value, d quad.Direction) graph.Value {
|
||||||
|
|
@ -460,7 +463,7 @@ func (qs *QuadStore) TripleDirection(val graph.Value, d quad.Direction) graph.Va
|
||||||
if offset != -1 {
|
if offset != -1 {
|
||||||
return &Token{
|
return &Token{
|
||||||
bucket: nodeBucket,
|
bucket: nodeBucket,
|
||||||
key: v[offset : offset+qs.hasher.Size()],
|
key: v.key[offset : offset+qs.hasher.Size()],
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
return qs.ValueOf(qs.Quad(v).Get(d))
|
return qs.ValueOf(qs.Quad(v).Get(d))
|
||||||
|
|
@ -470,7 +473,7 @@ func (qs *QuadStore) TripleDirection(val graph.Value, d quad.Direction) graph.Va
|
||||||
func compareTokens(a, b graph.Value) bool {
|
func compareTokens(a, b graph.Value) bool {
|
||||||
atok := a.(*Token)
|
atok := a.(*Token)
|
||||||
btok := b.(*Token)
|
btok := b.(*Token)
|
||||||
return bytes.Equal(atok.key, btok.key) && atok.bucket == btok.bucket
|
return bytes.Equal(atok.key, btok.key) && bytes.Equal(atok.bucket, btok.bucket)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (qs *QuadStore) FixedIterator() graph.FixedIterator {
|
func (qs *QuadStore) FixedIterator() graph.FixedIterator {
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue