summaryrefslogtreecommitdiffstats
path: root/vendor/gitea.com/lunny/nodb/t_zset.go
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/gitea.com/lunny/nodb/t_zset.go')
-rw-r--r--vendor/gitea.com/lunny/nodb/t_zset.go943
1 files changed, 943 insertions, 0 deletions
diff --git a/vendor/gitea.com/lunny/nodb/t_zset.go b/vendor/gitea.com/lunny/nodb/t_zset.go
new file mode 100644
index 0000000000..91e30049e0
--- /dev/null
+++ b/vendor/gitea.com/lunny/nodb/t_zset.go
@@ -0,0 +1,943 @@
+package nodb
+
+import (
+ "bytes"
+ "encoding/binary"
+ "errors"
+ "time"
+
+ "gitea.com/lunny/nodb/store"
+)
+
+const (
+ MinScore int64 = -1<<63 + 1
+ MaxScore int64 = 1<<63 - 1
+ InvalidScore int64 = -1 << 63
+
+ AggregateSum byte = 0
+ AggregateMin byte = 1
+ AggregateMax byte = 2
+)
+
+type ScorePair struct {
+ Score int64
+ Member []byte
+}
+
+var errZSizeKey = errors.New("invalid zsize key")
+var errZSetKey = errors.New("invalid zset key")
+var errZScoreKey = errors.New("invalid zscore key")
+var errScoreOverflow = errors.New("zset score overflow")
+var errInvalidAggregate = errors.New("invalid aggregate")
+var errInvalidWeightNum = errors.New("invalid weight number")
+var errInvalidSrcKeyNum = errors.New("invalid src key number")
+
+const (
+ zsetNScoreSep byte = '<'
+ zsetPScoreSep byte = zsetNScoreSep + 1
+ zsetStopScoreSep byte = zsetPScoreSep + 1
+
+ zsetStartMemSep byte = ':'
+ zsetStopMemSep byte = zsetStartMemSep + 1
+)
+
+func checkZSetKMSize(key []byte, member []byte) error {
+ if len(key) > MaxKeySize || len(key) == 0 {
+ return errKeySize
+ } else if len(member) > MaxZSetMemberSize || len(member) == 0 {
+ return errZSetMemberSize
+ }
+ return nil
+}
+
+func (db *DB) zEncodeSizeKey(key []byte) []byte {
+ buf := make([]byte, len(key)+2)
+ buf[0] = db.index
+ buf[1] = ZSizeType
+
+ copy(buf[2:], key)
+ return buf
+}
+
+func (db *DB) zDecodeSizeKey(ek []byte) ([]byte, error) {
+ if len(ek) < 2 || ek[0] != db.index || ek[1] != ZSizeType {
+ return nil, errZSizeKey
+ }
+
+ return ek[2:], nil
+}
+
+func (db *DB) zEncodeSetKey(key []byte, member []byte) []byte {
+ buf := make([]byte, len(key)+len(member)+5)
+
+ pos := 0
+ buf[pos] = db.index
+ pos++
+
+ buf[pos] = ZSetType
+ pos++
+
+ binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
+ pos += 2
+
+ copy(buf[pos:], key)
+ pos += len(key)
+
+ buf[pos] = zsetStartMemSep
+ pos++
+
+ copy(buf[pos:], member)
+
+ return buf
+}
+
+func (db *DB) zDecodeSetKey(ek []byte) ([]byte, []byte, error) {
+ if len(ek) < 5 || ek[0] != db.index || ek[1] != ZSetType {
+ return nil, nil, errZSetKey
+ }
+
+ keyLen := int(binary.BigEndian.Uint16(ek[2:]))
+ if keyLen+5 > len(ek) {
+ return nil, nil, errZSetKey
+ }
+
+ key := ek[4 : 4+keyLen]
+
+ if ek[4+keyLen] != zsetStartMemSep {
+ return nil, nil, errZSetKey
+ }
+
+ member := ek[5+keyLen:]
+ return key, member, nil
+}
+
+func (db *DB) zEncodeStartSetKey(key []byte) []byte {
+ k := db.zEncodeSetKey(key, nil)
+ return k
+}
+
+func (db *DB) zEncodeStopSetKey(key []byte) []byte {
+ k := db.zEncodeSetKey(key, nil)
+ k[len(k)-1] = zsetStartMemSep + 1
+ return k
+}
+
+func (db *DB) zEncodeScoreKey(key []byte, member []byte, score int64) []byte {
+ buf := make([]byte, len(key)+len(member)+14)
+
+ pos := 0
+ buf[pos] = db.index
+ pos++
+
+ buf[pos] = ZScoreType
+ pos++
+
+ binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
+ pos += 2
+
+ copy(buf[pos:], key)
+ pos += len(key)
+
+ if score < 0 {
+ buf[pos] = zsetNScoreSep
+ } else {
+ buf[pos] = zsetPScoreSep
+ }
+
+ pos++
+ binary.BigEndian.PutUint64(buf[pos:], uint64(score))
+ pos += 8
+
+ buf[pos] = zsetStartMemSep
+ pos++
+
+ copy(buf[pos:], member)
+ return buf
+}
+
+func (db *DB) zEncodeStartScoreKey(key []byte, score int64) []byte {
+ return db.zEncodeScoreKey(key, nil, score)
+}
+
+func (db *DB) zEncodeStopScoreKey(key []byte, score int64) []byte {
+ k := db.zEncodeScoreKey(key, nil, score)
+ k[len(k)-1] = zsetStopMemSep
+ return k
+}
+
+func (db *DB) zDecodeScoreKey(ek []byte) (key []byte, member []byte, score int64, err error) {
+ if len(ek) < 14 || ek[0] != db.index || ek[1] != ZScoreType {
+ err = errZScoreKey
+ return
+ }
+
+ keyLen := int(binary.BigEndian.Uint16(ek[2:]))
+ if keyLen+14 > len(ek) {
+ err = errZScoreKey
+ return
+ }
+
+ key = ek[4 : 4+keyLen]
+ pos := 4 + keyLen
+
+ if (ek[pos] != zsetNScoreSep) && (ek[pos] != zsetPScoreSep) {
+ err = errZScoreKey
+ return
+ }
+ pos++
+
+ score = int64(binary.BigEndian.Uint64(ek[pos:]))
+ pos += 8
+
+ if ek[pos] != zsetStartMemSep {
+ err = errZScoreKey
+ return
+ }
+
+ pos++
+
+ member = ek[pos:]
+ return
+}
+
+func (db *DB) zSetItem(t *batch, key []byte, score int64, member []byte) (int64, error) {
+ if score <= MinScore || score >= MaxScore {
+ return 0, errScoreOverflow
+ }
+
+ var exists int64 = 0
+ ek := db.zEncodeSetKey(key, member)
+
+ if v, err := db.bucket.Get(ek); err != nil {
+ return 0, err
+ } else if v != nil {
+ exists = 1
+
+ if s, err := Int64(v, err); err != nil {
+ return 0, err
+ } else {
+ sk := db.zEncodeScoreKey(key, member, s)
+ t.Delete(sk)
+ }
+ }
+
+ t.Put(ek, PutInt64(score))
+
+ sk := db.zEncodeScoreKey(key, member, score)
+ t.Put(sk, []byte{})
+
+ return exists, nil
+}
+
+func (db *DB) zDelItem(t *batch, key []byte, member []byte, skipDelScore bool) (int64, error) {
+ ek := db.zEncodeSetKey(key, member)
+ if v, err := db.bucket.Get(ek); err != nil {
+ return 0, err
+ } else if v == nil {
+ //not exists
+ return 0, nil
+ } else {
+ //exists
+ if !skipDelScore {
+ //we must del score
+ if s, err := Int64(v, err); err != nil {
+ return 0, err
+ } else {
+ sk := db.zEncodeScoreKey(key, member, s)
+ t.Delete(sk)
+ }
+ }
+ }
+
+ t.Delete(ek)
+
+ return 1, nil
+}
+
+func (db *DB) zDelete(t *batch, key []byte) int64 {
+ delMembCnt, _ := db.zRemRange(t, key, MinScore, MaxScore, 0, -1)
+ // todo : log err
+ return delMembCnt
+}
+
+func (db *DB) zExpireAt(key []byte, when int64) (int64, error) {
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ if zcnt, err := db.ZCard(key); err != nil || zcnt == 0 {
+ return 0, err
+ } else {
+ db.expireAt(t, ZSetType, key, when)
+ if err := t.Commit(); err != nil {
+ return 0, err
+ }
+ }
+ return 1, nil
+}
+
+func (db *DB) ZAdd(key []byte, args ...ScorePair) (int64, error) {
+ if len(args) == 0 {
+ return 0, nil
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ var num int64 = 0
+ for i := 0; i < len(args); i++ {
+ score := args[i].Score
+ member := args[i].Member
+
+ if err := checkZSetKMSize(key, member); err != nil {
+ return 0, err
+ }
+
+ if n, err := db.zSetItem(t, key, score, member); err != nil {
+ return 0, err
+ } else if n == 0 {
+ //add new
+ num++
+ }
+ }
+
+ if _, err := db.zIncrSize(t, key, num); err != nil {
+ return 0, err
+ }
+
+ //todo add binlog
+ err := t.Commit()
+ return num, err
+}
+
+func (db *DB) zIncrSize(t *batch, key []byte, delta int64) (int64, error) {
+ sk := db.zEncodeSizeKey(key)
+
+ size, err := Int64(db.bucket.Get(sk))
+ if err != nil {
+ return 0, err
+ } else {
+ size += delta
+ if size <= 0 {
+ size = 0
+ t.Delete(sk)
+ db.rmExpire(t, ZSetType, key)
+ } else {
+ t.Put(sk, PutInt64(size))
+ }
+ }
+
+ return size, nil
+}
+
+func (db *DB) ZCard(key []byte) (int64, error) {
+ if err := checkKeySize(key); err != nil {
+ return 0, err
+ }
+
+ sk := db.zEncodeSizeKey(key)
+ return Int64(db.bucket.Get(sk))
+}
+
+func (db *DB) ZScore(key []byte, member []byte) (int64, error) {
+ if err := checkZSetKMSize(key, member); err != nil {
+ return InvalidScore, err
+ }
+
+ var score int64 = InvalidScore
+
+ k := db.zEncodeSetKey(key, member)
+ if v, err := db.bucket.Get(k); err != nil {
+ return InvalidScore, err
+ } else if v == nil {
+ return InvalidScore, ErrScoreMiss
+ } else {
+ if score, err = Int64(v, nil); err != nil {
+ return InvalidScore, err
+ }
+ }
+
+ return score, nil
+}
+
+func (db *DB) ZRem(key []byte, members ...[]byte) (int64, error) {
+ if len(members) == 0 {
+ return 0, nil
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ var num int64 = 0
+ for i := 0; i < len(members); i++ {
+ if err := checkZSetKMSize(key, members[i]); err != nil {
+ return 0, err
+ }
+
+ if n, err := db.zDelItem(t, key, members[i], false); err != nil {
+ return 0, err
+ } else if n == 1 {
+ num++
+ }
+ }
+
+ if _, err := db.zIncrSize(t, key, -num); err != nil {
+ return 0, err
+ }
+
+ err := t.Commit()
+ return num, err
+}
+
+func (db *DB) ZIncrBy(key []byte, delta int64, member []byte) (int64, error) {
+ if err := checkZSetKMSize(key, member); err != nil {
+ return InvalidScore, err
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ ek := db.zEncodeSetKey(key, member)
+
+ var oldScore int64 = 0
+ v, err := db.bucket.Get(ek)
+ if err != nil {
+ return InvalidScore, err
+ } else if v == nil {
+ db.zIncrSize(t, key, 1)
+ } else {
+ if oldScore, err = Int64(v, err); err != nil {
+ return InvalidScore, err
+ }
+ }
+
+ newScore := oldScore + delta
+ if newScore >= MaxScore || newScore <= MinScore {
+ return InvalidScore, errScoreOverflow
+ }
+
+ sk := db.zEncodeScoreKey(key, member, newScore)
+ t.Put(sk, []byte{})
+ t.Put(ek, PutInt64(newScore))
+
+ if v != nil {
+ // so as to update score, we must delete the old one
+ oldSk := db.zEncodeScoreKey(key, member, oldScore)
+ t.Delete(oldSk)
+ }
+
+ err = t.Commit()
+ return newScore, err
+}
+
+func (db *DB) ZCount(key []byte, min int64, max int64) (int64, error) {
+ if err := checkKeySize(key); err != nil {
+ return 0, err
+ }
+ minKey := db.zEncodeStartScoreKey(key, min)
+ maxKey := db.zEncodeStopScoreKey(key, max)
+
+ rangeType := store.RangeROpen
+
+ it := db.bucket.RangeLimitIterator(minKey, maxKey, rangeType, 0, -1)
+ var n int64 = 0
+ for ; it.Valid(); it.Next() {
+ n++
+ }
+ it.Close()
+
+ return n, nil
+}
+
+func (db *DB) zrank(key []byte, member []byte, reverse bool) (int64, error) {
+ if err := checkZSetKMSize(key, member); err != nil {
+ return 0, err
+ }
+
+ k := db.zEncodeSetKey(key, member)
+
+ it := db.bucket.NewIterator()
+ defer it.Close()
+
+ if v := it.Find(k); v == nil {
+ return -1, nil
+ } else {
+ if s, err := Int64(v, nil); err != nil {
+ return 0, err
+ } else {
+ var rit *store.RangeLimitIterator
+
+ sk := db.zEncodeScoreKey(key, member, s)
+
+ if !reverse {
+ minKey := db.zEncodeStartScoreKey(key, MinScore)
+
+ rit = store.NewRangeIterator(it, &store.Range{minKey, sk, store.RangeClose})
+ } else {
+ maxKey := db.zEncodeStopScoreKey(key, MaxScore)
+ rit = store.NewRevRangeIterator(it, &store.Range{sk, maxKey, store.RangeClose})
+ }
+
+ var lastKey []byte = nil
+ var n int64 = 0
+
+ for ; rit.Valid(); rit.Next() {
+ n++
+
+ lastKey = rit.BufKey(lastKey)
+ }
+
+ if _, m, _, err := db.zDecodeScoreKey(lastKey); err == nil && bytes.Equal(m, member) {
+ n--
+ return n, nil
+ }
+ }
+ }
+
+ return -1, nil
+}
+
+func (db *DB) zIterator(key []byte, min int64, max int64, offset int, count int, reverse bool) *store.RangeLimitIterator {
+ minKey := db.zEncodeStartScoreKey(key, min)
+ maxKey := db.zEncodeStopScoreKey(key, max)
+
+ if !reverse {
+ return db.bucket.RangeLimitIterator(minKey, maxKey, store.RangeClose, offset, count)
+ } else {
+ return db.bucket.RevRangeLimitIterator(minKey, maxKey, store.RangeClose, offset, count)
+ }
+}
+
+func (db *DB) zRemRange(t *batch, key []byte, min int64, max int64, offset int, count int) (int64, error) {
+ if len(key) > MaxKeySize {
+ return 0, errKeySize
+ }
+
+ it := db.zIterator(key, min, max, offset, count, false)
+ var num int64 = 0
+ for ; it.Valid(); it.Next() {
+ sk := it.RawKey()
+ _, m, _, err := db.zDecodeScoreKey(sk)
+ if err != nil {
+ continue
+ }
+
+ if n, err := db.zDelItem(t, key, m, true); err != nil {
+ return 0, err
+ } else if n == 1 {
+ num++
+ }
+
+ t.Delete(sk)
+ }
+ it.Close()
+
+ if _, err := db.zIncrSize(t, key, -num); err != nil {
+ return 0, err
+ }
+
+ return num, nil
+}
+
+func (db *DB) zRange(key []byte, min int64, max int64, offset int, count int, reverse bool) ([]ScorePair, error) {
+ if len(key) > MaxKeySize {
+ return nil, errKeySize
+ }
+
+ if offset < 0 {
+ return []ScorePair{}, nil
+ }
+
+ nv := 64
+ if count > 0 {
+ nv = count
+ }
+
+ v := make([]ScorePair, 0, nv)
+
+ var it *store.RangeLimitIterator
+
+ //if reverse and offset is 0, count < 0, we may use forward iterator then reverse
+ //because store iterator prev is slower than next
+ if !reverse || (offset == 0 && count < 0) {
+ it = db.zIterator(key, min, max, offset, count, false)
+ } else {
+ it = db.zIterator(key, min, max, offset, count, true)
+ }
+
+ for ; it.Valid(); it.Next() {
+ _, m, s, err := db.zDecodeScoreKey(it.Key())
+ //may be we will check key equal?
+ if err != nil {
+ continue
+ }
+
+ v = append(v, ScorePair{Member: m, Score: s})
+ }
+ it.Close()
+
+ if reverse && (offset == 0 && count < 0) {
+ for i, j := 0, len(v)-1; i < j; i, j = i+1, j-1 {
+ v[i], v[j] = v[j], v[i]
+ }
+ }
+
+ return v, nil
+}
+
+func (db *DB) zParseLimit(key []byte, start int, stop int) (offset int, count int, err error) {
+ if start < 0 || stop < 0 {
+ //refer redis implementation
+ var size int64
+ size, err = db.ZCard(key)
+ if err != nil {
+ return
+ }
+
+ llen := int(size)
+
+ if start < 0 {
+ start = llen + start
+ }
+ if stop < 0 {
+ stop = llen + stop
+ }
+
+ if start < 0 {
+ start = 0
+ }
+
+ if start >= llen {
+ offset = -1
+ return
+ }
+ }
+
+ if start > stop {
+ offset = -1
+ return
+ }
+
+ offset = start
+ count = (stop - start) + 1
+ return
+}
+
+func (db *DB) ZClear(key []byte) (int64, error) {
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ rmCnt, err := db.zRemRange(t, key, MinScore, MaxScore, 0, -1)
+ if err == nil {
+ err = t.Commit()
+ }
+
+ return rmCnt, err
+}
+
+func (db *DB) ZMclear(keys ...[]byte) (int64, error) {
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ for _, key := range keys {
+ if _, err := db.zRemRange(t, key, MinScore, MaxScore, 0, -1); err != nil {
+ return 0, err
+ }
+ }
+
+ err := t.Commit()
+
+ return int64(len(keys)), err
+}
+
+func (db *DB) ZRange(key []byte, start int, stop int) ([]ScorePair, error) {
+ return db.ZRangeGeneric(key, start, stop, false)
+}
+
+//min and max must be inclusive
+//if no limit, set offset = 0 and count = -1
+func (db *DB) ZRangeByScore(key []byte, min int64, max int64,
+ offset int, count int) ([]ScorePair, error) {
+ return db.ZRangeByScoreGeneric(key, min, max, offset, count, false)
+}
+
+func (db *DB) ZRank(key []byte, member []byte) (int64, error) {
+ return db.zrank(key, member, false)
+}
+
+func (db *DB) ZRemRangeByRank(key []byte, start int, stop int) (int64, error) {
+ offset, count, err := db.zParseLimit(key, start, stop)
+ if err != nil {
+ return 0, err
+ }
+
+ var rmCnt int64
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ rmCnt, err = db.zRemRange(t, key, MinScore, MaxScore, offset, count)
+ if err == nil {
+ err = t.Commit()
+ }
+
+ return rmCnt, err
+}
+
+//min and max must be inclusive
+func (db *DB) ZRemRangeByScore(key []byte, min int64, max int64) (int64, error) {
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ rmCnt, err := db.zRemRange(t, key, min, max, 0, -1)
+ if err == nil {
+ err = t.Commit()
+ }
+
+ return rmCnt, err
+}
+
+func (db *DB) ZRevRange(key []byte, start int, stop int) ([]ScorePair, error) {
+ return db.ZRangeGeneric(key, start, stop, true)
+}
+
+func (db *DB) ZRevRank(key []byte, member []byte) (int64, error) {
+ return db.zrank(key, member, true)
+}
+
+//min and max must be inclusive
+//if no limit, set offset = 0 and count = -1
+func (db *DB) ZRevRangeByScore(key []byte, min int64, max int64, offset int, count int) ([]ScorePair, error) {
+ return db.ZRangeByScoreGeneric(key, min, max, offset, count, true)
+}
+
+func (db *DB) ZRangeGeneric(key []byte, start int, stop int, reverse bool) ([]ScorePair, error) {
+ offset, count, err := db.zParseLimit(key, start, stop)
+ if err != nil {
+ return nil, err
+ }
+
+ return db.zRange(key, MinScore, MaxScore, offset, count, reverse)
+}
+
+//min and max must be inclusive
+//if no limit, set offset = 0 and count = -1
+func (db *DB) ZRangeByScoreGeneric(key []byte, min int64, max int64,
+ offset int, count int, reverse bool) ([]ScorePair, error) {
+
+ return db.zRange(key, min, max, offset, count, reverse)
+}
+
+func (db *DB) zFlush() (drop int64, err error) {
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+ return db.flushType(t, ZSetType)
+}
+
+func (db *DB) ZExpire(key []byte, duration int64) (int64, error) {
+ if duration <= 0 {
+ return 0, errExpireValue
+ }
+
+ return db.zExpireAt(key, time.Now().Unix()+duration)
+}
+
+func (db *DB) ZExpireAt(key []byte, when int64) (int64, error) {
+ if when <= time.Now().Unix() {
+ return 0, errExpireValue
+ }
+
+ return db.zExpireAt(key, when)
+}
+
+func (db *DB) ZTTL(key []byte) (int64, error) {
+ if err := checkKeySize(key); err != nil {
+ return -1, err
+ }
+
+ return db.ttl(ZSetType, key)
+}
+
+func (db *DB) ZPersist(key []byte) (int64, error) {
+ if err := checkKeySize(key); err != nil {
+ return 0, err
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ n, err := db.rmExpire(t, ZSetType, key)
+ if err != nil {
+ return 0, err
+ }
+
+ err = t.Commit()
+ return n, err
+}
+
+func getAggregateFunc(aggregate byte) func(int64, int64) int64 {
+ switch aggregate {
+ case AggregateSum:
+ return func(a int64, b int64) int64 {
+ return a + b
+ }
+ case AggregateMax:
+ return func(a int64, b int64) int64 {
+ if a > b {
+ return a
+ }
+ return b
+ }
+ case AggregateMin:
+ return func(a int64, b int64) int64 {
+ if a > b {
+ return b
+ }
+ return a
+ }
+ }
+ return nil
+}
+
+func (db *DB) ZUnionStore(destKey []byte, srcKeys [][]byte, weights []int64, aggregate byte) (int64, error) {
+
+ var destMap = map[string]int64{}
+ aggregateFunc := getAggregateFunc(aggregate)
+ if aggregateFunc == nil {
+ return 0, errInvalidAggregate
+ }
+ if len(srcKeys) < 1 {
+ return 0, errInvalidSrcKeyNum
+ }
+ if weights != nil {
+ if len(srcKeys) != len(weights) {
+ return 0, errInvalidWeightNum
+ }
+ } else {
+ weights = make([]int64, len(srcKeys))
+ for i := 0; i < len(weights); i++ {
+ weights[i] = 1
+ }
+ }
+
+ for i, key := range srcKeys {
+ scorePairs, err := db.ZRange(key, 0, -1)
+ if err != nil {
+ return 0, err
+ }
+ for _, pair := range scorePairs {
+ if score, ok := destMap[String(pair.Member)]; !ok {
+ destMap[String(pair.Member)] = pair.Score * weights[i]
+ } else {
+ destMap[String(pair.Member)] = aggregateFunc(score, pair.Score*weights[i])
+ }
+ }
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ db.zDelete(t, destKey)
+
+ for member, score := range destMap {
+ if err := checkZSetKMSize(destKey, []byte(member)); err != nil {
+ return 0, err
+ }
+
+ if _, err := db.zSetItem(t, destKey, score, []byte(member)); err != nil {
+ return 0, err
+ }
+ }
+
+ var num = int64(len(destMap))
+ sk := db.zEncodeSizeKey(destKey)
+ t.Put(sk, PutInt64(num))
+
+ //todo add binlog
+ if err := t.Commit(); err != nil {
+ return 0, err
+ }
+ return num, nil
+}
+
+func (db *DB) ZInterStore(destKey []byte, srcKeys [][]byte, weights []int64, aggregate byte) (int64, error) {
+
+ aggregateFunc := getAggregateFunc(aggregate)
+ if aggregateFunc == nil {
+ return 0, errInvalidAggregate
+ }
+ if len(srcKeys) < 1 {
+ return 0, errInvalidSrcKeyNum
+ }
+ if weights != nil {
+ if len(srcKeys) != len(weights) {
+ return 0, errInvalidWeightNum
+ }
+ } else {
+ weights = make([]int64, len(srcKeys))
+ for i := 0; i < len(weights); i++ {
+ weights[i] = 1
+ }
+ }
+
+ var destMap = map[string]int64{}
+ scorePairs, err := db.ZRange(srcKeys[0], 0, -1)
+ if err != nil {
+ return 0, err
+ }
+ for _, pair := range scorePairs {
+ destMap[String(pair.Member)] = pair.Score * weights[0]
+ }
+
+ for i, key := range srcKeys[1:] {
+ scorePairs, err := db.ZRange(key, 0, -1)
+ if err != nil {
+ return 0, err
+ }
+ tmpMap := map[string]int64{}
+ for _, pair := range scorePairs {
+ if score, ok := destMap[String(pair.Member)]; ok {
+ tmpMap[String(pair.Member)] = aggregateFunc(score, pair.Score*weights[i+1])
+ }
+ }
+ destMap = tmpMap
+ }
+
+ t := db.zsetBatch
+ t.Lock()
+ defer t.Unlock()
+
+ db.zDelete(t, destKey)
+
+ for member, score := range destMap {
+ if err := checkZSetKMSize(destKey, []byte(member)); err != nil {
+ return 0, err
+ }
+ if _, err := db.zSetItem(t, destKey, score, []byte(member)); err != nil {
+ return 0, err
+ }
+ }
+
+ var num int64 = int64(len(destMap))
+ sk := db.zEncodeSizeKey(destKey)
+ t.Put(sk, PutInt64(num))
+ //todo add binlog
+ if err := t.Commit(); err != nil {
+ return 0, err
+ }
+ return num, nil
+}
+
+func (db *DB) ZScan(key []byte, count int, inclusive bool, match string) ([][]byte, error) {
+ return db.scan(ZSizeType, key, count, inclusive, match)
+}