aboutsummaryrefslogtreecommitdiffstats
path: root/vendor/gitea.com/lunny/nodb/t_set.go
diff options
context:
space:
mode:
Diffstat (limited to 'vendor/gitea.com/lunny/nodb/t_set.go')
-rw-r--r--vendor/gitea.com/lunny/nodb/t_set.go601
1 files changed, 0 insertions, 601 deletions
diff --git a/vendor/gitea.com/lunny/nodb/t_set.go b/vendor/gitea.com/lunny/nodb/t_set.go
deleted file mode 100644
index 0ff33e991a..0000000000
--- a/vendor/gitea.com/lunny/nodb/t_set.go
+++ /dev/null
@@ -1,601 +0,0 @@
-package nodb
-
-import (
- "encoding/binary"
- "errors"
- "time"
-
- "gitea.com/lunny/nodb/store"
-)
-
-var errSetKey = errors.New("invalid set key")
-var errSSizeKey = errors.New("invalid ssize key")
-
-const (
- setStartSep byte = ':'
- setStopSep byte = setStartSep + 1
- UnionType byte = 51
- DiffType byte = 52
- InterType byte = 53
-)
-
-func checkSetKMSize(key []byte, member []byte) error {
- if len(key) > MaxKeySize || len(key) == 0 {
- return errKeySize
- } else if len(member) > MaxSetMemberSize || len(member) == 0 {
- return errSetMemberSize
- }
- return nil
-}
-
-func (db *DB) sEncodeSizeKey(key []byte) []byte {
- buf := make([]byte, len(key)+2)
-
- buf[0] = db.index
- buf[1] = SSizeType
-
- copy(buf[2:], key)
- return buf
-}
-
-func (db *DB) sDecodeSizeKey(ek []byte) ([]byte, error) {
- if len(ek) < 2 || ek[0] != db.index || ek[1] != SSizeType {
- return nil, errSSizeKey
- }
-
- return ek[2:], nil
-}
-
-func (db *DB) sEncodeSetKey(key []byte, member []byte) []byte {
- buf := make([]byte, len(key)+len(member)+1+1+2+1)
-
- pos := 0
- buf[pos] = db.index
- pos++
- buf[pos] = SetType
- pos++
-
- binary.BigEndian.PutUint16(buf[pos:], uint16(len(key)))
- pos += 2
-
- copy(buf[pos:], key)
- pos += len(key)
-
- buf[pos] = setStartSep
- pos++
- copy(buf[pos:], member)
-
- return buf
-}
-
-func (db *DB) sDecodeSetKey(ek []byte) ([]byte, []byte, error) {
- if len(ek) < 5 || ek[0] != db.index || ek[1] != SetType {
- return nil, nil, errSetKey
- }
-
- pos := 2
- keyLen := int(binary.BigEndian.Uint16(ek[pos:]))
- pos += 2
-
- if keyLen+5 > len(ek) {
- return nil, nil, errSetKey
- }
-
- key := ek[pos : pos+keyLen]
- pos += keyLen
-
- if ek[pos] != hashStartSep {
- return nil, nil, errSetKey
- }
-
- pos++
- member := ek[pos:]
- return key, member, nil
-}
-
-func (db *DB) sEncodeStartKey(key []byte) []byte {
- return db.sEncodeSetKey(key, nil)
-}
-
-func (db *DB) sEncodeStopKey(key []byte) []byte {
- k := db.sEncodeSetKey(key, nil)
-
- k[len(k)-1] = setStopSep
-
- return k
-}
-
-func (db *DB) sFlush() (drop int64, err error) {
-
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- return db.flushType(t, SetType)
-}
-
-func (db *DB) sDelete(t *batch, key []byte) int64 {
- sk := db.sEncodeSizeKey(key)
- start := db.sEncodeStartKey(key)
- stop := db.sEncodeStopKey(key)
-
- var num int64 = 0
- it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
- for ; it.Valid(); it.Next() {
- t.Delete(it.RawKey())
- num++
- }
-
- it.Close()
-
- t.Delete(sk)
- return num
-}
-
-func (db *DB) sIncrSize(key []byte, delta int64) (int64, error) {
- t := db.setBatch
- sk := db.sEncodeSizeKey(key)
-
- var err error
- var size int64 = 0
- if size, err = Int64(db.bucket.Get(sk)); err != nil {
- return 0, err
- } else {
- size += delta
- if size <= 0 {
- size = 0
- t.Delete(sk)
- db.rmExpire(t, SetType, key)
- } else {
- t.Put(sk, PutInt64(size))
- }
- }
-
- return size, nil
-}
-
-func (db *DB) sExpireAt(key []byte, when int64) (int64, error) {
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- if scnt, err := db.SCard(key); err != nil || scnt == 0 {
- return 0, err
- } else {
- db.expireAt(t, SetType, key, when)
- if err := t.Commit(); err != nil {
- return 0, err
- }
-
- }
-
- return 1, nil
-}
-
-func (db *DB) sSetItem(key []byte, member []byte) (int64, error) {
- t := db.setBatch
- ek := db.sEncodeSetKey(key, member)
-
- var n int64 = 1
- if v, _ := db.bucket.Get(ek); v != nil {
- n = 0
- } else {
- if _, err := db.sIncrSize(key, 1); err != nil {
- return 0, err
- }
- }
-
- t.Put(ek, nil)
- return n, nil
-}
-
-func (db *DB) SAdd(key []byte, args ...[]byte) (int64, error) {
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- var err error
- var ek []byte
- var num int64 = 0
- for i := 0; i < len(args); i++ {
- if err := checkSetKMSize(key, args[i]); err != nil {
- return 0, err
- }
-
- ek = db.sEncodeSetKey(key, args[i])
-
- if v, err := db.bucket.Get(ek); err != nil {
- return 0, err
- } else if v == nil {
- num++
- }
-
- t.Put(ek, nil)
- }
-
- if _, err = db.sIncrSize(key, num); err != nil {
- return 0, err
- }
-
- err = t.Commit()
- return num, err
-
-}
-
-func (db *DB) SCard(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- sk := db.sEncodeSizeKey(key)
-
- return Int64(db.bucket.Get(sk))
-}
-
-func (db *DB) sDiffGeneric(keys ...[]byte) ([][]byte, error) {
- destMap := make(map[string]bool)
-
- members, err := db.SMembers(keys[0])
- if err != nil {
- return nil, err
- }
-
- for _, m := range members {
- destMap[String(m)] = true
- }
-
- for _, k := range keys[1:] {
- members, err := db.SMembers(k)
- if err != nil {
- return nil, err
- }
-
- for _, m := range members {
- if _, ok := destMap[String(m)]; !ok {
- continue
- } else if ok {
- delete(destMap, String(m))
- }
- }
- // O - A = O, O is zero set.
- if len(destMap) == 0 {
- return nil, nil
- }
- }
-
- slice := make([][]byte, len(destMap))
- idx := 0
- for k, v := range destMap {
- if !v {
- continue
- }
- slice[idx] = []byte(k)
- idx++
- }
-
- return slice, nil
-}
-
-func (db *DB) SDiff(keys ...[]byte) ([][]byte, error) {
- v, err := db.sDiffGeneric(keys...)
- return v, err
-}
-
-func (db *DB) SDiffStore(dstKey []byte, keys ...[]byte) (int64, error) {
- n, err := db.sStoreGeneric(dstKey, DiffType, keys...)
- return n, err
-}
-
-func (db *DB) sInterGeneric(keys ...[]byte) ([][]byte, error) {
- destMap := make(map[string]bool)
-
- members, err := db.SMembers(keys[0])
- if err != nil {
- return nil, err
- }
-
- for _, m := range members {
- destMap[String(m)] = true
- }
-
- for _, key := range keys[1:] {
- if err := checkKeySize(key); err != nil {
- return nil, err
- }
-
- members, err := db.SMembers(key)
- if err != nil {
- return nil, err
- } else if len(members) == 0 {
- return nil, err
- }
-
- tempMap := make(map[string]bool)
- for _, member := range members {
- if err := checkKeySize(member); err != nil {
- return nil, err
- }
- if _, ok := destMap[String(member)]; ok {
- tempMap[String(member)] = true //mark this item as selected
- }
- }
- destMap = tempMap //reduce the size of the result set
- if len(destMap) == 0 {
- return nil, nil
- }
- }
-
- slice := make([][]byte, len(destMap))
- idx := 0
- for k, v := range destMap {
- if !v {
- continue
- }
-
- slice[idx] = []byte(k)
- idx++
- }
-
- return slice, nil
-
-}
-
-func (db *DB) SInter(keys ...[]byte) ([][]byte, error) {
- v, err := db.sInterGeneric(keys...)
- return v, err
-
-}
-
-func (db *DB) SInterStore(dstKey []byte, keys ...[]byte) (int64, error) {
- n, err := db.sStoreGeneric(dstKey, InterType, keys...)
- return n, err
-}
-
-func (db *DB) SIsMember(key []byte, member []byte) (int64, error) {
- ek := db.sEncodeSetKey(key, member)
-
- var n int64 = 1
- if v, err := db.bucket.Get(ek); err != nil {
- return 0, err
- } else if v == nil {
- n = 0
- }
- return n, nil
-}
-
-func (db *DB) SMembers(key []byte) ([][]byte, error) {
- if err := checkKeySize(key); err != nil {
- return nil, err
- }
-
- start := db.sEncodeStartKey(key)
- stop := db.sEncodeStopKey(key)
-
- v := make([][]byte, 0, 16)
-
- it := db.bucket.RangeLimitIterator(start, stop, store.RangeROpen, 0, -1)
- for ; it.Valid(); it.Next() {
- _, m, err := db.sDecodeSetKey(it.Key())
- if err != nil {
- return nil, err
- }
-
- v = append(v, m)
- }
-
- it.Close()
-
- return v, nil
-}
-
-func (db *DB) SRem(key []byte, args ...[]byte) (int64, error) {
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- var ek []byte
- var v []byte
- var err error
-
- it := db.bucket.NewIterator()
- defer it.Close()
-
- var num int64 = 0
- for i := 0; i < len(args); i++ {
- if err := checkSetKMSize(key, args[i]); err != nil {
- return 0, err
- }
-
- ek = db.sEncodeSetKey(key, args[i])
-
- v = it.RawFind(ek)
- if v == nil {
- continue
- } else {
- num++
- t.Delete(ek)
- }
- }
-
- if _, err = db.sIncrSize(key, -num); err != nil {
- return 0, err
- }
-
- err = t.Commit()
- return num, err
-
-}
-
-func (db *DB) sUnionGeneric(keys ...[]byte) ([][]byte, error) {
- dstMap := make(map[string]bool)
-
- for _, key := range keys {
- if err := checkKeySize(key); err != nil {
- return nil, err
- }
-
- members, err := db.SMembers(key)
- if err != nil {
- return nil, err
- }
-
- for _, member := range members {
- dstMap[String(member)] = true
- }
- }
-
- slice := make([][]byte, len(dstMap))
- idx := 0
- for k, v := range dstMap {
- if !v {
- continue
- }
- slice[idx] = []byte(k)
- idx++
- }
-
- return slice, nil
-}
-
-func (db *DB) SUnion(keys ...[]byte) ([][]byte, error) {
- v, err := db.sUnionGeneric(keys...)
- return v, err
-}
-
-func (db *DB) SUnionStore(dstKey []byte, keys ...[]byte) (int64, error) {
- n, err := db.sStoreGeneric(dstKey, UnionType, keys...)
- return n, err
-}
-
-func (db *DB) sStoreGeneric(dstKey []byte, optType byte, keys ...[]byte) (int64, error) {
- if err := checkKeySize(dstKey); err != nil {
- return 0, err
- }
-
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- db.sDelete(t, dstKey)
-
- var err error
- var ek []byte
- var v [][]byte
-
- switch optType {
- case UnionType:
- v, err = db.sUnionGeneric(keys...)
- case DiffType:
- v, err = db.sDiffGeneric(keys...)
- case InterType:
- v, err = db.sInterGeneric(keys...)
- }
-
- if err != nil {
- return 0, err
- }
-
- for _, m := range v {
- if err := checkSetKMSize(dstKey, m); err != nil {
- return 0, err
- }
-
- ek = db.sEncodeSetKey(dstKey, m)
-
- if _, err := db.bucket.Get(ek); err != nil {
- return 0, err
- }
-
- t.Put(ek, nil)
- }
-
- var num = int64(len(v))
- sk := db.sEncodeSizeKey(dstKey)
- t.Put(sk, PutInt64(num))
-
- if err = t.Commit(); err != nil {
- return 0, err
- }
- return num, nil
-}
-
-func (db *DB) SClear(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- num := db.sDelete(t, key)
- db.rmExpire(t, SetType, key)
-
- err := t.Commit()
- return num, err
-}
-
-func (db *DB) SMclear(keys ...[]byte) (int64, error) {
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- for _, key := range keys {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- db.sDelete(t, key)
- db.rmExpire(t, SetType, key)
- }
-
- err := t.Commit()
- return int64(len(keys)), err
-}
-
-func (db *DB) SExpire(key []byte, duration int64) (int64, error) {
- if duration <= 0 {
- return 0, errExpireValue
- }
-
- return db.sExpireAt(key, time.Now().Unix()+duration)
-
-}
-
-func (db *DB) SExpireAt(key []byte, when int64) (int64, error) {
- if when <= time.Now().Unix() {
- return 0, errExpireValue
- }
-
- return db.sExpireAt(key, when)
-
-}
-
-func (db *DB) STTL(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return -1, err
- }
-
- return db.ttl(SetType, key)
-}
-
-func (db *DB) SPersist(key []byte) (int64, error) {
- if err := checkKeySize(key); err != nil {
- return 0, err
- }
-
- t := db.setBatch
- t.Lock()
- defer t.Unlock()
-
- n, err := db.rmExpire(t, SetType, key)
- if err != nil {
- return 0, err
- }
- err = t.Commit()
- return n, err
-}
-
-func (db *DB) SScan(key []byte, count int, inclusive bool, match string) ([][]byte, error) {
- return db.scan(SSizeType, key, count, inclusive, match)
-}