mirror of
https://codeberg.org/superseriousbusiness/gotosocial.git
synced 2024-12-19 23:52:04 +03:00
7ec1e1332e
* rewrite cache library as codeberg.org/gruf/go-structr, implement in gotosocial
* use actual go-structr release version (not just commit hash)
* revert go toolchain changes (damn you go for auto changing this)
* fix go mod woes
* ensure %w is used in calls to errs.Appendf()
* fix error checking
* fix possible panic
* remove unnecessary start/stop functions, move to main Cache{} struct, add note regarding which caches require start/stop
* fix copy-paste artifact... 😇
* fix all comment copy-paste artifacts
* remove dropID() function, now we can just use slices.DeleteFunc()
* use util.Deduplicate() instead of collate(), move collate to util
* move orderByIDs() to util package and "generify"
* add a util.DeleteIf() function, use this to delete entries on failed population
* use slices.DeleteFunc() instead of util.DeleteIf() (i had the logic mixed up in my head somehow lol)
* add note about how collate differs from deduplicate
731 lines
17 KiB
Go
731 lines
17 KiB
Go
package structr
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"reflect"
|
|
"sync"
|
|
)
|
|
|
|
// DefaultIgnoreErr is the default function used to
|
|
// ignore (i.e. not cache) incoming error results during
|
|
// Load() calls. By default ignores context pkg errors.
|
|
func DefaultIgnoreErr(err error) bool {
|
|
return errors.Is(err, context.Canceled) ||
|
|
errors.Is(err, context.DeadlineExceeded)
|
|
}
|
|
|
|
// Config defines config variables
|
|
// for initializing a struct cache.
|
|
type Config[StructType any] struct {
|
|
|
|
// Indices defines indices to create
|
|
// in the Cache for the receiving
|
|
// generic struct type parameter.
|
|
Indices []IndexConfig
|
|
|
|
// MaxSize defines the maximum number
|
|
// of results allowed in the Cache at
|
|
// one time, before old results start
|
|
// getting evicted.
|
|
MaxSize int
|
|
|
|
// IgnoreErr defines which errors to
|
|
// ignore (i.e. not cache) returned
|
|
// from load function callback calls.
|
|
// This may be left as nil, on which
|
|
// DefaultIgnoreErr will be used.
|
|
IgnoreErr func(error) bool
|
|
|
|
// CopyValue provides a means of copying
|
|
// cached values, to ensure returned values
|
|
// do not share memory with those in cache.
|
|
CopyValue func(StructType) StructType
|
|
|
|
// Invalidate is called when cache values
|
|
// (NOT errors) are invalidated, either
|
|
// as the values passed to Put() / Store(),
|
|
// or by the keys by calls to Invalidate().
|
|
Invalidate func(StructType)
|
|
}
|
|
|
|
// Cache provides a structure cache with automated
|
|
// indexing and lookups by any initialization-defined
|
|
// combination of fields (as long as serialization is
|
|
// supported by codeberg.org/gruf/go-mangler). This
|
|
// also supports caching of negative results by errors
|
|
// returned from the LoadOne() series of functions.
|
|
type Cache[StructType any] struct {
|
|
|
|
// indices used in storing passed struct
|
|
// types by user defined sets of fields.
|
|
indices []Index[StructType]
|
|
|
|
// keeps track of all indexed results,
|
|
// in order of last recently used (LRU).
|
|
lruList list[*result[StructType]]
|
|
|
|
// memory pools of common types.
|
|
llsPool []*list[*result[StructType]]
|
|
resPool []*result[StructType]
|
|
keyPool []*indexkey[StructType]
|
|
|
|
// max cache size, imposes size
|
|
// limit on the lruList in order
|
|
// to evict old entries.
|
|
maxSize int
|
|
|
|
// hook functions.
|
|
ignore func(error) bool
|
|
copy func(StructType) StructType
|
|
invalid func(StructType)
|
|
|
|
// protective mutex, guards:
|
|
// - Cache{}.lruList
|
|
// - Index{}.data
|
|
// - Cache{} hook fns
|
|
// - Cache{} pools
|
|
mutex sync.Mutex
|
|
}
|
|
|
|
// Init initializes the cache with given configuration
|
|
// including struct fields to index, and necessary fns.
|
|
func (c *Cache[T]) Init(config Config[T]) {
|
|
if len(config.Indices) == 0 {
|
|
panic("no indices provided")
|
|
}
|
|
|
|
if config.IgnoreErr == nil {
|
|
config.IgnoreErr = DefaultIgnoreErr
|
|
}
|
|
|
|
if config.CopyValue == nil {
|
|
panic("copy value function must be provided")
|
|
}
|
|
|
|
if config.MaxSize < 2 {
|
|
panic("minimum cache size is 2 for LRU to work")
|
|
}
|
|
|
|
// Safely copy over
|
|
// provided config.
|
|
c.mutex.Lock()
|
|
c.indices = make([]Index[T], len(config.Indices))
|
|
for i, config := range config.Indices {
|
|
c.indices[i].init(config)
|
|
}
|
|
c.ignore = config.IgnoreErr
|
|
c.copy = config.CopyValue
|
|
c.invalid = config.Invalidate
|
|
c.maxSize = config.MaxSize
|
|
c.mutex.Unlock()
|
|
}
|
|
|
|
// Index selects index with given name from cache, else panics.
|
|
func (c *Cache[T]) Index(name string) *Index[T] {
|
|
for i := range c.indices {
|
|
if c.indices[i].name == name {
|
|
return &c.indices[i]
|
|
}
|
|
}
|
|
panic("unknown index: " + name)
|
|
}
|
|
|
|
// GetOne fetches one value from the cache stored under index, using key generated from key parts.
|
|
// Note that given number of key parts MUST match expected number and types of the given index name.
|
|
func (c *Cache[T]) GetOne(index string, keyParts ...any) (T, bool) {
|
|
// Get index with name.
|
|
idx := c.Index(index)
|
|
|
|
// Generate index key from provided parts.
|
|
key, ok := idx.keygen.FromParts(keyParts...)
|
|
if !ok {
|
|
var zero T
|
|
return zero, false
|
|
}
|
|
|
|
// Fetch one value for key.
|
|
return c.GetOneBy(idx, key)
|
|
}
|
|
|
|
// GetOneBy fetches value from cache stored under index, using precalculated index key.
|
|
func (c *Cache[T]) GetOneBy(index *Index[T], key string) (T, bool) {
|
|
if index == nil {
|
|
panic("no index given")
|
|
} else if !index.unique {
|
|
panic("cannot get one by non-unique index")
|
|
}
|
|
values := c.GetBy(index, key)
|
|
if len(values) == 0 {
|
|
var zero T
|
|
return zero, false
|
|
}
|
|
return values[0], true
|
|
}
|
|
|
|
// Get fetches values from the cache stored under index, using keys generated from given key parts.
|
|
// Note that each number of key parts MUST match expected number and types of the given index name.
|
|
func (c *Cache[T]) Get(index string, keysParts ...[]any) []T {
|
|
// Get index with name.
|
|
idx := c.Index(index)
|
|
|
|
// Preallocate expected keys slice length.
|
|
keys := make([]string, 0, len(keysParts))
|
|
|
|
// Acquire buf.
|
|
buf := getBuf()
|
|
|
|
for _, parts := range keysParts {
|
|
// Reset buf.
|
|
buf.Reset()
|
|
|
|
// Generate key from provided parts into buffer.
|
|
if !idx.keygen.AppendFromParts(buf, parts...) {
|
|
continue
|
|
}
|
|
|
|
// Get string copy of
|
|
// genarated idx key.
|
|
key := string(buf.B)
|
|
|
|
// Append key to keys.
|
|
keys = append(keys, key)
|
|
}
|
|
|
|
// Done with buf.
|
|
putBuf(buf)
|
|
|
|
// Continue fetching values.
|
|
return c.GetBy(idx, keys...)
|
|
}
|
|
|
|
// GetBy fetches values from the cache stored under index, using precalculated index keys.
|
|
func (c *Cache[T]) GetBy(index *Index[T], keys ...string) []T {
|
|
if index == nil {
|
|
panic("no index given")
|
|
}
|
|
|
|
// Preallocate a slice of est. len.
|
|
values := make([]T, 0, len(keys))
|
|
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Check cache init.
|
|
if c.copy == nil {
|
|
c.mutex.Unlock()
|
|
panic("not initialized")
|
|
}
|
|
|
|
// Check index for all keys.
|
|
for _, key := range keys {
|
|
|
|
// Get indexed results.
|
|
list := index.data[key]
|
|
|
|
if list != nil {
|
|
// Concatenate all results with values.
|
|
list.rangefn(func(e *elem[*result[T]]) {
|
|
if e.Value.err != nil {
|
|
return
|
|
}
|
|
|
|
// Append a copy of value.
|
|
value := c.copy(e.Value.value)
|
|
values = append(values, value)
|
|
|
|
// Push to front of LRU list, USING
|
|
// THE RESULT'S LRU ENTRY, NOT THE
|
|
// INDEX KEY ENTRY. VERY IMPORTANT!!
|
|
c.lruList.moveFront(&e.Value.entry)
|
|
})
|
|
}
|
|
}
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
|
|
return values
|
|
}
|
|
|
|
// Put will insert the given values into cache,
|
|
// calling any invalidate hook on each value.
|
|
func (c *Cache[T]) Put(values ...T) {
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Get func ptrs.
|
|
invalid := c.invalid
|
|
|
|
// Check cache init.
|
|
if c.copy == nil {
|
|
c.mutex.Unlock()
|
|
panic("not initialized")
|
|
}
|
|
|
|
// Store all the passed values.
|
|
for _, value := range values {
|
|
c.store(nil, "", value, nil)
|
|
}
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
|
|
if invalid != nil {
|
|
// Pass all invalidated values
|
|
// to given user hook (if set).
|
|
for _, value := range values {
|
|
invalid(value)
|
|
}
|
|
}
|
|
}
|
|
|
|
// LoadOne fetches one result from the cache stored under index, using key generated from key parts.
|
|
// In the case that no result is found, the provided load callback will be used to hydrate the cache.
|
|
// Note that given number of key parts MUST match expected number and types of the given index name.
|
|
func (c *Cache[T]) LoadOne(index string, load func() (T, error), keyParts ...any) (T, error) {
|
|
// Get index with name.
|
|
idx := c.Index(index)
|
|
|
|
// Generate cache from from provided parts.
|
|
key, _ := idx.keygen.FromParts(keyParts...)
|
|
|
|
// Continue loading this result.
|
|
return c.LoadOneBy(idx, load, key)
|
|
}
|
|
|
|
// LoadOneBy fetches one result from the cache stored under index, using precalculated index key.
|
|
// In the case that no result is found, provided load callback will be used to hydrate the cache.
|
|
func (c *Cache[T]) LoadOneBy(index *Index[T], load func() (T, error), key string) (T, error) {
|
|
if index == nil {
|
|
panic("no index given")
|
|
} else if !index.unique {
|
|
panic("cannot get one by non-unique index")
|
|
}
|
|
|
|
var (
|
|
// whether a result was found
|
|
// (and so val / err are set).
|
|
ok bool
|
|
|
|
// separate value / error ptrs
|
|
// as the result is liable to
|
|
// change outside of lock.
|
|
val T
|
|
err error
|
|
)
|
|
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Get func ptrs.
|
|
ignore := c.ignore
|
|
|
|
// Check init'd.
|
|
if c.copy == nil ||
|
|
ignore == nil {
|
|
c.mutex.Unlock()
|
|
panic("not initialized")
|
|
}
|
|
|
|
// Get indexed results.
|
|
list := index.data[key]
|
|
|
|
if ok = (list != nil && list.head != nil); ok {
|
|
e := list.head
|
|
|
|
// Extract val / err.
|
|
val = e.Value.value
|
|
err = e.Value.err
|
|
|
|
if err == nil {
|
|
// We only ever ret
|
|
// a COPY of value.
|
|
val = c.copy(val)
|
|
}
|
|
|
|
// Push to front of LRU list, USING
|
|
// THE RESULT'S LRU ENTRY, NOT THE
|
|
// INDEX KEY ENTRY. VERY IMPORTANT!!
|
|
c.lruList.moveFront(&e.Value.entry)
|
|
}
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
|
|
if ok {
|
|
// result found!
|
|
return val, err
|
|
}
|
|
|
|
// Load new result.
|
|
val, err = load()
|
|
|
|
// Check for ignored
|
|
// (transient) errors.
|
|
if ignore(err) {
|
|
return val, err
|
|
}
|
|
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Index this new loaded result.
|
|
// Note this handles copying of
|
|
// the provided value, so it is
|
|
// safe for us to return as-is.
|
|
c.store(index, key, val, err)
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
|
|
return val, err
|
|
}
|
|
|
|
// Load fetches values from the cache stored under index, using keys generated from given key parts. The provided get callback is used
|
|
// to load groups of values from the cache by the key generated from the key parts provided to the inner callback func, where the returned
|
|
// boolean indicates whether any values are currently stored. After the get callback has returned, the cache will then call provided load
|
|
// callback to hydrate the cache with any other values. Example usage here is that you may see which values are cached using 'get', and load
|
|
// the remaining uncached values using 'load', to minimize database queries. Cached error results are not included or returned by this func.
|
|
// Note that given number of key parts MUST match expected number and types of the given index name, in those provided to the get callback.
|
|
func (c *Cache[T]) Load(index string, get func(load func(keyParts ...any) bool), load func() ([]T, error)) (values []T, err error) {
|
|
return c.LoadBy(c.Index(index), get, load)
|
|
}
|
|
|
|
// LoadBy fetches values from the cache stored under index, using precalculated index key. The provided get callback is used to load
|
|
// groups of values from the cache by the key generated from the key parts provided to the inner callback func, where the returned boolea
|
|
// indicates whether any values are currently stored. After the get callback has returned, the cache will then call provided load callback
|
|
// to hydrate the cache with any other values. Example usage here is that you may see which values are cached using 'get', and load the
|
|
// remaining uncached values using 'load', to minimize database queries. Cached error results are not included or returned by this func.
|
|
// Note that given number of key parts MUST match expected number and types of the given index name, in those provided to the get callback.
|
|
func (c *Cache[T]) LoadBy(index *Index[T], get func(load func(keyParts ...any) bool), load func() ([]T, error)) (values []T, err error) {
|
|
if index == nil {
|
|
panic("no index given")
|
|
}
|
|
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Check init'd.
|
|
if c.copy == nil {
|
|
c.mutex.Unlock()
|
|
panic("not initialized")
|
|
}
|
|
|
|
var unlocked bool
|
|
defer func() {
|
|
// Deferred unlock to catch
|
|
// any user function panics.
|
|
if !unlocked {
|
|
c.mutex.Unlock()
|
|
}
|
|
}()
|
|
|
|
// Acquire buf.
|
|
buf := getBuf()
|
|
|
|
// Pass cache check to user func.
|
|
get(func(keyParts ...any) bool {
|
|
|
|
// Reset buf.
|
|
buf.Reset()
|
|
|
|
// Generate index key from provided key parts.
|
|
if !index.keygen.AppendFromParts(buf, keyParts...) {
|
|
return false
|
|
}
|
|
|
|
// Get temp generated key str,
|
|
// (not needed after return).
|
|
keyStr := buf.String()
|
|
|
|
// Get all indexed results.
|
|
list := index.data[keyStr]
|
|
|
|
if list != nil && list.len > 0 {
|
|
// Value length before
|
|
// any below appends.
|
|
before := len(values)
|
|
|
|
// Concatenate all results with values.
|
|
list.rangefn(func(e *elem[*result[T]]) {
|
|
if e.Value.err != nil {
|
|
return
|
|
}
|
|
|
|
// Append a copy of value.
|
|
value := c.copy(e.Value.value)
|
|
values = append(values, value)
|
|
|
|
// Push to front of LRU list, USING
|
|
// THE RESULT'S LRU ENTRY, NOT THE
|
|
// INDEX KEY ENTRY. VERY IMPORTANT!!
|
|
c.lruList.moveFront(&e.Value.entry)
|
|
})
|
|
|
|
// Only if values changed did
|
|
// we actually find anything.
|
|
return len(values) != before
|
|
}
|
|
|
|
return false
|
|
})
|
|
|
|
// Done with buf.
|
|
putBuf(buf)
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
unlocked = true
|
|
|
|
// Load uncached values.
|
|
uncached, err := load()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Insert uncached.
|
|
c.Put(uncached...)
|
|
|
|
// Append uncached to return values.
|
|
values = append(values, uncached...)
|
|
|
|
return
|
|
}
|
|
|
|
// Store will call the given store callback, on non-error then
|
|
// passing the provided value to the Put() function. On error
|
|
// return the value is still passed to stored invalidate hook.
|
|
func (c *Cache[T]) Store(value T, store func() error) error {
|
|
// Store value.
|
|
err := store()
|
|
|
|
if err != nil {
|
|
// Get func ptrs.
|
|
c.mutex.Lock()
|
|
invalid := c.invalid
|
|
c.mutex.Unlock()
|
|
|
|
// On error don't store
|
|
// value, but still pass
|
|
// to invalidate hook.
|
|
if invalid != nil {
|
|
invalid(value)
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// Store value.
|
|
c.Put(value)
|
|
|
|
return nil
|
|
}
|
|
|
|
// Invalidate generates index key from parts and invalidates all stored under it.
|
|
func (c *Cache[T]) Invalidate(index string, keyParts ...any) {
|
|
// Get index with name.
|
|
idx := c.Index(index)
|
|
|
|
// Generate cache from from provided parts.
|
|
key, ok := idx.keygen.FromParts(keyParts...)
|
|
if !ok {
|
|
return
|
|
}
|
|
|
|
// Continue invalidation.
|
|
c.InvalidateBy(idx, key)
|
|
}
|
|
|
|
// InvalidateBy invalidates all results stored under index key.
|
|
func (c *Cache[T]) InvalidateBy(index *Index[T], key string) {
|
|
if index == nil {
|
|
panic("no index given")
|
|
}
|
|
|
|
var values []T
|
|
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Get func ptrs.
|
|
invalid := c.invalid
|
|
|
|
// Delete all results under key from index, collecting
|
|
// value results and dropping them from all their indices.
|
|
index_delete(c, index, key, func(del *result[T]) {
|
|
if del.err == nil {
|
|
values = append(values, del.value)
|
|
}
|
|
c.delete(del)
|
|
})
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
|
|
if invalid != nil {
|
|
// Pass all invalidated values
|
|
// to given user hook (if set).
|
|
for _, value := range values {
|
|
invalid(value)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Trim will truncate the cache to ensure it
|
|
// stays within given percentage of MaxSize.
|
|
func (c *Cache[T]) Trim(perc float64) {
|
|
// Acquire lock.
|
|
c.mutex.Lock()
|
|
|
|
// Calculate number of cache items to drop.
|
|
max := (perc / 100) * float64(c.maxSize)
|
|
diff := c.lruList.len - int(max)
|
|
|
|
if diff <= 0 {
|
|
// Trim not needed.
|
|
c.mutex.Unlock()
|
|
return
|
|
}
|
|
|
|
// Iterate over 'diff' results
|
|
// from back (oldest) of cache.
|
|
for i := 0; i < diff; i++ {
|
|
|
|
// Get oldest LRU element.
|
|
oldest := c.lruList.tail
|
|
|
|
if oldest == nil {
|
|
// reached end.
|
|
break
|
|
}
|
|
|
|
// Drop oldest from cache.
|
|
c.delete(oldest.Value)
|
|
}
|
|
|
|
// Done with lock.
|
|
c.mutex.Unlock()
|
|
}
|
|
|
|
// Clear empties the cache by calling .Trim(0).
|
|
func (c *Cache[T]) Clear() { c.Trim(0) }
|
|
|
|
// Clean drops unused items from its memory pools.
|
|
// Useful to free memory if cache has downsized.
|
|
func (c *Cache[T]) Clean() {
|
|
c.mutex.Lock()
|
|
c.llsPool = nil
|
|
c.resPool = nil
|
|
c.keyPool = nil
|
|
c.mutex.Unlock()
|
|
}
|
|
|
|
// Len returns the current length of cache.
|
|
func (c *Cache[T]) Len() int {
|
|
c.mutex.Lock()
|
|
l := c.lruList.len
|
|
c.mutex.Unlock()
|
|
return l
|
|
}
|
|
|
|
// Cap returns the maximum capacity (size) of cache.
|
|
func (c *Cache[T]) Cap() int {
|
|
c.mutex.Lock()
|
|
m := c.maxSize
|
|
c.mutex.Unlock()
|
|
return m
|
|
}
|
|
|
|
// store will store the given value / error result in the cache, storing it under the
|
|
// already provided index + key if provided, else generating keys from provided value.
|
|
func (c *Cache[T]) store(index *Index[T], key string, value T, err error) {
|
|
// Acquire new result.
|
|
res := result_acquire(c)
|
|
|
|
if index != nil {
|
|
// Append result to the provided
|
|
// precalculated key and its index.
|
|
index_append(c, index, key, res)
|
|
|
|
} else if err != nil {
|
|
|
|
// This is an error result without
|
|
// an index provided, nothing we
|
|
// can do here so release result.
|
|
result_release(c, res)
|
|
return
|
|
}
|
|
|
|
// Set and check the result error.
|
|
if res.err = err; res.err == nil {
|
|
|
|
// This is value result, we need to
|
|
// store it under all other indices
|
|
// other than the provided.
|
|
//
|
|
// Create COPY of value.
|
|
res.value = c.copy(value)
|
|
|
|
// Get reflected value of incoming
|
|
// value, used during cache key gen.
|
|
rvalue := reflect.ValueOf(value)
|
|
|
|
// Acquire buf.
|
|
buf := getBuf()
|
|
|
|
for i := range c.indices {
|
|
// Get current index ptr.
|
|
idx := &(c.indices[i])
|
|
|
|
if idx == index {
|
|
// Already stored under
|
|
// this index, ignore.
|
|
continue
|
|
}
|
|
|
|
// Generate key from reflect value,
|
|
// (this ignores zero value keys).
|
|
buf.Reset() // reset buf first
|
|
if !idx.keygen.appendFromRValue(buf, rvalue) {
|
|
continue
|
|
}
|
|
|
|
// Alloc key copy.
|
|
key := string(buf.B)
|
|
|
|
// Append result to index at key.
|
|
index_append(c, idx, key, res)
|
|
}
|
|
|
|
// Done with buf.
|
|
putBuf(buf)
|
|
}
|
|
|
|
if c.lruList.len > c.maxSize {
|
|
// Cache has hit max size!
|
|
// Drop the oldest element.
|
|
res := c.lruList.tail.Value
|
|
c.delete(res)
|
|
}
|
|
}
|
|
|
|
// delete will delete the given result from the cache, deleting
|
|
// it from all indices it is stored under, and main LRU list.
|
|
func (c *Cache[T]) delete(res *result[T]) {
|
|
for len(res.keys) != 0 {
|
|
|
|
// Pop indexkey at end of list.
|
|
ikey := res.keys[len(res.keys)-1]
|
|
res.keys = res.keys[:len(res.keys)-1]
|
|
|
|
// Drop this result from list at key.
|
|
index_deleteOne(c, ikey.index, ikey)
|
|
|
|
// Release ikey to pool.
|
|
indexkey_release(c, ikey)
|
|
}
|
|
|
|
// Release res to pool.
|
|
result_release(c, res)
|
|
}
|