mirror of
				https://github.com/superseriousbusiness/gotosocial.git
				synced 2025-10-30 23:32:25 -05:00 
			
		
		
		
	* Implement conversations API * Sort and page conversations by last status ID * Appease linter * Fix deleting conversations and statuses * Refactor to make migrations automatic * Lint * Update tests post-merge * Fixes from live-fire testing * Linter caught a format problem * Refactor tests, fix cache * Negative test for non-DMs * Run conversations advanced migration on testrig startup as well as regular server startup * Document (lack of) side effects of API method for deleting a conversation * Make not-found check less nested for readability * Rename PutConversation to UpsertConversation * Use util.Ptr instead of IIFE * Reduce cache used by conversations * Remove unnecessary TableExpr/ColumnExpr * Use struct tags for both unique constraints on Conversation * Make it clear how paging with GetDirectStatusIDsBatch should be used * Let conversation paging skip conversations it can't render * Use Bun NewDropTable * Convert delete raw query to Bun * Convert update raw query to Bun * Convert latestConversationStatusesTempTable raw query partially to Bun * Convert conversationStatusesTempTable raw query partially to Bun * Rename field used to store result of MaxDirectStatusID * Move advanced migrations to their own tiny processor * Catch up util function name with main * Remove json.… wrappers * Remove redundant check * Combine error checks * Replace map with slice of structs * Address processor/type converter comments - Add context info for errors - Extract some common processor code into shared methods - Move conversation eligibility check ahead of populating conversation * Add error context when dropping temp tables
		
			
				
	
	
		
			242 lines
		
	
	
	
		
			7.7 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			242 lines
		
	
	
	
		
			7.7 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // GoToSocial
 | |
| // Copyright (C) GoToSocial Authors admin@gotosocial.org
 | |
| // SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful,
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package cache
 | |
| 
 | |
| import (
 | |
| 	"slices"
 | |
| 
 | |
| 	"codeberg.org/gruf/go-cache/v3/simple"
 | |
| 	"codeberg.org/gruf/go-structr"
 | |
| )
 | |
| 
 | |
| // SliceCache wraps a simple.Cache to provide simple loader-callback
 | |
| // functions for fetching + caching slices of objects (e.g. IDs).
 | |
| type SliceCache[T any] struct {
 | |
| 	cache simple.Cache[string, []T]
 | |
| }
 | |
| 
 | |
| // Init initializes the cache with given length + capacity.
 | |
| func (c *SliceCache[T]) Init(len, cap int) {
 | |
| 	c.cache = simple.Cache[string, []T]{}
 | |
| 	c.cache.Init(len, cap)
 | |
| }
 | |
| 
 | |
| // Load will attempt to load an existing slice from cache for key, else calling load function and caching the result.
 | |
| func (c *SliceCache[T]) Load(key string, load func() ([]T, error)) ([]T, error) {
 | |
| 	// Look for cached values.
 | |
| 	data, ok := c.cache.Get(key)
 | |
| 
 | |
| 	if !ok {
 | |
| 		var err error
 | |
| 
 | |
| 		// Not cached, load!
 | |
| 		data, err = load()
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		// Store the data.
 | |
| 		c.cache.Set(key, data)
 | |
| 	}
 | |
| 
 | |
| 	// Return data clone for safety.
 | |
| 	return slices.Clone(data), nil
 | |
| }
 | |
| 
 | |
| // Invalidate: see simple.Cache{}.InvalidateAll().
 | |
| func (c *SliceCache[T]) Invalidate(keys ...string) {
 | |
| 	_ = c.cache.InvalidateAll(keys...)
 | |
| }
 | |
| 
 | |
| // Trim: see simple.Cache{}.Trim().
 | |
| func (c *SliceCache[T]) Trim(perc float64) {
 | |
| 	c.cache.Trim(perc)
 | |
| }
 | |
| 
 | |
| // Clear: see simple.Cache{}.Clear().
 | |
| func (c *SliceCache[T]) Clear() {
 | |
| 	c.cache.Clear()
 | |
| }
 | |
| 
 | |
| // Len: see simple.Cache{}.Len().
 | |
| func (c *SliceCache[T]) Len() int {
 | |
| 	return c.cache.Len()
 | |
| }
 | |
| 
 | |
| // Cap: see simple.Cache{}.Cap().
 | |
| func (c *SliceCache[T]) Cap() int {
 | |
| 	return c.cache.Cap()
 | |
| }
 | |
| 
 | |
| // StructCache wraps a structr.Cache{} to simple index caching
 | |
| // by name (also to ease update to library version that introduced
 | |
| // this). (in the future it may be worth embedding these indexes by
 | |
| // name under the main database caches struct which would reduce
 | |
| // time required to access cached values).
 | |
| type StructCache[StructType any] struct {
 | |
| 	cache structr.Cache[StructType]
 | |
| 	index map[string]*structr.Index
 | |
| }
 | |
| 
 | |
| // Init initializes the cache with given structr.CacheConfig{}.
 | |
| func (c *StructCache[T]) Init(config structr.CacheConfig[T]) {
 | |
| 	c.index = make(map[string]*structr.Index, len(config.Indices))
 | |
| 	c.cache = structr.Cache[T]{}
 | |
| 	c.cache.Init(config)
 | |
| 	for _, cfg := range config.Indices {
 | |
| 		c.index[cfg.Fields] = c.cache.Index(cfg.Fields)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // GetOne calls structr.Cache{}.GetOne(), using a cached structr.Index{} by 'index' name.
 | |
| // Note: this also handles conversion of the untyped (any) keys to structr.Key{} via structr.Index{}.
 | |
| func (c *StructCache[T]) GetOne(index string, key ...any) (T, bool) {
 | |
| 	i := c.index[index]
 | |
| 	return c.cache.GetOne(i, i.Key(key...))
 | |
| }
 | |
| 
 | |
| // Get calls structr.Cache{}.Get(), using a cached structr.Index{} by 'index' name.
 | |
| // Note: this also handles conversion of the untyped (any) keys to structr.Key{} via structr.Index{}.
 | |
| func (c *StructCache[T]) Get(index string, keys ...[]any) []T {
 | |
| 	i := c.index[index]
 | |
| 	return c.cache.Get(i, i.Keys(keys...)...)
 | |
| }
 | |
| 
 | |
| // Put: see structr.Cache{}.Put().
 | |
| func (c *StructCache[T]) Put(values ...T) {
 | |
| 	c.cache.Put(values...)
 | |
| }
 | |
| 
 | |
| // LoadOne calls structr.Cache{}.LoadOne(), using a cached structr.Index{} by 'index' name.
 | |
| // Note: this also handles conversion of the untyped (any) keys to structr.Key{} via structr.Index{}.
 | |
| func (c *StructCache[T]) LoadOne(index string, load func() (T, error), key ...any) (T, error) {
 | |
| 	i := c.index[index]
 | |
| 	return c.cache.LoadOne(i, i.Key(key...), load)
 | |
| }
 | |
| 
 | |
| // LoadIDs calls structr.Cache{}.Load(), using a cached structr.Index{} by 'index' name. Note: this also handles
 | |
| // conversion of the ID strings to structr.Key{} via structr.Index{}. Strong typing is used for caller convenience.
 | |
| //
 | |
| // If you need to load multiple cache keys other than by ID strings, please create another convenience wrapper.
 | |
| func (c *StructCache[T]) LoadIDs(index string, ids []string, load func([]string) ([]T, error)) ([]T, error) {
 | |
| 	i := c.index[index]
 | |
| 	if i == nil {
 | |
| 		// we only perform this check here as
 | |
| 		// we're going to use the index before
 | |
| 		// passing it to cache in main .Load().
 | |
| 		panic("missing index for cache type")
 | |
| 	}
 | |
| 
 | |
| 	// Generate cache keys for ID types.
 | |
| 	keys := make([]structr.Key, len(ids))
 | |
| 	for x, id := range ids {
 | |
| 		keys[x] = i.Key(id)
 | |
| 	}
 | |
| 
 | |
| 	// Pass loader callback with wrapper onto main cache load function.
 | |
| 	return c.cache.Load(i, keys, func(uncached []structr.Key) ([]T, error) {
 | |
| 		uncachedIDs := make([]string, len(uncached))
 | |
| 		for i := range uncached {
 | |
| 			uncachedIDs[i] = uncached[i].Values()[0].(string)
 | |
| 		}
 | |
| 		return load(uncachedIDs)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| // LoadIDs2Part works as LoadIDs, except using a two-part key,
 | |
| // where the first part is an ID shared by all the objects,
 | |
| // and the second part is a list of per-object IDs.
 | |
| func (c *StructCache[T]) LoadIDs2Part(index string, id1 string, id2s []string, load func(string, []string) ([]T, error)) ([]T, error) {
 | |
| 	i := c.index[index]
 | |
| 	if i == nil {
 | |
| 		// we only perform this check here as
 | |
| 		// we're going to use the index before
 | |
| 		// passing it to cache in main .Load().
 | |
| 		panic("missing index for cache type")
 | |
| 	}
 | |
| 
 | |
| 	// Generate cache keys for two-part IDs.
 | |
| 	keys := make([]structr.Key, len(id2s))
 | |
| 	for x, id2 := range id2s {
 | |
| 		keys[x] = i.Key(id1, id2)
 | |
| 	}
 | |
| 
 | |
| 	// Pass loader callback with wrapper onto main cache load function.
 | |
| 	return c.cache.Load(i, keys, func(uncached []structr.Key) ([]T, error) {
 | |
| 		uncachedIDs := make([]string, len(uncached))
 | |
| 		for i := range uncached {
 | |
| 			uncachedIDs[i] = uncached[i].Values()[1].(string)
 | |
| 		}
 | |
| 		return load(id1, uncachedIDs)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| // Store: see structr.Cache{}.Store().
 | |
| func (c *StructCache[T]) Store(value T, store func() error) error {
 | |
| 	return c.cache.Store(value, store)
 | |
| }
 | |
| 
 | |
| // Invalidate calls structr.Cache{}.Invalidate(), using a cached structr.Index{} by 'index' name.
 | |
| // Note: this also handles conversion of the untyped (any) keys to structr.Key{} via structr.Index{}.
 | |
| func (c *StructCache[T]) Invalidate(index string, key ...any) {
 | |
| 	i := c.index[index]
 | |
| 	c.cache.Invalidate(i, i.Key(key...))
 | |
| }
 | |
| 
 | |
| // InvalidateIDs calls structr.Cache{}.Invalidate(), using a cached structr.Index{} by 'index' name. Note: this also
 | |
| // handles conversion of the ID strings to structr.Key{} via structr.Index{}. Strong typing is used for caller convenience.
 | |
| //
 | |
| // If you need to invalidate multiple cache keys other than by ID strings, please create another convenience wrapper.
 | |
| func (c *StructCache[T]) InvalidateIDs(index string, ids []string) {
 | |
| 	i := c.index[index]
 | |
| 	if i == nil {
 | |
| 		// we only perform this check here as
 | |
| 		// we're going to use the index before
 | |
| 		// passing it to cache in main .Load().
 | |
| 		panic("missing index for cache type")
 | |
| 	}
 | |
| 
 | |
| 	// Generate cache keys for ID types.
 | |
| 	keys := make([]structr.Key, len(ids))
 | |
| 	for x, id := range ids {
 | |
| 		keys[x] = i.Key(id)
 | |
| 	}
 | |
| 
 | |
| 	// Pass to main invalidate func.
 | |
| 	c.cache.Invalidate(i, keys...)
 | |
| }
 | |
| 
 | |
| // Trim: see structr.Cache{}.Trim().
 | |
| func (c *StructCache[T]) Trim(perc float64) {
 | |
| 	c.cache.Trim(perc)
 | |
| }
 | |
| 
 | |
| // Clear: see structr.Cache{}.Clear().
 | |
| func (c *StructCache[T]) Clear() {
 | |
| 	c.cache.Clear()
 | |
| }
 | |
| 
 | |
| // Len: see structr.Cache{}.Len().
 | |
| func (c *StructCache[T]) Len() int {
 | |
| 	return c.cache.Len()
 | |
| }
 | |
| 
 | |
| // Cap: see structr.Cache{}.Cap().
 | |
| func (c *StructCache[T]) Cap() int {
 | |
| 	return c.cache.Cap()
 | |
| }
 |