mirror of
				https://github.com/superseriousbusiness/gotosocial.git
				synced 2025-10-30 23:32:25 -05:00 
			
		
		
		
	* start work rewriting timeline cache type
* further work rewriting timeline caching
* more work integration new timeline code
* remove old code
* add local timeline, fix up merge conflicts
* remove old use of go-bytes
* implement new timeline code into more areas of codebase, pull in latest go-mangler, go-mutexes, go-structr
* remove old timeline package, add local timeline cache
* remove references to old timeline types that needed starting up in tests
* start adding page validation
* fix test-identified timeline cache package issues
* fix up more tests, fix missing required changes, etc
* add exclusion for test.out in gitignore
* clarify some things better in code comments
* tweak cache size limits
* fix list timeline cache fetching
* further list timeline fixes
* linter, ssssssssshhhhhhhhhhhh please
* fix linter hints
* reslice the output if it's beyond length of 'lim'
* remove old timeline initialization code, bump go-structr to v0.9.4
* continued from previous commit
* improved code comments
* don't allow multiple entries for BoostOfID values to prevent repeated boosts of same boosts
* finish writing more code comments
* some variable renaming, for ease of following
* change the way we update lo,hi paging values during timeline load
* improved code comments for updated / returned lo , hi paging values
* finish writing code comments for the StatusTimeline{} type itself
* fill in more code comments
* update go-structr version to latest with changed timeline unique indexing logic
* have a local and public timeline *per user*
* rewrite calls to public / local timeline calls
* remove the zero length check, as lo, hi values might still be set
* simplify timeline cache loading, fix lo/hi returns, fix timeline invalidation side-effects missing for some federated actions
* swap the lo, hi values 🤦
* add (now) missing slice reverse of tag timeline statuses when paging ASC
* remove local / public caches (is out of scope for this work), share more timeline code
* remove unnecessary change
* again, remove more unused code
* remove unused function to appease the linter
* move boost checking to prepare function
* fix use of timeline.lastOrder, fix incorrect range functions used
* remove comments for repeat code
* remove the boost logic from prepare function
* do a maximum of 5 loads, not 10
* add repeat boost filtering logic, update go-structr, general improvements
* more code comments
* add important note
* fix timeline tests now that timelines are returned in page order
* remove unused field
* add StatusTimeline{} tests
* add more status timeline tests
* start adding preloading support
* ensure repeat boosts are marked in preloaded entries
* share a bunch of the database load code in timeline cache, don't clear timelines on relationship change
* add logic to allow dynamic clear / preloading of timelines
* comment-out unused functions, but leave in place as we might end-up using them
* fix timeline preload state check
* much improved status timeline code comments
* more code comments, don't bother inserting statuses if timeline not preloaded
* shift around some logic to make sure things aren't accidentally left set
* finish writing code comments
* remove trim-after-insert behaviour
* fix-up some comments referring to old logic
* remove unsetting of lo, hi
* fix preload repeatBoost checking logic
* don't return on status filter errors, these are usually transient
* better concurrency safety in Clear() and Done()
* fix test broken due to addition of preloader
* fix repeatBoost logic that doesn't account for already-hidden repeatBoosts
* ensure edit submodels are dropped on cache insertion
* update code-comment to expand CAS accronym
* use a plus1hULID() instead of 24h
* remove unused functions
* add note that public / local timeline requester can be nil
* fix incorrect visibility filtering of tag timeline statuses
* ensure we filter home timeline statuses on local only
* some small re-orderings to confirm query params in correct places
* fix the local only home timeline filter func
		
	
			
		
			
				
	
	
		
			318 lines
		
	
	
	
		
			7.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			318 lines
		
	
	
	
		
			7.5 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // GoToSocial
 | |
| // Copyright (C) GoToSocial Authors admin@gotosocial.org
 | |
| // SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful,
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package paging
 | |
| 
 | |
| import (
 | |
| 	"net/url"
 | |
| 	"slices"
 | |
| 	"strconv"
 | |
| )
 | |
| 
 | |
| type Page struct {
 | |
| 	// Min is the Page's lower limit value.
 | |
| 	Min Boundary
 | |
| 
 | |
| 	// Max is this Page's upper limit value.
 | |
| 	Max Boundary
 | |
| 
 | |
| 	// Limit will limit the returned
 | |
| 	// page of items to at most 'limit'.
 | |
| 	Limit int
 | |
| }
 | |
| 
 | |
| // GetMin is a small helper function to return minimum boundary value (checking for nil page).
 | |
| func (p *Page) GetMin() string {
 | |
| 	if p == nil {
 | |
| 		return ""
 | |
| 	}
 | |
| 	return p.Min.Value
 | |
| }
 | |
| 
 | |
| // GetMax is a small helper function to return maximum boundary value (checking for nil page).
 | |
| func (p *Page) GetMax() string {
 | |
| 	if p == nil {
 | |
| 		return ""
 | |
| 	}
 | |
| 	return p.Max.Value
 | |
| }
 | |
| 
 | |
| // GetLimit is a small helper function to return limit (checking for nil page and unusable limit).
 | |
| func (p *Page) GetLimit() int {
 | |
| 	if p == nil || p.Limit < 0 {
 | |
| 		return 0
 | |
| 	}
 | |
| 	return p.Limit
 | |
| }
 | |
| 
 | |
| // GetOrder is a small helper function to return page sort ordering (checking for nil page).
 | |
| func (p *Page) GetOrder() Order {
 | |
| 	if p == nil {
 | |
| 		return 0
 | |
| 	}
 | |
| 	return p.Order()
 | |
| }
 | |
| 
 | |
| // Order is a small helper function to return page sort ordering.
 | |
| func (p *Page) Order() Order {
 | |
| 	switch {
 | |
| 	case p.Min.Order != 0:
 | |
| 		return p.Min.Order
 | |
| 	case p.Max.Order != 0:
 | |
| 		return p.Max.Order
 | |
| 	default:
 | |
| 		return 0
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // Page will page the given slice of input according
 | |
| // to the receiving Page's minimum, maximum and limit.
 | |
| // NOTE: input slice MUST be sorted according to the order is
 | |
| // expected to be paged in, i.e. it is currently sorted
 | |
| // according to Page.Order(). Sorted data isn't always according
 | |
| // to string inequalities so this CANNOT be checked here.
 | |
| func (p *Page) Page(in []string) []string {
 | |
| 	if p == nil {
 | |
| 		// no paging.
 | |
| 		return in
 | |
| 	}
 | |
| 
 | |
| 	if p.Order().Ascending() {
 | |
| 		// Sort type is ascending, input
 | |
| 		// data is assumed to be ascending.
 | |
| 
 | |
| 		if minIdx := p.Min.Find(in); minIdx != -1 {
 | |
| 			// Reslice skipping up to min.
 | |
| 			in = in[minIdx+1:]
 | |
| 		}
 | |
| 
 | |
| 		if maxIdx := p.Max.Find(in); maxIdx != -1 {
 | |
| 			// Reslice stripping past max.
 | |
| 			in = in[:maxIdx]
 | |
| 		}
 | |
| 
 | |
| 		if p.Limit > 0 && p.Limit < len(in) {
 | |
| 			// Reslice input to limit.
 | |
| 			in = in[:p.Limit]
 | |
| 		}
 | |
| 
 | |
| 		if len(in) > 1 {
 | |
| 			// Clone input before
 | |
| 			// any modifications.
 | |
| 			in = slices.Clone(in)
 | |
| 
 | |
| 			// Output slice must
 | |
| 			// ALWAYS be descending.
 | |
| 			slices.Reverse(in)
 | |
| 		}
 | |
| 	} else {
 | |
| 		// Default sort is descending,
 | |
| 		// catching all cases when NOT
 | |
| 		// ascending (even zero value).
 | |
| 
 | |
| 		if maxIdx := p.Max.Find(in); maxIdx != -1 {
 | |
| 			// Reslice skipping up to max.
 | |
| 			in = in[maxIdx+1:]
 | |
| 		}
 | |
| 
 | |
| 		if minIdx := p.Min.Find(in); minIdx != -1 {
 | |
| 			// Reslice stripping past min.
 | |
| 			in = in[:minIdx]
 | |
| 		}
 | |
| 
 | |
| 		if p.Limit > 0 && p.Limit < len(in) {
 | |
| 			// Reslice input to limit.
 | |
| 			in = in[:p.Limit]
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return in
 | |
| }
 | |
| 
 | |
| // Page_PageFunc is functionally equivalent to Page{}.Page(), but for an arbitrary type with ID.
 | |
| // Note: this is not a Page{} method as Go generics are not supported in method receiver functions.
 | |
| func Page_PageFunc[WithID any](p *Page, in []WithID, get func(WithID) string) []WithID { //nolint:revive
 | |
| 	if p == nil {
 | |
| 		// no paging.
 | |
| 		return in
 | |
| 	}
 | |
| 
 | |
| 	if p.Order().Ascending() {
 | |
| 		// Sort type is ascending, input
 | |
| 		// data is assumed to be ascending.
 | |
| 
 | |
| 		if minIdx := Boundary_FindFunc(p.Min, in, get); minIdx != -1 {
 | |
| 			// Reslice skipping up to min.
 | |
| 			in = in[minIdx+1:]
 | |
| 		}
 | |
| 
 | |
| 		if maxIdx := Boundary_FindFunc(p.Max, in, get); maxIdx != -1 {
 | |
| 			// Reslice stripping past max.
 | |
| 			in = in[:maxIdx]
 | |
| 		}
 | |
| 
 | |
| 		if p.Limit > 0 && p.Limit < len(in) {
 | |
| 			// Reslice input to limit.
 | |
| 			in = in[:p.Limit]
 | |
| 		}
 | |
| 
 | |
| 		if len(in) > 1 {
 | |
| 			// Clone input before
 | |
| 			// any modifications.
 | |
| 			in = slices.Clone(in)
 | |
| 
 | |
| 			// Output slice must
 | |
| 			// ALWAYS be descending.
 | |
| 			slices.Reverse(in)
 | |
| 		}
 | |
| 	} else {
 | |
| 		// Default sort is descending,
 | |
| 		// catching all cases when NOT
 | |
| 		// ascending (even zero value).
 | |
| 
 | |
| 		if maxIdx := Boundary_FindFunc(p.Max, in, get); maxIdx != -1 {
 | |
| 			// Reslice skipping up to max.
 | |
| 			in = in[maxIdx+1:]
 | |
| 		}
 | |
| 
 | |
| 		if minIdx := Boundary_FindFunc(p.Min, in, get); minIdx != -1 {
 | |
| 			// Reslice stripping past min.
 | |
| 			in = in[:minIdx]
 | |
| 		}
 | |
| 
 | |
| 		if p.Limit > 0 && p.Limit < len(in) {
 | |
| 			// Reslice input to limit.
 | |
| 			in = in[:p.Limit]
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return in
 | |
| }
 | |
| 
 | |
| // Prev creates a new instance for the next returnable page, using
 | |
| // given max value. This will always assume DESCENDING for Mastodon
 | |
| // API compatibility, but in case of change it can support both.
 | |
| func (p *Page) Next(lo, hi string) *Page {
 | |
| 	if p == nil || lo == "" || hi == "" {
 | |
| 		// no paging.
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	// Create new page.
 | |
| 	p2 := new(Page)
 | |
| 
 | |
| 	// Set original limit.
 | |
| 	p2.Limit = p.Limit
 | |
| 
 | |
| 	// NOTE:
 | |
| 	// We ALWAYS assume the order
 | |
| 	// when creating next / prev
 | |
| 	// links is DESCENDING. It will
 | |
| 	// always use prev: ?max_name
 | |
| 	p2.Min = p.Min.new("")
 | |
| 	p2.Max = p.Max.new(lo)
 | |
| 	p2.Min.Order = OrderDescending
 | |
| 	p2.Max.Order = OrderDescending
 | |
| 
 | |
| 	return p2
 | |
| }
 | |
| 
 | |
| // Prev creates a new instance for the prev returnable page, using
 | |
| // given min value. This will always assume DESCENDING for Mastodon
 | |
| // API compatibility, but in case of change it can support both.
 | |
| func (p *Page) Prev(lo, hi string) *Page {
 | |
| 	if p == nil || lo == "" || hi == "" {
 | |
| 		// no paging.
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	// Create new page.
 | |
| 	p2 := new(Page)
 | |
| 
 | |
| 	// Set original limit.
 | |
| 	p2.Limit = p.Limit
 | |
| 
 | |
| 	// NOTE:
 | |
| 	// We ALWAYS assume the order
 | |
| 	// when creating next / prev
 | |
| 	// links is DESCENDING. It will
 | |
| 	// always use prev: ?min_name
 | |
| 	p2.Min = p.Min.new(hi)
 | |
| 	p2.Max = p.Max.new("")
 | |
| 	p2.Min.Order = OrderDescending
 | |
| 	p2.Max.Order = OrderDescending
 | |
| 
 | |
| 	return p2
 | |
| }
 | |
| 
 | |
| // ToLink performs ToLinkURL() and calls .String() on the resulting URL.
 | |
| func (p *Page) ToLink(proto, host, path string, queryParams url.Values) string {
 | |
| 	u := p.ToLinkURL(proto, host, path, queryParams)
 | |
| 	if u == nil {
 | |
| 		return ""
 | |
| 	}
 | |
| 	return u.String()
 | |
| }
 | |
| 
 | |
| // ToLink builds a URL link for given endpoint information and extra query parameters,
 | |
| // appending this Page's minimum / maximum boundaries and available limit (if any).
 | |
| func (p *Page) ToLinkURL(proto, host, path string, queryParams url.Values) *url.URL {
 | |
| 	if p == nil {
 | |
| 		// no paging.
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	if queryParams == nil {
 | |
| 		// Allocate new query parameters.
 | |
| 		queryParams = make(url.Values)
 | |
| 	} else {
 | |
| 		// Before edit clone existing params.
 | |
| 		queryParams = cloneQuery(queryParams)
 | |
| 	}
 | |
| 
 | |
| 	if p.Min.Value != "" {
 | |
| 		// Set page-minimum cursor value.
 | |
| 		queryParams.Set(p.Min.Name, p.Min.Value)
 | |
| 	}
 | |
| 
 | |
| 	if p.Max.Value != "" {
 | |
| 		// Set page-maximum cursor value.
 | |
| 		queryParams.Set(p.Max.Name, p.Max.Value)
 | |
| 	}
 | |
| 
 | |
| 	if p.Limit > 0 {
 | |
| 		// A page limit query parameter is available.
 | |
| 		queryParams.Set("limit", strconv.Itoa(p.Limit))
 | |
| 	}
 | |
| 
 | |
| 	// Build URL string.
 | |
| 	return &url.URL{
 | |
| 		Scheme:   proto,
 | |
| 		Host:     host,
 | |
| 		Path:     path,
 | |
| 		RawQuery: queryParams.Encode(),
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // cloneQuery clones input map of url values.
 | |
| func cloneQuery(src url.Values) url.Values {
 | |
| 	dst := make(url.Values, len(src))
 | |
| 	for k, vs := range src {
 | |
| 		dst[k] = slices.Clone(vs)
 | |
| 	}
 | |
| 	return dst
 | |
| }
 |