mirror of
https://github.com/superseriousbusiness/gotosocial.git
synced 2025-10-29 07:12:25 -05:00
[performance] cache follow, follow request and block ID lists (#2027)
This commit is contained in:
parent
de148e9f9f
commit
ed2477ebea
29 changed files with 1283 additions and 335 deletions
227
internal/paging/paging.go
Normal file
227
internal/paging/paging.go
Normal file
|
|
@ -0,0 +1,227 @@
|
|||
// GoToSocial
|
||||
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
||||
// SPDX-License-Identifier: AGPL-3.0-or-later
|
||||
//
|
||||
// This program is free software: you can redistribute it and/or modify
|
||||
// it under the terms of the GNU Affero General Public License as published by
|
||||
// the Free Software Foundation, either version 3 of the License, or
|
||||
// (at your option) any later version.
|
||||
//
|
||||
// This program is distributed in the hope that it will be useful,
|
||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
// GNU Affero General Public License for more details.
|
||||
//
|
||||
// You should have received a copy of the GNU Affero General Public License
|
||||
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
package paging
|
||||
|
||||
import "golang.org/x/exp/slices"
|
||||
|
||||
// Pager provides a means of paging serialized IDs,
|
||||
// using the terminology of our API endpoint queries.
|
||||
type Pager struct {
|
||||
// SinceID will limit the returned
|
||||
// page of IDs to contain newer than
|
||||
// since ID (excluding it). Result
|
||||
// will be returned DESCENDING.
|
||||
SinceID string
|
||||
|
||||
// MinID will limit the returned
|
||||
// page of IDs to contain newer than
|
||||
// min ID (excluding it). Result
|
||||
// will be returned ASCENDING.
|
||||
MinID string
|
||||
|
||||
// MaxID will limit the returned
|
||||
// page of IDs to contain older
|
||||
// than (excluding) this max ID.
|
||||
MaxID string
|
||||
|
||||
// Limit will limit the returned
|
||||
// page of IDs to at most 'limit'.
|
||||
Limit int
|
||||
}
|
||||
|
||||
// Page will page the given slice of GoToSocial IDs according
|
||||
// to the receiving Pager's SinceID, MinID, MaxID and Limits.
|
||||
// NOTE THE INPUT SLICE MUST BE SORTED IN ASCENDING ORDER
|
||||
// (I.E. OLDEST ITEMS AT LOWEST INDICES, NEWER AT HIGHER).
|
||||
func (p *Pager) PageAsc(ids []string) []string {
|
||||
if p == nil {
|
||||
// no paging.
|
||||
return ids
|
||||
}
|
||||
|
||||
var asc bool
|
||||
|
||||
if p.SinceID != "" {
|
||||
// If a sinceID is given, we
|
||||
// page down i.e. descending.
|
||||
asc = false
|
||||
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.SinceID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "from here"
|
||||
ids = ids[i+1:]
|
||||
break
|
||||
}
|
||||
}
|
||||
} else if p.MinID != "" {
|
||||
// We only support minID if
|
||||
// no sinceID is provided.
|
||||
//
|
||||
// If a minID is given, we
|
||||
// page up, i.e. ascending.
|
||||
asc = true
|
||||
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.MinID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "from here"
|
||||
ids = ids[i+1:]
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if p.MaxID != "" {
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.MaxID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "up to here"
|
||||
ids = ids[:i]
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if !asc && len(ids) > 1 {
|
||||
var (
|
||||
// Start at front.
|
||||
i = 0
|
||||
|
||||
// Start at back.
|
||||
j = len(ids) - 1
|
||||
)
|
||||
|
||||
// Clone input IDs before
|
||||
// we perform modifications.
|
||||
ids = slices.Clone(ids)
|
||||
|
||||
for i < j {
|
||||
// Swap i,j index values in slice.
|
||||
ids[i], ids[j] = ids[j], ids[i]
|
||||
|
||||
// incr + decr,
|
||||
// looping until
|
||||
// they meet in
|
||||
// the middle.
|
||||
i++
|
||||
j--
|
||||
}
|
||||
}
|
||||
|
||||
if p.Limit > 0 && p.Limit < len(ids) {
|
||||
// Reslice IDs to given limit.
|
||||
ids = ids[:p.Limit]
|
||||
}
|
||||
|
||||
return ids
|
||||
}
|
||||
|
||||
// Page will page the given slice of GoToSocial IDs according
|
||||
// to the receiving Pager's SinceID, MinID, MaxID and Limits.
|
||||
// NOTE THE INPUT SLICE MUST BE SORTED IN ASCENDING ORDER.
|
||||
// (I.E. NEWEST ITEMS AT LOWEST INDICES, OLDER AT HIGHER).
|
||||
func (p *Pager) PageDesc(ids []string) []string {
|
||||
if p == nil {
|
||||
// no paging.
|
||||
return ids
|
||||
}
|
||||
|
||||
var asc bool
|
||||
|
||||
if p.MaxID != "" {
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.MaxID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "from here"
|
||||
ids = ids[i+1:]
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if p.SinceID != "" {
|
||||
// If a sinceID is given, we
|
||||
// page down i.e. descending.
|
||||
asc = false
|
||||
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.SinceID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "up to here"
|
||||
ids = ids[:i]
|
||||
break
|
||||
}
|
||||
}
|
||||
} else if p.MinID != "" {
|
||||
// We only support minID if
|
||||
// no sinceID is provided.
|
||||
//
|
||||
// If a minID is given, we
|
||||
// page up, i.e. ascending.
|
||||
asc = true
|
||||
|
||||
for i := 0; i < len(ids); i++ {
|
||||
if ids[i] == p.MinID {
|
||||
// Hit the boundary.
|
||||
// Reslice to be:
|
||||
// "up to here"
|
||||
ids = ids[:i]
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if asc && len(ids) > 1 {
|
||||
var (
|
||||
// Start at front.
|
||||
i = 0
|
||||
|
||||
// Start at back.
|
||||
j = len(ids) - 1
|
||||
)
|
||||
|
||||
// Clone input IDs before
|
||||
// we perform modifications.
|
||||
ids = slices.Clone(ids)
|
||||
|
||||
for i < j {
|
||||
// Swap i,j index values in slice.
|
||||
ids[i], ids[j] = ids[j], ids[i]
|
||||
|
||||
// incr + decr,
|
||||
// looping until
|
||||
// they meet in
|
||||
// the middle.
|
||||
i++
|
||||
j--
|
||||
}
|
||||
}
|
||||
|
||||
if p.Limit > 0 && p.Limit < len(ids) {
|
||||
// Reslice IDs to given limit.
|
||||
ids = ids[:p.Limit]
|
||||
}
|
||||
|
||||
return ids
|
||||
}
|
||||
171
internal/paging/paging_test.go
Normal file
171
internal/paging/paging_test.go
Normal file
|
|
@ -0,0 +1,171 @@
|
|||
// GoToSocial
|
||||
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
||||
// SPDX-License-Identifier: AGPL-3.0-or-later
|
||||
//
|
||||
// This program is free software: you can redistribute it and/or modify
|
||||
// it under the terms of the GNU Affero General Public License as published by
|
||||
// the Free Software Foundation, either version 3 of the License, or
|
||||
// (at your option) any later version.
|
||||
//
|
||||
// This program is distributed in the hope that it will be useful,
|
||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
// GNU Affero General Public License for more details.
|
||||
//
|
||||
// You should have received a copy of the GNU Affero General Public License
|
||||
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
package paging_test
|
||||
|
||||
import (
|
||||
"testing"
|
||||
|
||||
"github.com/superseriousbusiness/gotosocial/internal/paging"
|
||||
"golang.org/x/exp/slices"
|
||||
)
|
||||
|
||||
type Case struct {
|
||||
// Name is the test case name.
|
||||
Name string
|
||||
|
||||
// Input contains test case input ID slice.
|
||||
Input []string
|
||||
|
||||
// Expect contains expected test case output.
|
||||
Expect []string
|
||||
|
||||
// Page contains the paging function to use.
|
||||
Page func([]string) []string
|
||||
}
|
||||
|
||||
var cases = []Case{
|
||||
{
|
||||
Name: "min_id and max_id set",
|
||||
Input: []string{
|
||||
"064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
"064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
},
|
||||
Expect: []string{
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
},
|
||||
Page: (&paging.Pager{
|
||||
MinID: "064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
MaxID: "064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
}).PageAsc,
|
||||
},
|
||||
{
|
||||
Name: "min_id, max_id and limit set",
|
||||
Input: []string{
|
||||
"064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
"064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
},
|
||||
Expect: []string{
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
},
|
||||
Page: (&paging.Pager{
|
||||
MinID: "064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
MaxID: "064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
Limit: 5,
|
||||
}).PageAsc,
|
||||
},
|
||||
{
|
||||
Name: "min_id, max_id and too-large limit set",
|
||||
Input: []string{
|
||||
"064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
"064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
},
|
||||
Expect: []string{
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
},
|
||||
Page: (&paging.Pager{
|
||||
MinID: "064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
MaxID: "064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
Limit: 100,
|
||||
}).PageAsc,
|
||||
},
|
||||
{
|
||||
Name: "since_id and max_id set",
|
||||
Input: []string{
|
||||
"064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
"064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
},
|
||||
Expect: []string{
|
||||
"064Q5D7VK8H7WMJS399SHEPCB0",
|
||||
"064Q5D7VJYFBYSAH86KDBKZ6AC",
|
||||
"064Q5D7VJMWXZD3S1KT7RD51N8",
|
||||
"064Q5D7VJADJTPA3GW8WAX10TW",
|
||||
"064Q5D7VJ073XG9ZTWHA2KHN10",
|
||||
"064Q5D7VHMSW9DF3GCS088VAZC",
|
||||
"064Q5D7VH5F0JXG6W5NCQ3JCWW",
|
||||
"064Q5D7VGPTC4NK5T070VYSSF8",
|
||||
},
|
||||
Page: (&paging.Pager{
|
||||
SinceID: "064Q5D7VG6TPPQ46T09MHJ96FW",
|
||||
MaxID: "064Q5D7VKG5EQ43TYP71B4K6K0",
|
||||
}).PageAsc,
|
||||
},
|
||||
}
|
||||
|
||||
func TestPage(t *testing.T) {
|
||||
for _, c := range cases {
|
||||
t.Run(c.Name, func(t *testing.T) {
|
||||
// Page the input slice.
|
||||
out := c.Page(c.Input)
|
||||
|
||||
// Check paged output is as expected.
|
||||
if !slices.Equal(out, c.Expect) {
|
||||
t.Errorf("\nreceived=%v\nexpect%v\n", out, c.Expect)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
Loading…
Add table
Add a link
Reference in a new issue