2023-09-07 15:58:37 +01:00
|
|
|
// GoToSocial
|
|
|
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
|
|
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package paging
|
|
|
|
|
|
|
|
import (
|
|
|
|
"net/url"
|
2023-11-20 12:22:28 +00:00
|
|
|
"slices"
|
2023-09-07 15:58:37 +01:00
|
|
|
"strconv"
|
|
|
|
)
|
|
|
|
|
|
|
|
type Page struct {
|
|
|
|
// Min is the Page's lower limit value.
|
|
|
|
Min Boundary
|
|
|
|
|
|
|
|
// Max is this Page's upper limit value.
|
|
|
|
Max Boundary
|
|
|
|
|
|
|
|
// Limit will limit the returned
|
|
|
|
// page of items to at most 'limit'.
|
|
|
|
Limit int
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetMin is a small helper function to return minimum boundary value (checking for nil page).
|
|
|
|
func (p *Page) GetMin() string {
|
|
|
|
if p == nil {
|
|
|
|
return ""
|
|
|
|
}
|
|
|
|
return p.Min.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetMax is a small helper function to return maximum boundary value (checking for nil page).
|
|
|
|
func (p *Page) GetMax() string {
|
|
|
|
if p == nil {
|
|
|
|
return ""
|
|
|
|
}
|
|
|
|
return p.Max.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetLimit is a small helper function to return limit (checking for nil page and unusable limit).
|
|
|
|
func (p *Page) GetLimit() int {
|
|
|
|
if p == nil || p.Limit < 0 {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
return p.Limit
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetOrder is a small helper function to return page sort ordering (checking for nil page).
|
|
|
|
func (p *Page) GetOrder() Order {
|
|
|
|
if p == nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
return p.order()
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Page) order() Order {
|
|
|
|
switch {
|
2023-09-12 14:00:35 +01:00
|
|
|
case p.Min.Order != 0:
|
2023-09-07 15:58:37 +01:00
|
|
|
return p.Min.Order
|
2023-09-12 14:00:35 +01:00
|
|
|
case p.Max.Order != 0:
|
2023-09-07 15:58:37 +01:00
|
|
|
return p.Max.Order
|
|
|
|
default:
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Page will page the given slice of input according
|
|
|
|
// to the receiving Page's minimum, maximum and limit.
|
|
|
|
// NOTE: input slice MUST be sorted according to the order is
|
|
|
|
// expected to be paged in, i.e. it is currently sorted
|
|
|
|
// according to Page.Order(). Sorted data isn't always according
|
|
|
|
// to string inequalities so this CANNOT be checked here.
|
|
|
|
func (p *Page) Page(in []string) []string {
|
|
|
|
if p == nil {
|
|
|
|
// no paging.
|
|
|
|
return in
|
|
|
|
}
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.order().Ascending() {
|
2023-09-07 15:58:37 +01:00
|
|
|
// Sort type is ascending, input
|
|
|
|
// data is assumed to be ascending.
|
|
|
|
|
|
|
|
if minIdx := p.Min.Find(in); minIdx != -1 {
|
|
|
|
// Reslice skipping up to min.
|
|
|
|
in = in[minIdx+1:]
|
|
|
|
}
|
|
|
|
|
|
|
|
if maxIdx := p.Max.Find(in); maxIdx != -1 {
|
|
|
|
// Reslice stripping past max.
|
|
|
|
in = in[:maxIdx]
|
|
|
|
}
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.Limit > 0 && p.Limit < len(in) {
|
|
|
|
// Reslice input to limit.
|
|
|
|
in = in[:p.Limit]
|
|
|
|
}
|
|
|
|
|
2023-09-07 15:58:37 +01:00
|
|
|
if len(in) > 1 {
|
|
|
|
// Clone input before
|
|
|
|
// any modifications.
|
|
|
|
in = slices.Clone(in)
|
|
|
|
|
|
|
|
// Output slice must
|
|
|
|
// ALWAYS be descending.
|
2023-11-20 12:22:28 +00:00
|
|
|
slices.Reverse(in)
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
2023-09-12 14:00:35 +01:00
|
|
|
} else {
|
|
|
|
// Default sort is descending,
|
|
|
|
// catching all cases when NOT
|
|
|
|
// ascending (even zero value).
|
|
|
|
|
|
|
|
if maxIdx := p.Max.Find(in); maxIdx != -1 {
|
|
|
|
// Reslice skipping up to max.
|
|
|
|
in = in[maxIdx+1:]
|
|
|
|
}
|
|
|
|
|
|
|
|
if minIdx := p.Min.Find(in); minIdx != -1 {
|
|
|
|
// Reslice stripping past min.
|
|
|
|
in = in[:minIdx]
|
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.Limit > 0 && p.Limit < len(in) {
|
|
|
|
// Reslice input to limit.
|
|
|
|
in = in[:p.Limit]
|
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return in
|
|
|
|
}
|
|
|
|
|
2023-11-20 12:22:28 +00:00
|
|
|
// Page_PageFunc is functionally equivalent to Page{}.Page(), but for an arbitrary type with ID.
|
|
|
|
// Note: this is not a Page{} method as Go generics are not supported in method receiver functions.
|
|
|
|
func Page_PageFunc[WithID any](p *Page, in []WithID, get func(WithID) string) []WithID { //nolint:revive
|
|
|
|
if p == nil {
|
|
|
|
// no paging.
|
|
|
|
return in
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.order().Ascending() {
|
|
|
|
// Sort type is ascending, input
|
|
|
|
// data is assumed to be ascending.
|
|
|
|
|
|
|
|
if minIdx := Boundary_FindFunc(p.Min, in, get); minIdx != -1 {
|
|
|
|
// Reslice skipping up to min.
|
|
|
|
in = in[minIdx+1:]
|
|
|
|
}
|
|
|
|
|
|
|
|
if maxIdx := Boundary_FindFunc(p.Max, in, get); maxIdx != -1 {
|
|
|
|
// Reslice stripping past max.
|
|
|
|
in = in[:maxIdx]
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.Limit > 0 && p.Limit < len(in) {
|
|
|
|
// Reslice input to limit.
|
|
|
|
in = in[:p.Limit]
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(in) > 1 {
|
|
|
|
// Clone input before
|
|
|
|
// any modifications.
|
|
|
|
in = slices.Clone(in)
|
|
|
|
|
|
|
|
// Output slice must
|
|
|
|
// ALWAYS be descending.
|
|
|
|
slices.Reverse(in)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Default sort is descending,
|
|
|
|
// catching all cases when NOT
|
|
|
|
// ascending (even zero value).
|
|
|
|
|
|
|
|
if maxIdx := Boundary_FindFunc(p.Max, in, get); maxIdx != -1 {
|
|
|
|
// Reslice skipping up to max.
|
|
|
|
in = in[maxIdx+1:]
|
|
|
|
}
|
|
|
|
|
|
|
|
if minIdx := Boundary_FindFunc(p.Min, in, get); minIdx != -1 {
|
|
|
|
// Reslice stripping past min.
|
|
|
|
in = in[:minIdx]
|
|
|
|
}
|
|
|
|
|
|
|
|
if p.Limit > 0 && p.Limit < len(in) {
|
|
|
|
// Reslice input to limit.
|
|
|
|
in = in[:p.Limit]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return in
|
|
|
|
}
|
|
|
|
|
2023-09-07 15:58:37 +01:00
|
|
|
// Next creates a new instance for the next returnable page, using
|
|
|
|
// given max value. This preserves original limit and max key name.
|
2023-09-12 14:00:35 +01:00
|
|
|
func (p *Page) Next(lo, hi string) *Page {
|
|
|
|
if p == nil || lo == "" || hi == "" {
|
2023-09-07 15:58:37 +01:00
|
|
|
// no paging.
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create new page.
|
|
|
|
p2 := new(Page)
|
|
|
|
|
|
|
|
// Set original limit.
|
|
|
|
p2.Limit = p.Limit
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.order().Ascending() {
|
|
|
|
// When ascending, next page
|
|
|
|
// needs to start with min at
|
|
|
|
// the next highest value.
|
|
|
|
p2.Min = p.Min.new(hi)
|
|
|
|
p2.Max = p.Max.new("")
|
|
|
|
} else {
|
|
|
|
// When descending, next page
|
|
|
|
// needs to start with max at
|
|
|
|
// the next lowest value.
|
|
|
|
p2.Min = p.Min.new("")
|
|
|
|
p2.Max = p.Max.new(lo)
|
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
|
|
|
|
return p2
|
|
|
|
}
|
|
|
|
|
|
|
|
// Prev creates a new instance for the prev returnable page, using
|
|
|
|
// given min value. This preserves original limit and min key name.
|
2023-09-12 14:00:35 +01:00
|
|
|
func (p *Page) Prev(lo, hi string) *Page {
|
|
|
|
if p == nil || lo == "" || hi == "" {
|
2023-09-07 15:58:37 +01:00
|
|
|
// no paging.
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create new page.
|
|
|
|
p2 := new(Page)
|
|
|
|
|
|
|
|
// Set original limit.
|
|
|
|
p2.Limit = p.Limit
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.order().Ascending() {
|
|
|
|
// When ascending, prev page
|
|
|
|
// needs to start with max at
|
|
|
|
// the next lowest value.
|
|
|
|
p2.Min = p.Min.new("")
|
|
|
|
p2.Max = p.Max.new(lo)
|
|
|
|
} else {
|
|
|
|
// When descending, next page
|
|
|
|
// needs to start with max at
|
|
|
|
// the next lowest value.
|
|
|
|
p2.Min = p.Min.new(hi)
|
|
|
|
p2.Max = p.Max.new("")
|
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
|
|
|
|
return p2
|
|
|
|
}
|
|
|
|
|
2023-09-20 16:49:46 +01:00
|
|
|
// ToLink performs ToLinkURL() and calls .String() on the resulting URL.
|
|
|
|
func (p *Page) ToLink(proto, host, path string, queryParams url.Values) string {
|
|
|
|
u := p.ToLinkURL(proto, host, path, queryParams)
|
|
|
|
if u == nil {
|
|
|
|
return ""
|
|
|
|
}
|
|
|
|
return u.String()
|
|
|
|
}
|
|
|
|
|
2023-09-07 15:58:37 +01:00
|
|
|
// ToLink builds a URL link for given endpoint information and extra query parameters,
|
|
|
|
// appending this Page's minimum / maximum boundaries and available limit (if any).
|
2023-09-20 16:49:46 +01:00
|
|
|
func (p *Page) ToLinkURL(proto, host, path string, queryParams url.Values) *url.URL {
|
2023-09-07 15:58:37 +01:00
|
|
|
if p == nil {
|
|
|
|
// no paging.
|
2023-09-20 16:49:46 +01:00
|
|
|
return nil
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if queryParams == nil {
|
|
|
|
// Allocate new query parameters.
|
|
|
|
queryParams = make(url.Values)
|
2023-11-20 12:22:28 +00:00
|
|
|
} else {
|
|
|
|
// Before edit clone existing params.
|
|
|
|
queryParams = cloneQuery(queryParams)
|
2023-09-12 14:00:35 +01:00
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.Min.Value != "" {
|
2023-09-07 15:58:37 +01:00
|
|
|
// A page-minimum query parameter is available.
|
2023-11-20 12:22:28 +00:00
|
|
|
queryParams.Set(p.Min.Name, p.Min.Value)
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
|
|
|
|
2023-09-12 14:00:35 +01:00
|
|
|
if p.Max.Value != "" {
|
2023-09-07 15:58:37 +01:00
|
|
|
// A page-maximum query parameter is available.
|
2023-11-20 12:22:28 +00:00
|
|
|
queryParams.Set(p.Max.Name, p.Max.Value)
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if p.Limit > 0 {
|
2023-09-12 14:00:35 +01:00
|
|
|
// A page limit query parameter is available.
|
2023-11-20 12:22:28 +00:00
|
|
|
queryParams.Set("limit", strconv.Itoa(p.Limit))
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Build URL string.
|
2023-09-20 16:49:46 +01:00
|
|
|
return &url.URL{
|
2023-09-07 15:58:37 +01:00
|
|
|
Scheme: proto,
|
|
|
|
Host: host,
|
|
|
|
Path: path,
|
2023-09-12 14:00:35 +01:00
|
|
|
RawQuery: queryParams.Encode(),
|
2023-09-20 16:49:46 +01:00
|
|
|
}
|
2023-09-07 15:58:37 +01:00
|
|
|
}
|
2023-11-20 12:22:28 +00:00
|
|
|
|
|
|
|
// cloneQuery clones input map of url values.
|
|
|
|
func cloneQuery(src url.Values) url.Values {
|
|
|
|
dst := make(url.Values, len(src))
|
|
|
|
for k, vs := range src {
|
|
|
|
dst[k] = slices.Clone(vs)
|
|
|
|
}
|
|
|
|
return dst
|
|
|
|
}
|