mirror of
https://codeberg.org/superseriousbusiness/gotosocial.git
synced 2024-12-30 12:58:16 +03:00
87cff71af9
* persist queued worker tasks to database on shutdown, fill worker queues from database on startup
* ensure the tasks are sorted by creation time before pushing them
* add migration to insert WorkerTask{} into database, add test for worker task persistence
* add test for recovering worker queues from database
* quick tweak
* whoops we ended up with double cleaner job scheduling
* insert each task separately, because bun is throwing some reflection error??
* add specific checking of cancelled worker contexts
* add http request signing to deliveries recovered from database
* add test for outgoing public key ID being correctly set on delivery
* replace select with Queue.PopCtx()
* get rid of loop now we don't use it
* remove field now we don't use it
* ensure that signing func is set
* header values weren't being copied over 🤦
* use ptr for httpclient.Request in delivery
* move worker queue filling to later in server init process
* fix rebase issues
* make logging less shouty
* use slices.Delete() instead of copying / reslicing
* have database return tasks in ascending order instead of sorting them
* add a 1 minute timeout to persisting worker queues
314 lines
7.4 KiB
Go
314 lines
7.4 KiB
Go
// GoToSocial
|
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package delivery
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"slices"
|
|
"time"
|
|
|
|
"codeberg.org/gruf/go-runners"
|
|
"codeberg.org/gruf/go-structr"
|
|
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
|
"github.com/superseriousbusiness/gotosocial/internal/httpclient"
|
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
|
"github.com/superseriousbusiness/gotosocial/internal/queue"
|
|
"github.com/superseriousbusiness/gotosocial/internal/util"
|
|
)
|
|
|
|
// WorkerPool wraps multiple Worker{}s in
|
|
// a singular struct for easy multi start/stop.
|
|
type WorkerPool struct {
|
|
|
|
// Client defines httpclient.Client{}
|
|
// passed to each of delivery pool Worker{}s.
|
|
Client *httpclient.Client
|
|
|
|
// Queue is the embedded queue.StructQueue{}
|
|
// passed to each of delivery pool Worker{}s.
|
|
Queue queue.StructQueue[*Delivery]
|
|
|
|
// internal fields.
|
|
workers []*Worker
|
|
}
|
|
|
|
// Init will initialize the Worker{} pool
|
|
// with given http client, request queue to pull
|
|
// from and number of delivery workers to spawn.
|
|
func (p *WorkerPool) Init(client *httpclient.Client) {
|
|
p.Client = client
|
|
p.Queue.Init(structr.QueueConfig[*Delivery]{
|
|
Indices: []structr.IndexConfig{
|
|
{Fields: "ActorID", Multiple: true},
|
|
{Fields: "ObjectID", Multiple: true},
|
|
{Fields: "TargetID", Multiple: true},
|
|
},
|
|
})
|
|
}
|
|
|
|
// Start will attempt to start 'n' Worker{}s.
|
|
func (p *WorkerPool) Start(n int) {
|
|
// Check whether workers are
|
|
// set (is already running).
|
|
ok := (len(p.workers) > 0)
|
|
if ok {
|
|
return
|
|
}
|
|
|
|
// Allocate new workers slice.
|
|
p.workers = make([]*Worker, n)
|
|
for i := range p.workers {
|
|
|
|
// Allocate new Worker{}.
|
|
p.workers[i] = new(Worker)
|
|
p.workers[i].Client = p.Client
|
|
p.workers[i].Queue = &p.Queue
|
|
|
|
// Attempt to start worker.
|
|
// Return bool not useful
|
|
// here, as true = started,
|
|
// false = already running.
|
|
_ = p.workers[i].Start()
|
|
}
|
|
}
|
|
|
|
// Stop will attempt to stop contained Worker{}s.
|
|
func (p *WorkerPool) Stop() {
|
|
// Check whether workers are
|
|
// set (is currently running).
|
|
ok := (len(p.workers) == 0)
|
|
if ok {
|
|
return
|
|
}
|
|
|
|
// Stop all running workers.
|
|
for i := range p.workers {
|
|
|
|
// return bool not useful
|
|
// here, as true = stopped,
|
|
// false = never running.
|
|
_ = p.workers[i].Stop()
|
|
}
|
|
|
|
// Unset workers slice.
|
|
p.workers = p.workers[:0]
|
|
}
|
|
|
|
// Worker wraps an httpclient.Client{} to feed
|
|
// from queue.StructQueue{} for ActivityPub reqs
|
|
// to deliver. It does so while prioritizing new
|
|
// queued requests over backlogged retries.
|
|
type Worker struct {
|
|
|
|
// Client is the httpclient.Client{} that
|
|
// delivery worker will use for requests.
|
|
Client *httpclient.Client
|
|
|
|
// Queue is the Delivery{} message queue
|
|
// that delivery worker will feed from.
|
|
Queue *queue.StructQueue[*Delivery]
|
|
|
|
// internal fields.
|
|
backlog []*Delivery
|
|
service runners.Service
|
|
}
|
|
|
|
// Start will attempt to start the Worker{}.
|
|
func (w *Worker) Start() bool {
|
|
return w.service.GoRun(w.run)
|
|
}
|
|
|
|
// Stop will attempt to stop the Worker{}.
|
|
func (w *Worker) Stop() bool {
|
|
return w.service.Stop()
|
|
}
|
|
|
|
// run wraps process to restart on any panic.
|
|
func (w *Worker) run(ctx context.Context) {
|
|
if w.Client == nil || w.Queue == nil {
|
|
panic("not yet initialized")
|
|
}
|
|
log.Debugf(ctx, "%p: starting worker", w)
|
|
defer log.Debugf(ctx, "%p: stopped worker", w)
|
|
util.Must(func() { w.process(ctx) })
|
|
}
|
|
|
|
// process is the main delivery worker processing routine.
|
|
func (w *Worker) process(ctx context.Context) bool {
|
|
if w.Client == nil || w.Queue == nil {
|
|
// we perform this check here just
|
|
// to ensure the compiler knows these
|
|
// variables aren't nil in the loop,
|
|
// even if already checked by caller.
|
|
panic("not yet initialized")
|
|
}
|
|
|
|
loop:
|
|
for {
|
|
// Before trying to get
|
|
// next delivery, check
|
|
// context still valid.
|
|
if ctx.Err() != nil {
|
|
return true
|
|
}
|
|
|
|
// Get next delivery.
|
|
dlv, ok := w.next(ctx)
|
|
if !ok {
|
|
return true
|
|
}
|
|
|
|
// Check whether backoff required.
|
|
const min = 100 * time.Millisecond
|
|
if d := dlv.backoff(); d > min {
|
|
|
|
// Start backoff sleep timer.
|
|
backoff := time.NewTimer(d)
|
|
|
|
select {
|
|
case <-ctx.Done():
|
|
// Main ctx
|
|
// cancelled.
|
|
backoff.Stop()
|
|
return true
|
|
|
|
case <-w.Queue.Wait():
|
|
// A new message was
|
|
// queued, re-add this
|
|
// to backlog + retry.
|
|
w.pushBacklog(dlv)
|
|
backoff.Stop()
|
|
continue loop
|
|
|
|
case <-backoff.C:
|
|
// success!
|
|
}
|
|
}
|
|
|
|
// Attempt delivery of AP request.
|
|
rsp, retry, err := w.Client.DoOnce(
|
|
dlv.Request,
|
|
)
|
|
|
|
switch {
|
|
case err == nil:
|
|
// Ensure body closed.
|
|
_ = rsp.Body.Close()
|
|
continue loop
|
|
|
|
case errors.Is(err, context.Canceled) &&
|
|
ctx.Err() != nil:
|
|
// In the case of our own context
|
|
// being cancelled, push delivery
|
|
// back onto queue for persisting.
|
|
//
|
|
// Note we specifically check against
|
|
// context.Canceled here as it will
|
|
// be faster than the mutex lock of
|
|
// ctx.Err(), so gives an initial
|
|
// faster check in the if-clause.
|
|
w.Queue.Push(dlv)
|
|
continue loop
|
|
|
|
case !retry:
|
|
// Drop deliveries when no
|
|
// retry requested, or they
|
|
// reached max (either).
|
|
continue loop
|
|
}
|
|
|
|
// Determine next delivery attempt.
|
|
backoff := dlv.Request.BackOff()
|
|
dlv.next = time.Now().Add(backoff)
|
|
|
|
// Push to backlog.
|
|
w.pushBacklog(dlv)
|
|
}
|
|
}
|
|
|
|
// next gets the next available delivery, blocking until available if necessary.
|
|
func (w *Worker) next(ctx context.Context) (*Delivery, bool) {
|
|
// Try a fast-pop of queued
|
|
// delivery before anything.
|
|
dlv, ok := w.Queue.Pop()
|
|
|
|
if !ok {
|
|
// Check the backlog.
|
|
if len(w.backlog) > 0 {
|
|
|
|
// Sort by 'next' time.
|
|
sortDeliveries(w.backlog)
|
|
|
|
// Pop next delivery.
|
|
dlv := w.popBacklog()
|
|
|
|
return dlv, true
|
|
}
|
|
|
|
// Block on next delivery push
|
|
// OR worker context canceled.
|
|
dlv, ok = w.Queue.PopCtx(ctx)
|
|
if !ok {
|
|
return nil, false
|
|
}
|
|
}
|
|
|
|
// Replace request context for worker state canceling.
|
|
ctx = gtscontext.WithValues(ctx, dlv.Request.Context())
|
|
dlv.Request.Request = dlv.Request.Request.WithContext(ctx)
|
|
|
|
return dlv, true
|
|
}
|
|
|
|
// popBacklog pops next available from the backlog.
|
|
func (w *Worker) popBacklog() *Delivery {
|
|
if len(w.backlog) == 0 {
|
|
return nil
|
|
}
|
|
|
|
// Pop from backlog.
|
|
dlv := w.backlog[0]
|
|
|
|
// Shift backlog down by one.
|
|
copy(w.backlog, w.backlog[1:])
|
|
w.backlog = w.backlog[:len(w.backlog)-1]
|
|
|
|
return dlv
|
|
}
|
|
|
|
// pushBacklog pushes the given delivery to backlog.
|
|
func (w *Worker) pushBacklog(dlv *Delivery) {
|
|
w.backlog = append(w.backlog, dlv)
|
|
}
|
|
|
|
// sortDeliveries sorts deliveries according
|
|
// to when is the first requiring re-attempt.
|
|
func sortDeliveries(d []*Delivery) {
|
|
slices.SortFunc(d, func(a, b *Delivery) int {
|
|
const k = +1
|
|
switch {
|
|
case a.next.Before(b.next):
|
|
return +k
|
|
case b.next.Before(a.next):
|
|
return -k
|
|
default:
|
|
return 0
|
|
}
|
|
})
|
|
}
|