420 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			420 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
| // Copyright 2020 The Gitea Authors. All rights reserved.
 | |
| // SPDX-License-Identifier: MIT
 | |
| 
 | |
| package queue
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"fmt"
 | |
| 	"runtime/pprof"
 | |
| 	"sync"
 | |
| 	"sync/atomic"
 | |
| 	"time"
 | |
| 
 | |
| 	"code.gitea.io/gitea/modules/json"
 | |
| 	"code.gitea.io/gitea/modules/log"
 | |
| 	"code.gitea.io/gitea/modules/util"
 | |
| )
 | |
| 
 | |
| // ByteFIFOQueueConfiguration is the configuration for a ByteFIFOQueue
 | |
| type ByteFIFOQueueConfiguration struct {
 | |
| 	WorkerPoolConfiguration
 | |
| 	Workers     int
 | |
| 	WaitOnEmpty bool
 | |
| }
 | |
| 
 | |
| var _ Queue = &ByteFIFOQueue{}
 | |
| 
 | |
| // ByteFIFOQueue is a Queue formed from a ByteFIFO and WorkerPool
 | |
| type ByteFIFOQueue struct {
 | |
| 	*WorkerPool
 | |
| 	byteFIFO           ByteFIFO
 | |
| 	typ                Type
 | |
| 	shutdownCtx        context.Context
 | |
| 	shutdownCtxCancel  context.CancelFunc
 | |
| 	terminateCtx       context.Context
 | |
| 	terminateCtxCancel context.CancelFunc
 | |
| 	exemplar           interface{}
 | |
| 	workers            int
 | |
| 	name               string
 | |
| 	lock               sync.Mutex
 | |
| 	waitOnEmpty        bool
 | |
| 	pushed             chan struct{}
 | |
| }
 | |
| 
 | |
| // NewByteFIFOQueue creates a new ByteFIFOQueue
 | |
| func NewByteFIFOQueue(typ Type, byteFIFO ByteFIFO, handle HandlerFunc, cfg, exemplar interface{}) (*ByteFIFOQueue, error) {
 | |
| 	configInterface, err := toConfig(ByteFIFOQueueConfiguration{}, cfg)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	config := configInterface.(ByteFIFOQueueConfiguration)
 | |
| 
 | |
| 	terminateCtx, terminateCtxCancel := context.WithCancel(context.Background())
 | |
| 	shutdownCtx, shutdownCtxCancel := context.WithCancel(terminateCtx)
 | |
| 
 | |
| 	q := &ByteFIFOQueue{
 | |
| 		byteFIFO:           byteFIFO,
 | |
| 		typ:                typ,
 | |
| 		shutdownCtx:        shutdownCtx,
 | |
| 		shutdownCtxCancel:  shutdownCtxCancel,
 | |
| 		terminateCtx:       terminateCtx,
 | |
| 		terminateCtxCancel: terminateCtxCancel,
 | |
| 		exemplar:           exemplar,
 | |
| 		workers:            config.Workers,
 | |
| 		name:               config.Name,
 | |
| 		waitOnEmpty:        config.WaitOnEmpty,
 | |
| 		pushed:             make(chan struct{}, 1),
 | |
| 	}
 | |
| 	q.WorkerPool = NewWorkerPool(func(data ...Data) (failed []Data) {
 | |
| 		for _, unhandled := range handle(data...) {
 | |
| 			if fail := q.PushBack(unhandled); fail != nil {
 | |
| 				failed = append(failed, fail)
 | |
| 			}
 | |
| 		}
 | |
| 		return failed
 | |
| 	}, config.WorkerPoolConfiguration)
 | |
| 
 | |
| 	return q, nil
 | |
| }
 | |
| 
 | |
| // Name returns the name of this queue
 | |
| func (q *ByteFIFOQueue) Name() string {
 | |
| 	return q.name
 | |
| }
 | |
| 
 | |
| // Push pushes data to the fifo
 | |
| func (q *ByteFIFOQueue) Push(data Data) error {
 | |
| 	return q.PushFunc(data, nil)
 | |
| }
 | |
| 
 | |
| // PushBack pushes data to the fifo
 | |
| func (q *ByteFIFOQueue) PushBack(data Data) error {
 | |
| 	if !assignableTo(data, q.exemplar) {
 | |
| 		return fmt.Errorf("unable to assign data: %v to same type as exemplar: %v in %s", data, q.exemplar, q.name)
 | |
| 	}
 | |
| 	bs, err := json.Marshal(data)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	defer func() {
 | |
| 		select {
 | |
| 		case q.pushed <- struct{}{}:
 | |
| 		default:
 | |
| 		}
 | |
| 	}()
 | |
| 	return q.byteFIFO.PushBack(q.terminateCtx, bs)
 | |
| }
 | |
| 
 | |
| // PushFunc pushes data to the fifo
 | |
| func (q *ByteFIFOQueue) PushFunc(data Data, fn func() error) error {
 | |
| 	if !assignableTo(data, q.exemplar) {
 | |
| 		return fmt.Errorf("unable to assign data: %v to same type as exemplar: %v in %s", data, q.exemplar, q.name)
 | |
| 	}
 | |
| 	bs, err := json.Marshal(data)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	defer func() {
 | |
| 		select {
 | |
| 		case q.pushed <- struct{}{}:
 | |
| 		default:
 | |
| 		}
 | |
| 	}()
 | |
| 	return q.byteFIFO.PushFunc(q.terminateCtx, bs, fn)
 | |
| }
 | |
| 
 | |
| // IsEmpty checks if the queue is empty
 | |
| func (q *ByteFIFOQueue) IsEmpty() bool {
 | |
| 	q.lock.Lock()
 | |
| 	defer q.lock.Unlock()
 | |
| 	if !q.WorkerPool.IsEmpty() {
 | |
| 		return false
 | |
| 	}
 | |
| 	return q.byteFIFO.Len(q.terminateCtx) == 0
 | |
| }
 | |
| 
 | |
| // NumberInQueue returns the number in the queue
 | |
| func (q *ByteFIFOQueue) NumberInQueue() int64 {
 | |
| 	q.lock.Lock()
 | |
| 	defer q.lock.Unlock()
 | |
| 	return q.byteFIFO.Len(q.terminateCtx) + q.WorkerPool.NumberInQueue()
 | |
| }
 | |
| 
 | |
| // Flush flushes the ByteFIFOQueue
 | |
| func (q *ByteFIFOQueue) Flush(timeout time.Duration) error {
 | |
| 	select {
 | |
| 	case q.pushed <- struct{}{}:
 | |
| 	default:
 | |
| 	}
 | |
| 	return q.WorkerPool.Flush(timeout)
 | |
| }
 | |
| 
 | |
| // Run runs the bytefifo queue
 | |
| func (q *ByteFIFOQueue) Run(atShutdown, atTerminate func(func())) {
 | |
| 	pprof.SetGoroutineLabels(q.baseCtx)
 | |
| 	atShutdown(q.Shutdown)
 | |
| 	atTerminate(q.Terminate)
 | |
| 	log.Debug("%s: %s Starting", q.typ, q.name)
 | |
| 
 | |
| 	_ = q.AddWorkers(q.workers, 0)
 | |
| 
 | |
| 	log.Trace("%s: %s Now running", q.typ, q.name)
 | |
| 	q.readToChan()
 | |
| 
 | |
| 	<-q.shutdownCtx.Done()
 | |
| 	log.Trace("%s: %s Waiting til done", q.typ, q.name)
 | |
| 	q.Wait()
 | |
| 
 | |
| 	log.Trace("%s: %s Waiting til cleaned", q.typ, q.name)
 | |
| 	q.CleanUp(q.terminateCtx)
 | |
| 	q.terminateCtxCancel()
 | |
| }
 | |
| 
 | |
| const maxBackOffTime = time.Second * 3
 | |
| 
 | |
| func (q *ByteFIFOQueue) readToChan() {
 | |
| 	// handle quick cancels
 | |
| 	select {
 | |
| 	case <-q.shutdownCtx.Done():
 | |
| 		// tell the pool to shutdown.
 | |
| 		q.baseCtxCancel()
 | |
| 		return
 | |
| 	default:
 | |
| 	}
 | |
| 
 | |
| 	// Default backoff values
 | |
| 	backOffTime := time.Millisecond * 100
 | |
| 	backOffTimer := time.NewTimer(0)
 | |
| 	util.StopTimer(backOffTimer)
 | |
| 
 | |
| 	paused, _ := q.IsPausedIsResumed()
 | |
| 
 | |
| loop:
 | |
| 	for {
 | |
| 		select {
 | |
| 		case <-paused:
 | |
| 			log.Trace("Queue %s pausing", q.name)
 | |
| 			_, resumed := q.IsPausedIsResumed()
 | |
| 
 | |
| 			select {
 | |
| 			case <-resumed:
 | |
| 				paused, _ = q.IsPausedIsResumed()
 | |
| 				log.Trace("Queue %s resuming", q.name)
 | |
| 				if q.HasNoWorkerScaling() {
 | |
| 					log.Warn(
 | |
| 						"Queue: %s is configured to be non-scaling and has no workers - this configuration is likely incorrect.\n"+
 | |
| 							"The queue will be paused to prevent data-loss with the assumption that you will add workers and unpause as required.", q.name)
 | |
| 					q.Pause()
 | |
| 					continue loop
 | |
| 				}
 | |
| 			case <-q.shutdownCtx.Done():
 | |
| 				// tell the pool to shutdown.
 | |
| 				q.baseCtxCancel()
 | |
| 				return
 | |
| 			case data, ok := <-q.dataChan:
 | |
| 				if !ok {
 | |
| 					return
 | |
| 				}
 | |
| 				if err := q.PushBack(data); err != nil {
 | |
| 					log.Error("Unable to push back data into queue %s", q.name)
 | |
| 				}
 | |
| 				atomic.AddInt64(&q.numInQueue, -1)
 | |
| 			}
 | |
| 		default:
 | |
| 		}
 | |
| 
 | |
| 		// empty the pushed channel
 | |
| 		select {
 | |
| 		case <-q.pushed:
 | |
| 		default:
 | |
| 		}
 | |
| 
 | |
| 		err := q.doPop()
 | |
| 
 | |
| 		util.StopTimer(backOffTimer)
 | |
| 
 | |
| 		if err != nil {
 | |
| 			if err == errQueueEmpty && q.waitOnEmpty {
 | |
| 				log.Trace("%s: %s Waiting on Empty", q.typ, q.name)
 | |
| 
 | |
| 				// reset the backoff time but don't set the timer
 | |
| 				backOffTime = 100 * time.Millisecond
 | |
| 			} else if err == errUnmarshal {
 | |
| 				// reset the timer and backoff
 | |
| 				backOffTime = 100 * time.Millisecond
 | |
| 				backOffTimer.Reset(backOffTime)
 | |
| 			} else {
 | |
| 				//  backoff
 | |
| 				backOffTimer.Reset(backOffTime)
 | |
| 			}
 | |
| 
 | |
| 			// Need to Backoff
 | |
| 			select {
 | |
| 			case <-q.shutdownCtx.Done():
 | |
| 				// Oops we've been shutdown whilst backing off
 | |
| 				// Make sure the worker pool is shutdown too
 | |
| 				q.baseCtxCancel()
 | |
| 				return
 | |
| 			case <-q.pushed:
 | |
| 				// Data has been pushed to the fifo (or flush has been called)
 | |
| 				// reset the backoff time
 | |
| 				backOffTime = 100 * time.Millisecond
 | |
| 				continue loop
 | |
| 			case <-backOffTimer.C:
 | |
| 				// Calculate the next backoff time
 | |
| 				backOffTime += backOffTime / 2
 | |
| 				if backOffTime > maxBackOffTime {
 | |
| 					backOffTime = maxBackOffTime
 | |
| 				}
 | |
| 				continue loop
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Reset the backoff time
 | |
| 		backOffTime = 100 * time.Millisecond
 | |
| 
 | |
| 		select {
 | |
| 		case <-q.shutdownCtx.Done():
 | |
| 			// Oops we've been shutdown
 | |
| 			// Make sure the worker pool is shutdown too
 | |
| 			q.baseCtxCancel()
 | |
| 			return
 | |
| 		default:
 | |
| 			continue loop
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| var (
 | |
| 	errQueueEmpty = fmt.Errorf("empty queue")
 | |
| 	errEmptyBytes = fmt.Errorf("empty bytes")
 | |
| 	errUnmarshal  = fmt.Errorf("failed to unmarshal")
 | |
| )
 | |
| 
 | |
| func (q *ByteFIFOQueue) doPop() error {
 | |
| 	q.lock.Lock()
 | |
| 	defer q.lock.Unlock()
 | |
| 	bs, err := q.byteFIFO.Pop(q.shutdownCtx)
 | |
| 	if err != nil {
 | |
| 		if err == context.Canceled {
 | |
| 			q.baseCtxCancel()
 | |
| 			return err
 | |
| 		}
 | |
| 		log.Error("%s: %s Error on Pop: %v", q.typ, q.name, err)
 | |
| 		return err
 | |
| 	}
 | |
| 	if len(bs) == 0 {
 | |
| 		if q.waitOnEmpty && q.byteFIFO.Len(q.shutdownCtx) == 0 {
 | |
| 			return errQueueEmpty
 | |
| 		}
 | |
| 		return errEmptyBytes
 | |
| 	}
 | |
| 
 | |
| 	data, err := unmarshalAs(bs, q.exemplar)
 | |
| 	if err != nil {
 | |
| 		log.Error("%s: %s Failed to unmarshal with error: %v", q.typ, q.name, err)
 | |
| 		return errUnmarshal
 | |
| 	}
 | |
| 
 | |
| 	log.Trace("%s %s: Task found: %#v", q.typ, q.name, data)
 | |
| 	q.WorkerPool.Push(data)
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // Shutdown processing from this queue
 | |
| func (q *ByteFIFOQueue) Shutdown() {
 | |
| 	log.Trace("%s: %s Shutting down", q.typ, q.name)
 | |
| 	select {
 | |
| 	case <-q.shutdownCtx.Done():
 | |
| 		return
 | |
| 	default:
 | |
| 	}
 | |
| 	q.shutdownCtxCancel()
 | |
| 	log.Debug("%s: %s Shutdown", q.typ, q.name)
 | |
| }
 | |
| 
 | |
| // IsShutdown returns a channel which is closed when this Queue is shutdown
 | |
| func (q *ByteFIFOQueue) IsShutdown() <-chan struct{} {
 | |
| 	return q.shutdownCtx.Done()
 | |
| }
 | |
| 
 | |
| // Terminate this queue and close the queue
 | |
| func (q *ByteFIFOQueue) Terminate() {
 | |
| 	log.Trace("%s: %s Terminating", q.typ, q.name)
 | |
| 	q.Shutdown()
 | |
| 	select {
 | |
| 	case <-q.terminateCtx.Done():
 | |
| 		return
 | |
| 	default:
 | |
| 	}
 | |
| 	if log.IsDebug() {
 | |
| 		log.Debug("%s: %s Closing with %d tasks left in queue", q.typ, q.name, q.byteFIFO.Len(q.terminateCtx))
 | |
| 	}
 | |
| 	q.terminateCtxCancel()
 | |
| 	if err := q.byteFIFO.Close(); err != nil {
 | |
| 		log.Error("Error whilst closing internal byte fifo in %s: %s: %v", q.typ, q.name, err)
 | |
| 	}
 | |
| 	q.baseCtxFinished()
 | |
| 	log.Debug("%s: %s Terminated", q.typ, q.name)
 | |
| }
 | |
| 
 | |
| // IsTerminated returns a channel which is closed when this Queue is terminated
 | |
| func (q *ByteFIFOQueue) IsTerminated() <-chan struct{} {
 | |
| 	return q.terminateCtx.Done()
 | |
| }
 | |
| 
 | |
| var _ UniqueQueue = &ByteFIFOUniqueQueue{}
 | |
| 
 | |
| // ByteFIFOUniqueQueue represents a UniqueQueue formed from a UniqueByteFifo
 | |
| type ByteFIFOUniqueQueue struct {
 | |
| 	ByteFIFOQueue
 | |
| }
 | |
| 
 | |
| // NewByteFIFOUniqueQueue creates a new ByteFIFOUniqueQueue
 | |
| func NewByteFIFOUniqueQueue(typ Type, byteFIFO UniqueByteFIFO, handle HandlerFunc, cfg, exemplar interface{}) (*ByteFIFOUniqueQueue, error) {
 | |
| 	configInterface, err := toConfig(ByteFIFOQueueConfiguration{}, cfg)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	config := configInterface.(ByteFIFOQueueConfiguration)
 | |
| 	terminateCtx, terminateCtxCancel := context.WithCancel(context.Background())
 | |
| 	shutdownCtx, shutdownCtxCancel := context.WithCancel(terminateCtx)
 | |
| 
 | |
| 	q := &ByteFIFOUniqueQueue{
 | |
| 		ByteFIFOQueue: ByteFIFOQueue{
 | |
| 			byteFIFO:           byteFIFO,
 | |
| 			typ:                typ,
 | |
| 			shutdownCtx:        shutdownCtx,
 | |
| 			shutdownCtxCancel:  shutdownCtxCancel,
 | |
| 			terminateCtx:       terminateCtx,
 | |
| 			terminateCtxCancel: terminateCtxCancel,
 | |
| 			exemplar:           exemplar,
 | |
| 			workers:            config.Workers,
 | |
| 			name:               config.Name,
 | |
| 		},
 | |
| 	}
 | |
| 	q.WorkerPool = NewWorkerPool(func(data ...Data) (failed []Data) {
 | |
| 		for _, unhandled := range handle(data...) {
 | |
| 			if fail := q.PushBack(unhandled); fail != nil {
 | |
| 				failed = append(failed, fail)
 | |
| 			}
 | |
| 		}
 | |
| 		return failed
 | |
| 	}, config.WorkerPoolConfiguration)
 | |
| 
 | |
| 	return q, nil
 | |
| }
 | |
| 
 | |
| // Has checks if the provided data is in the queue
 | |
| func (q *ByteFIFOUniqueQueue) Has(data Data) (bool, error) {
 | |
| 	if !assignableTo(data, q.exemplar) {
 | |
| 		return false, fmt.Errorf("unable to assign data: %v to same type as exemplar: %v in %s", data, q.exemplar, q.name)
 | |
| 	}
 | |
| 	bs, err := json.Marshal(data)
 | |
| 	if err != nil {
 | |
| 		return false, err
 | |
| 	}
 | |
| 	return q.byteFIFO.(UniqueByteFIFO).Has(q.terminateCtx, bs)
 | |
| }
 |