// Copyright 2019 The Gitea Authors. All rights reserved. // Use of this source code is governed by a MIT-style // license that can be found in the LICENSE file. package queue import ( "context" "fmt" "sync" "sync/atomic" "time" "code.gitea.io/gitea/modules/log" ) // PersistableChannelQueueType is the type for persistable queue const PersistableChannelQueueType Type = "persistable-channel" // PersistableChannelQueueConfiguration is the configuration for a PersistableChannelQueue type PersistableChannelQueueConfiguration struct { Name string DataDir string BatchLength int QueueLength int Timeout time.Duration MaxAttempts int Workers int MaxWorkers int BlockTimeout time.Duration BoostTimeout time.Duration BoostWorkers int } // PersistableChannelQueue wraps a channel queue and level queue together // The disk level queue will be used to store data at shutdown and terminate - and will be restored // on start up. type PersistableChannelQueue struct { channelQueue *ChannelQueue delayedStarter lock sync.Mutex closed chan struct{} } // NewPersistableChannelQueue creates a wrapped batched channel queue with persistable level queue backend when shutting down // This differs from a wrapped queue in that the persistent queue is only used to persist at shutdown/terminate func NewPersistableChannelQueue(handle HandlerFunc, cfg, exemplar interface{}) (Queue, error) { configInterface, err := toConfig(PersistableChannelQueueConfiguration{}, cfg) if err != nil { return nil, err } config := configInterface.(PersistableChannelQueueConfiguration) channelQueue, err := NewChannelQueue(handle, ChannelQueueConfiguration{ WorkerPoolConfiguration: WorkerPoolConfiguration{ QueueLength: config.QueueLength, BatchLength: config.BatchLength, BlockTimeout: config.BlockTimeout, BoostTimeout: config.BoostTimeout, BoostWorkers: config.BoostWorkers, MaxWorkers: config.MaxWorkers, }, Workers: config.Workers, Name: config.Name + "-channel", }, exemplar) if err != nil { return nil, err } // the level backend only needs temporary workers to catch up with the previously dropped work levelCfg := LevelQueueConfiguration{ ByteFIFOQueueConfiguration: ByteFIFOQueueConfiguration{ WorkerPoolConfiguration: WorkerPoolConfiguration{ QueueLength: config.QueueLength, BatchLength: config.BatchLength, BlockTimeout: 1 * time.Second, BoostTimeout: 5 * time.Minute, BoostWorkers: 5, MaxWorkers: 6, }, Workers: 1, Name: config.Name + "-level", }, DataDir: config.DataDir, } levelQueue, err := NewLevelQueue(handle, levelCfg, exemplar) if err == nil { queue := &PersistableChannelQueue{ channelQueue: channelQueue.(*ChannelQueue), delayedStarter: delayedStarter{ internal: levelQueue.(*LevelQueue), name: config.Name, }, closed: make(chan struct{}), } _ = GetManager().Add(queue, PersistableChannelQueueType, config, exemplar) return queue, nil } if IsErrInvalidConfiguration(err) { // Retrying ain't gonna make this any better... return nil, ErrInvalidConfiguration{cfg: cfg} } queue := &PersistableChannelQueue{ channelQueue: channelQueue.(*ChannelQueue), delayedStarter: delayedStarter{ cfg: levelCfg, underlying: LevelQueueType, timeout: config.Timeout, maxAttempts: config.MaxAttempts, name: config.Name, }, closed: make(chan struct{}), } _ = GetManager().Add(queue, PersistableChannelQueueType, config, exemplar) return queue, nil } // Name returns the name of this queue func (q *PersistableChannelQueue) Name() string { return q.delayedStarter.name } // Push will push the indexer data to queue func (q *PersistableChannelQueue) Push(data Data) error { select { case <-q.closed: return q.internal.Push(data) default: return q.channelQueue.Push(data) } } // Run starts to run the queue func (q *PersistableChannelQueue) Run(atShutdown, atTerminate func(context.Context, func())) { log.Debug("PersistableChannelQueue: %s Starting", q.delayedStarter.name) q.lock.Lock() if q.internal == nil { err := q.setInternal(atShutdown, q.channelQueue.handle, q.channelQueue.exemplar) q.lock.Unlock() if err != nil { log.Fatal("Unable to create internal queue for %s Error: %v", q.Name(), err) return } } else { q.lock.Unlock() } atShutdown(context.Background(), q.Shutdown) atTerminate(context.Background(), q.Terminate) // Just run the level queue - we shut it down later go q.internal.Run(func(_ context.Context, _ func()) {}, func(_ context.Context, _ func()) {}) go func() { _ = q.channelQueue.AddWorkers(q.channelQueue.workers, 0) }() log.Trace("PersistableChannelQueue: %s Waiting til closed", q.delayedStarter.name) <-q.closed log.Trace("PersistableChannelQueue: %s Cancelling pools", q.delayedStarter.name) q.channelQueue.cancel() q.internal.(*LevelQueue).cancel() log.Trace("PersistableChannelQueue: %s Waiting til done", q.delayedStarter.name) q.channelQueue.Wait() q.internal.(*LevelQueue).Wait() // Redirect all remaining data in the chan to the internal channel go func() { log.Trace("PersistableChannelQueue: %s Redirecting remaining data", q.delayedStarter.name) for data := range q.channelQueue.dataChan { _ = q.internal.Push(data) atomic.AddInt64(&q.channelQueue.numInQueue, -1) } log.Trace("PersistableChannelQueue: %s Done Redirecting remaining data", q.delayedStarter.name) }() log.Trace("PersistableChannelQueue: %s Done main loop", q.delayedStarter.name) } // Flush flushes the queue and blocks till the queue is empty func (q *PersistableChannelQueue) Flush(timeout time.Duration) error { var ctx context.Context var cancel context.CancelFunc if timeout > 0 { ctx, cancel = context.WithTimeout(context.Background(), timeout) } else { ctx, cancel = context.WithCancel(context.Background()) } defer cancel() return q.FlushWithContext(ctx) } // FlushWithContext flushes the queue and blocks till the queue is empty func (q *PersistableChannelQueue) FlushWithContext(ctx context.Context) error { errChan := make(chan error, 1) go func() { errChan <- q.channelQueue.FlushWithContext(ctx) }() go func() { q.lock.Lock() if q.internal == nil { q.lock.Unlock() errChan <- fmt.Errorf("not ready to flush internal queue %s yet", q.Name()) return } q.lock.Unlock() errChan <- q.internal.FlushWithContext(ctx) }() err1 := <-errChan err2 := <-errChan if err1 != nil { return err1 } return err2 } // IsEmpty checks if a queue is empty func (q *PersistableChannelQueue) IsEmpty() bool { if !q.channelQueue.IsEmpty() { return false } q.lock.Lock() defer q.lock.Unlock() if q.internal == nil { return false } return q.internal.IsEmpty() } // Shutdown processing this queue func (q *PersistableChannelQueue) Shutdown() { log.Trace("PersistableChannelQueue: %s Shutting down", q.delayedStarter.name) q.lock.Lock() defer q.lock.Unlock() select { case <-q.closed: default: if q.internal != nil { q.internal.(*LevelQueue).Shutdown() } close(q.closed) log.Debug("PersistableChannelQueue: %s Shutdown", q.delayedStarter.name) } } // Terminate this queue and close the queue func (q *PersistableChannelQueue) Terminate() { log.Trace("PersistableChannelQueue: %s Terminating", q.delayedStarter.name) q.Shutdown() q.lock.Lock() defer q.lock.Unlock() if q.internal != nil { q.internal.(*LevelQueue).Terminate() } log.Debug("PersistableChannelQueue: %s Terminated", q.delayedStarter.name) } func init() { queuesMap[PersistableChannelQueueType] = NewPersistableChannelQueue }