mirror of
https://github.com/d0zingcat/gocryptotrader.git
synced 2026-05-13 15:09:42 +00:00
* Bybit: Fix race in TestUpdateAccountInfo and TestWSHandleData * DriveBy rename TestWSHandleData * This doesn't address running with -race=2+ due to the singleton * Accounts: Add account.GetService() * exchange: Assertify TestSetupDefaults * Exchanges: Add account.Service override for testing * Exchanges: Remove duplicate IsWebsocketEnabled test from TestSetupDefaults * Dispatch: Replace nil checks with NilGuard * Engine: Remove deprecated printAccountHoldingsChangeSummary * Dispatcher: Add EnsureRunning method * Accounts: Move singleton accounts service to exchange Accounts * Move singleton accounts service to exchange Accounts This maintains the concept of a global store, whilst allowing exchanges to override it when needed, particularly for testing. APIServer: * Remove getAllActiveAccounts from apiserver Deprecated apiserver only thing using this, so remove it instead of updating it * Update comment for UpdateAccountBalances everywhere * Docs: Add punctuation to function comments * Bybit: Coverage for wsProcessWalletPushData Save
349 lines
7.8 KiB
Go
349 lines
7.8 KiB
Go
package dispatch
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/gofrs/uuid"
|
|
"github.com/thrasher-corp/gocryptotrader/common"
|
|
"github.com/thrasher-corp/gocryptotrader/log"
|
|
)
|
|
|
|
// Public errors.
|
|
var (
|
|
ErrNotRunning = errors.New("dispatcher not running")
|
|
ErrDispatcherAlreadyRunning = errors.New("dispatcher already running")
|
|
)
|
|
|
|
var (
|
|
errDispatchShutdown = errors.New("dispatcher did not shutdown properly, routines failed to close")
|
|
errDispatcherUUIDNotFoundInRouteList = errors.New("dispatcher uuid not found in route list")
|
|
errTypeAssertionFailure = errors.New("type assertion failure")
|
|
errChannelNotFoundInUUIDRef = errors.New("dispatcher channel not found in uuid reference slice")
|
|
errUUIDCollision = errors.New("dispatcher collision detected, uuid already exists")
|
|
errDispatcherJobsAtLimit = errors.New("dispatcher jobs at limit")
|
|
errChannelIsNil = errors.New("channel is nil")
|
|
errUUIDGeneratorFunctionIsNil = errors.New("UUID generator function is nil")
|
|
|
|
limitMessage = "%w [%d] current worker count [%d]. Spawn more workers via --dispatchworkers=x, or increase the jobs limit via --dispatchjobslimit=x"
|
|
)
|
|
|
|
// Name is an exported subsystem name.
|
|
const Name = "dispatch"
|
|
|
|
func init() {
|
|
dispatcher = NewDispatcher()
|
|
}
|
|
|
|
// NewDispatcher creates a new Dispatcher for relaying data.
|
|
func NewDispatcher() *Dispatcher {
|
|
return &Dispatcher{
|
|
routes: make(map[uuid.UUID][]chan any),
|
|
outbound: sync.Pool{
|
|
New: func() any { return make(chan any) },
|
|
},
|
|
}
|
|
}
|
|
|
|
// Start starts the dispatch system and spawns workers.
|
|
func Start(workers, jobsLimit int) error {
|
|
dispatcher.m.Lock()
|
|
defer dispatcher.m.Unlock()
|
|
return dispatcher.start(workers, jobsLimit)
|
|
}
|
|
|
|
// EnsureRunning starts the global dispatcher if it's not already running.
|
|
func EnsureRunning(workers, jobsLimit int) error {
|
|
dispatcher.m.Lock()
|
|
defer dispatcher.m.Unlock()
|
|
if dispatcher.running {
|
|
return nil
|
|
}
|
|
return dispatcher.start(workers, jobsLimit)
|
|
}
|
|
|
|
// Stop will halt the dispatch service.
|
|
func Stop() error {
|
|
log.Debugln(log.DispatchMgr, "Dispatch manager shutting down...")
|
|
return dispatcher.stop()
|
|
}
|
|
|
|
// IsRunning checks to see if the dispatch service is running.
|
|
func IsRunning() bool {
|
|
return dispatcher.isRunning()
|
|
}
|
|
|
|
// start sets defaults and config and spawns workers.
|
|
// Does not provide locking protection.
|
|
func (d *Dispatcher) start(workers, channelCapacity int) error {
|
|
if err := common.NilGuard(d); err != nil {
|
|
return err
|
|
}
|
|
|
|
if d.running {
|
|
return ErrDispatcherAlreadyRunning
|
|
}
|
|
|
|
d.running = true
|
|
|
|
if workers < 1 {
|
|
log.Warnf(log.DispatchMgr, "Dispatcher workers cannot be zero, using default value %d\n", DefaultMaxWorkers)
|
|
workers = DefaultMaxWorkers
|
|
}
|
|
if channelCapacity < 1 {
|
|
log.Warnf(log.DispatchMgr, "Dispatcher jobs limit cannot be zero, using default values %d\n", DefaultJobsLimit)
|
|
channelCapacity = DefaultJobsLimit
|
|
}
|
|
d.jobs = make(chan job, channelCapacity)
|
|
d.maxWorkers = workers
|
|
d.shutdown = make(chan struct{})
|
|
|
|
for range d.maxWorkers {
|
|
d.wg.Add(1)
|
|
go d.relayer()
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// stop stops the service and shuts down all worker routines.
|
|
func (d *Dispatcher) stop() error {
|
|
if err := common.NilGuard(d); err != nil {
|
|
return err
|
|
}
|
|
|
|
d.m.Lock()
|
|
defer d.m.Unlock()
|
|
|
|
if !d.running {
|
|
return ErrNotRunning
|
|
}
|
|
|
|
d.running = false
|
|
|
|
// Stop all jobs
|
|
close(d.jobs)
|
|
|
|
// Release finished workers
|
|
close(d.shutdown)
|
|
|
|
ch := make(chan struct{}, 1)
|
|
go func(ch chan<- struct{}) {
|
|
d.wg.Wait()
|
|
ch <- struct{}{}
|
|
}(ch)
|
|
|
|
select {
|
|
case <-ch:
|
|
case <-time.After(time.Second):
|
|
return errDispatchShutdown
|
|
}
|
|
|
|
// Wait for all relayers to have exited, including any blocking channel writes, before closing channels
|
|
d.routesMtx.Lock()
|
|
for key, pipes := range d.routes {
|
|
for i := range pipes {
|
|
close(pipes[i])
|
|
}
|
|
d.routes[key] = nil
|
|
}
|
|
d.routesMtx.Unlock()
|
|
|
|
log.Debugln(log.DispatchMgr, "Dispatch manager shutdown")
|
|
|
|
return nil
|
|
}
|
|
|
|
// isRunning returns if the dispatch system is running.
|
|
func (d *Dispatcher) isRunning() bool {
|
|
if d == nil {
|
|
return false
|
|
}
|
|
|
|
d.m.RLock()
|
|
defer d.m.RUnlock()
|
|
return d.running
|
|
}
|
|
|
|
// relayer routine relays communications across the defined routes.
|
|
func (d *Dispatcher) relayer() {
|
|
for {
|
|
select {
|
|
case j := <-d.jobs:
|
|
if j.ID.IsNil() {
|
|
// empty jobs from `channelCapacity` length are sent upon shutdown
|
|
// every real job created has an ID set
|
|
continue
|
|
}
|
|
d.routesMtx.Lock()
|
|
pipes, ok := d.routes[j.ID]
|
|
if !ok {
|
|
log.Warnf(log.DispatchMgr, "%v: %v\n", errDispatcherUUIDNotFoundInRouteList, j.ID)
|
|
d.routesMtx.Unlock()
|
|
continue
|
|
}
|
|
for i := range pipes {
|
|
d.wg.Add(1)
|
|
go func(p chan any) {
|
|
defer d.wg.Done()
|
|
select {
|
|
case p <- j.Data:
|
|
case <-d.shutdown: // Avoids race on blocking consumer when we go to stop
|
|
}
|
|
}(pipes[i])
|
|
}
|
|
d.routesMtx.Unlock()
|
|
case <-d.shutdown:
|
|
d.wg.Done()
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// publish relays data to the subscribed subsystems.
|
|
func (d *Dispatcher) publish(id uuid.UUID, data any) error {
|
|
if err := common.NilGuard(d, data); err != nil {
|
|
return err
|
|
}
|
|
|
|
if id.IsNil() {
|
|
return errIDNotSet
|
|
}
|
|
|
|
d.m.RLock()
|
|
defer d.m.RUnlock()
|
|
|
|
if !d.running {
|
|
return nil
|
|
}
|
|
|
|
select {
|
|
case d.jobs <- job{data, id}: // Push job into job channel.
|
|
return nil
|
|
default:
|
|
return fmt.Errorf(limitMessage, errDispatcherJobsAtLimit, len(d.jobs), d.maxWorkers)
|
|
}
|
|
}
|
|
|
|
// Subscribe subscribes a system and returns a communication chan, this does not ensure initial push.
|
|
func (d *Dispatcher) subscribe(id uuid.UUID) (chan any, error) {
|
|
if err := common.NilGuard(d); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if id.IsNil() {
|
|
return nil, errIDNotSet
|
|
}
|
|
|
|
d.m.RLock()
|
|
defer d.m.RUnlock()
|
|
|
|
if !d.running {
|
|
return nil, ErrNotRunning
|
|
}
|
|
|
|
d.routesMtx.Lock()
|
|
defer d.routesMtx.Unlock()
|
|
if _, ok := d.routes[id]; !ok {
|
|
return nil, errDispatcherUUIDNotFoundInRouteList
|
|
}
|
|
|
|
// Get an unused channel from the channel pool
|
|
ch, ok := d.outbound.Get().(chan any)
|
|
if !ok {
|
|
return nil, errTypeAssertionFailure
|
|
}
|
|
|
|
d.routes[id] = append(d.routes[id], ch)
|
|
atomic.AddInt32(&d.subscriberCount, 1)
|
|
return ch, nil
|
|
}
|
|
|
|
// Unsubscribe unsubs a routine from the dispatcher.
|
|
func (d *Dispatcher) unsubscribe(id uuid.UUID, usedChan chan any) error {
|
|
if err := common.NilGuard(d); err != nil {
|
|
return err
|
|
}
|
|
|
|
if id.IsNil() {
|
|
return errIDNotSet
|
|
}
|
|
|
|
if usedChan == nil {
|
|
return errChannelIsNil
|
|
}
|
|
|
|
d.m.RLock()
|
|
defer d.m.RUnlock()
|
|
|
|
if !d.running {
|
|
// reference will already be released in the stop function
|
|
return nil
|
|
}
|
|
|
|
d.routesMtx.Lock()
|
|
defer d.routesMtx.Unlock()
|
|
pipes, ok := d.routes[id]
|
|
if !ok {
|
|
return errDispatcherUUIDNotFoundInRouteList
|
|
}
|
|
|
|
for i := range pipes {
|
|
if pipes[i] != usedChan {
|
|
continue
|
|
}
|
|
// Delete individual reference
|
|
pipes[i] = pipes[len(pipes)-1]
|
|
pipes[len(pipes)-1] = nil
|
|
d.routes[id] = pipes[:len(pipes)-1]
|
|
atomic.AddInt32(&d.subscriberCount, -1)
|
|
|
|
// Drain and put the used chan back in pool; only if it is not closed.
|
|
select {
|
|
case _, ok = <-usedChan:
|
|
default:
|
|
}
|
|
|
|
if ok {
|
|
d.outbound.Put(usedChan)
|
|
}
|
|
return nil
|
|
}
|
|
return errChannelNotFoundInUUIDRef
|
|
}
|
|
|
|
// GetNewID returns a new ID.
|
|
func (d *Dispatcher) getNewID(genFn func() (uuid.UUID, error)) (uuid.UUID, error) {
|
|
if err := common.NilGuard(d); err != nil {
|
|
return uuid.Nil, err
|
|
}
|
|
|
|
if genFn == nil {
|
|
return uuid.Nil, errUUIDGeneratorFunctionIsNil
|
|
}
|
|
|
|
// Continue to allow the generation, input and return of UUIDs even if
|
|
// service is not currently enabled.
|
|
|
|
d.m.RLock()
|
|
defer d.m.RUnlock()
|
|
|
|
// Generate new uuid
|
|
newID, err := genFn()
|
|
if err != nil {
|
|
return uuid.Nil, err
|
|
}
|
|
|
|
d.routesMtx.Lock()
|
|
defer d.routesMtx.Unlock()
|
|
// Check to see if it already exists
|
|
if _, ok := d.routes[newID]; ok {
|
|
return uuid.Nil, errUUIDCollision
|
|
}
|
|
// Write the key into system
|
|
d.routes[newID] = nil
|
|
return newID, nil
|
|
}
|