mirror of
https://github.com/d0zingcat/gocryptotrader.git
synced 2026-05-13 15:09:42 +00:00
Engine: Dispatch service (#346)
* Added dispatch service * Added orderbook streaming capabilities * Assigned correct orderbook.base exchange name * Fixed Requested niterinos Add in cli orderbook QA tool to gctcli Add exchange orderbook streaming * Add ticker streaming support through dispatch package * Added in some more info on error returns for orderbook.go * fix linter issues * Fix some issues * Update * Fix requested * move dispatch out of exchanges folder to its own independant folder * Fix requested * change orderbook string to tickers * Limit orderbooks to 50 and made dispatch system more stateless in operation * lower cases for update/retrieve/sub exchange name * Adds in asset validation and lower case conversion on cli * Remove comment * Moved timer to a higher scope so its not constantly initialised just reset per instance and removed returning unused channel on error * Rm unused release function in dispatch.go Reset timer and bleed buffered timer chan if needed in dispatch.go Added in ticker.Stop() and timer.Stop() functions for worker routine return in dispatch.go Index aggregated bid and ask functions for orderbook.go Added in dummy slice for wsorderbook_test.go * Moved drain to above Reset so potential race would not occur in dispatch.go Fix various linter issues dispatch.go * Fix some issues * change to start/stop service, added in service state change via cli, updated logger * fix requested * Add worker amount init spawning * fix linter issues * Fix more linter issues * More fixes * Fix race issue on releasing pipe channel on a close after shutting down dispatcher system * Moved all types to dispatch_types.go && remove panic * Moved types into serperate file && improve test coverage * RM unnecessary select case for draining channel && fixed error string * Added orderbook_types file and improved code coverage * gofmt file * reinstated select cases on drain because I am silly * Remove error for drop worker * Added more test cases * not checking error issue fix * remove func causing race in test, this has required protection via an exported function * set Gemini websocket orderbook exchange name
This commit is contained in:
committed by
Adrian Gallagher
parent
4a0fcc7f0f
commit
db317a2447
357
dispatch/dispatch.go
Normal file
357
dispatch/dispatch.go
Normal file
@@ -0,0 +1,357 @@
|
||||
package dispatch
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"sync"
|
||||
"sync/atomic"
|
||||
"time"
|
||||
|
||||
"github.com/gofrs/uuid"
|
||||
log "github.com/thrasher-corp/gocryptotrader/logger"
|
||||
)
|
||||
|
||||
func init() {
|
||||
dispatcher = &Dispatcher{
|
||||
routes: make(map[uuid.UUID][]chan interface{}),
|
||||
jobs: make(chan *job, DefaultJobBuffer),
|
||||
outbound: sync.Pool{
|
||||
New: func() interface{} {
|
||||
// Create unbuffered channel for data pass
|
||||
return make(chan interface{})
|
||||
},
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Start starts the dispatch system by spawning workers and allocating memory
|
||||
func Start(workers int64) error {
|
||||
if dispatcher == nil {
|
||||
return errors.New(errNotInitialised)
|
||||
}
|
||||
|
||||
mtx.Lock()
|
||||
defer mtx.Unlock()
|
||||
return dispatcher.start(workers)
|
||||
}
|
||||
|
||||
// Stop attempts to stop the dispatch service, this will close all pipe channels
|
||||
// flush job list and drop all workers
|
||||
func Stop() error {
|
||||
if dispatcher == nil {
|
||||
return errors.New(errNotInitialised)
|
||||
}
|
||||
|
||||
log.Debugln(log.DispatchMgr, "Dispatch manager shutting down...")
|
||||
|
||||
mtx.Lock()
|
||||
defer mtx.Unlock()
|
||||
return dispatcher.stop()
|
||||
}
|
||||
|
||||
// IsRunning checks to see if the dispatch service is running
|
||||
func IsRunning() bool {
|
||||
if dispatcher == nil {
|
||||
return false
|
||||
}
|
||||
|
||||
return dispatcher.isRunning()
|
||||
}
|
||||
|
||||
// DropWorker drops a worker routine
|
||||
func DropWorker() error {
|
||||
if dispatcher == nil {
|
||||
return errors.New(errNotInitialised)
|
||||
}
|
||||
|
||||
dispatcher.dropWorker()
|
||||
return nil
|
||||
}
|
||||
|
||||
// SpawnWorker starts a new worker routine
|
||||
func SpawnWorker() error {
|
||||
if dispatcher == nil {
|
||||
return errors.New(errNotInitialised)
|
||||
}
|
||||
return dispatcher.spawnWorker()
|
||||
}
|
||||
|
||||
// start compares atomic running value, sets defaults, overides with
|
||||
// configuration, then spawns workers
|
||||
func (d *Dispatcher) start(workers int64) error {
|
||||
if atomic.LoadUint32(&d.running) == 1 {
|
||||
return errors.New(errAlreadyStarted)
|
||||
}
|
||||
|
||||
if workers < 1 {
|
||||
log.Warn(log.DispatchMgr,
|
||||
"Dispatcher: workers cannot be zero using default values")
|
||||
workers = DefaultMaxWorkers
|
||||
}
|
||||
|
||||
d.maxWorkers = workers
|
||||
d.shutdown = make(chan *sync.WaitGroup)
|
||||
|
||||
if atomic.LoadInt64(&d.count) != 0 {
|
||||
return errors.New("dispatcher leaked workers found")
|
||||
}
|
||||
|
||||
for i := int64(0); i < d.maxWorkers; i++ {
|
||||
err := d.spawnWorker()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
atomic.SwapUint32(&d.running, 1)
|
||||
return nil
|
||||
}
|
||||
|
||||
// stop stops the service and shuts down all worker routines
|
||||
func (d *Dispatcher) stop() error {
|
||||
if !atomic.CompareAndSwapUint32(&d.running, 1, 0) {
|
||||
return errors.New(errCannotShutdown)
|
||||
}
|
||||
close(d.shutdown)
|
||||
ch := make(chan struct{})
|
||||
timer := time.NewTimer(1 * time.Second)
|
||||
defer func() {
|
||||
if !timer.Stop() {
|
||||
select {
|
||||
case <-timer.C:
|
||||
default:
|
||||
}
|
||||
}
|
||||
}()
|
||||
go func(ch chan struct{}) { d.wg.Wait(); ch <- struct{}{} }(ch)
|
||||
select {
|
||||
case <-ch:
|
||||
// close all routes
|
||||
for key := range d.routes {
|
||||
for i := range d.routes[key] {
|
||||
close(d.routes[key][i])
|
||||
}
|
||||
|
||||
d.routes[key] = nil
|
||||
}
|
||||
|
||||
for len(d.jobs) != 0 { // drain jobs channel for old data
|
||||
<-d.jobs
|
||||
}
|
||||
|
||||
log.Debugln(log.DispatchMgr, "Dispatch manager shutdown.")
|
||||
|
||||
return nil
|
||||
case <-timer.C:
|
||||
return errors.New(errShutdownRoutines)
|
||||
}
|
||||
}
|
||||
|
||||
// isRunning returns if the dispatch system is running
|
||||
func (d *Dispatcher) isRunning() bool {
|
||||
return atomic.LoadUint32(&d.running) == 1
|
||||
}
|
||||
|
||||
// dropWorker deallocates a worker routine
|
||||
func (d *Dispatcher) dropWorker() {
|
||||
wg := sync.WaitGroup{}
|
||||
wg.Add(1)
|
||||
d.shutdown <- &wg
|
||||
wg.Wait()
|
||||
}
|
||||
|
||||
// spawnWorker allocates a new worker for job processing
|
||||
func (d *Dispatcher) spawnWorker() error {
|
||||
if atomic.LoadInt64(&d.count) >= d.maxWorkers {
|
||||
return errors.New("dispatcher cannot spawn more workers; ceiling reached")
|
||||
}
|
||||
var spawnWg sync.WaitGroup
|
||||
spawnWg.Add(1)
|
||||
go d.relayer(&spawnWg)
|
||||
spawnWg.Wait()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Relayer routine relays communications across the defined routes
|
||||
func (d *Dispatcher) relayer(i *sync.WaitGroup) {
|
||||
atomic.AddInt64(&d.count, 1)
|
||||
d.wg.Add(1)
|
||||
timeout := time.NewTimer(0)
|
||||
i.Done()
|
||||
for {
|
||||
select {
|
||||
case j := <-d.jobs:
|
||||
d.rMtx.RLock()
|
||||
if _, ok := d.routes[j.ID]; !ok {
|
||||
d.rMtx.RUnlock()
|
||||
continue
|
||||
}
|
||||
// Channel handshake timeout feature if a channel is blocked for any
|
||||
// period of time due to an issue with the receiving routine.
|
||||
// This will wait on channel then fall over to the next route when
|
||||
// the timer actuates and continue over the route list. Have to
|
||||
// iterate across full length of routes so every routine can get
|
||||
// their new info, cannot be buffered as we dont want to have an old
|
||||
// orderbook etc contained in a buffered channel when a routine
|
||||
// actually is ready for a receive.
|
||||
// TODO: Need to consider optimal timer length
|
||||
for i := range d.routes[j.ID] {
|
||||
if !timeout.Stop() { // Stop timer before reset
|
||||
// Drain channel if timer has already actuated
|
||||
select {
|
||||
case <-timeout.C:
|
||||
default:
|
||||
}
|
||||
}
|
||||
|
||||
timeout.Reset(DefaultHandshakeTimeout)
|
||||
select {
|
||||
case d.routes[j.ID][i] <- j.Data:
|
||||
case <-timeout.C:
|
||||
}
|
||||
}
|
||||
d.rMtx.RUnlock()
|
||||
|
||||
case v := <-d.shutdown:
|
||||
if !timeout.Stop() {
|
||||
select {
|
||||
case <-timeout.C:
|
||||
default:
|
||||
}
|
||||
}
|
||||
atomic.AddInt64(&d.count, -1)
|
||||
if v != nil {
|
||||
v.Done()
|
||||
}
|
||||
d.wg.Done()
|
||||
return
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// publish relays data to the subscribed subsystems
|
||||
func (d *Dispatcher) publish(id uuid.UUID, data interface{}) error {
|
||||
if data == nil {
|
||||
return errors.New("dispatcher data cannot be nil")
|
||||
}
|
||||
|
||||
if id == (uuid.UUID{}) {
|
||||
return errors.New("dispatcher uuid not set")
|
||||
}
|
||||
|
||||
if atomic.LoadUint32(&d.running) == 0 {
|
||||
return nil
|
||||
}
|
||||
|
||||
// Create a new job to publish
|
||||
newJob := &job{
|
||||
Data: data,
|
||||
ID: id,
|
||||
}
|
||||
|
||||
// Push job on stack here
|
||||
select {
|
||||
case d.jobs <- newJob:
|
||||
default:
|
||||
return fmt.Errorf("dispatcher buffer at max capacity [%d] current worker count [%d], spawn more workers via --dispatchworkers=x",
|
||||
len(d.jobs),
|
||||
atomic.LoadInt64(&d.count))
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// Subscribe subscribes a system and returns a communication chan, this does not
|
||||
// ensure initial push. If your routine is out of sync with heartbeat and the
|
||||
// system does not get a change, its up to you to in turn get initial state.
|
||||
func (d *Dispatcher) subscribe(id uuid.UUID) (chan interface{}, error) {
|
||||
if atomic.LoadUint32(&d.running) == 0 {
|
||||
return nil, errors.New(errNotInitialised)
|
||||
}
|
||||
|
||||
// Read lock to read route list
|
||||
d.rMtx.RLock()
|
||||
_, ok := d.routes[id]
|
||||
d.rMtx.RUnlock()
|
||||
if !ok {
|
||||
return nil, errors.New("dispatcher uuid not found in route list")
|
||||
}
|
||||
|
||||
// Get an unused channel from the channel pool
|
||||
unusedChan := d.outbound.Get().(chan interface{})
|
||||
|
||||
// Lock for writing to the route list
|
||||
d.rMtx.Lock()
|
||||
d.routes[id] = append(d.routes[id], unusedChan)
|
||||
d.rMtx.Unlock()
|
||||
|
||||
return unusedChan, nil
|
||||
}
|
||||
|
||||
// Unsubscribe unsubs a routine from the dispatcher
|
||||
func (d *Dispatcher) unsubscribe(id uuid.UUID, usedChan chan interface{}) error {
|
||||
if atomic.LoadUint32(&d.running) == 0 {
|
||||
// reference will already be released in the stop function
|
||||
return nil
|
||||
}
|
||||
|
||||
// Read lock to read route list
|
||||
d.rMtx.RLock()
|
||||
_, ok := d.routes[id]
|
||||
d.rMtx.RUnlock()
|
||||
if !ok {
|
||||
return errors.New("dispatcher uuid does not reference any channels")
|
||||
}
|
||||
|
||||
// Lock for write to delete references
|
||||
d.rMtx.Lock()
|
||||
for i := range d.routes[id] {
|
||||
if d.routes[id][i] != usedChan {
|
||||
continue
|
||||
}
|
||||
// Delete individual reference
|
||||
d.routes[id][i] = d.routes[id][len(d.routes[id])-1]
|
||||
d.routes[id][len(d.routes[id])-1] = nil
|
||||
d.routes[id] = d.routes[id][:len(d.routes[id])-1]
|
||||
|
||||
d.rMtx.Unlock()
|
||||
|
||||
// Drain and put the used chan back in pool; only if it is not closed.
|
||||
select {
|
||||
case _, ok := <-usedChan:
|
||||
if !ok {
|
||||
return nil
|
||||
}
|
||||
default:
|
||||
}
|
||||
|
||||
d.outbound.Put(usedChan)
|
||||
return nil
|
||||
}
|
||||
d.rMtx.Unlock()
|
||||
return errors.New("dispatcher channel not found in uuid reference slice")
|
||||
}
|
||||
|
||||
// GetNewID returns a new ID
|
||||
func (d *Dispatcher) getNewID() (uuid.UUID, error) {
|
||||
// Generate new uuid
|
||||
newID, err := uuid.NewV4()
|
||||
if err != nil {
|
||||
return uuid.UUID{}, err
|
||||
}
|
||||
|
||||
// Check to see if it already exists
|
||||
d.rMtx.RLock()
|
||||
_, ok := d.routes[newID]
|
||||
d.rMtx.RUnlock()
|
||||
if ok {
|
||||
return newID, errors.New("dispatcher collision detected, uuid already exists")
|
||||
}
|
||||
|
||||
// Write the key into system
|
||||
d.rMtx.Lock()
|
||||
d.routes[newID] = nil
|
||||
d.rMtx.Unlock()
|
||||
|
||||
return newID, nil
|
||||
}
|
||||
307
dispatch/dispatch_test.go
Normal file
307
dispatch/dispatch_test.go
Normal file
@@ -0,0 +1,307 @@
|
||||
package dispatch
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"os"
|
||||
"sync"
|
||||
"testing"
|
||||
|
||||
"github.com/gofrs/uuid"
|
||||
)
|
||||
|
||||
var mux *Mux
|
||||
|
||||
func TestMain(m *testing.M) {
|
||||
err := Start(DefaultMaxWorkers)
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
os.Exit(1)
|
||||
}
|
||||
|
||||
cpyDispatch = dispatcher
|
||||
mux = GetNewMux()
|
||||
cpyMux = mux
|
||||
os.Exit(m.Run())
|
||||
}
|
||||
|
||||
var cpyDispatch *Dispatcher
|
||||
var cpyMux *Mux
|
||||
|
||||
func TestDispatcher(t *testing.T) {
|
||||
dispatcher = nil
|
||||
err := Stop()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = Start(10)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
if IsRunning() {
|
||||
t.Error("should be false")
|
||||
}
|
||||
|
||||
err = DropWorker()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = SpawnWorker()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
dispatcher = cpyDispatch
|
||||
|
||||
if !IsRunning() {
|
||||
t.Error("should be true")
|
||||
}
|
||||
|
||||
err = Start(10)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = DropWorker()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = DropWorker()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = SpawnWorker()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = SpawnWorker()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = SpawnWorker()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = Stop()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = Stop()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = Start(0)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
payload := "something"
|
||||
|
||||
err = dispatcher.publish(uuid.UUID{}, &payload)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = dispatcher.publish(uuid.UUID{}, nil)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
id, errrrrrrrr := dispatcher.getNewID()
|
||||
if errrrrrrrr != nil {
|
||||
t.Error(errrrrrrrr)
|
||||
}
|
||||
|
||||
err = dispatcher.publish(id, &payload)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = dispatcher.stop()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = dispatcher.publish(id, &payload)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
_, err = dispatcher.subscribe(id)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = dispatcher.start(10)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
someID, err := uuid.NewV4()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
_, err = dispatcher.subscribe(someID)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
randomChan := make(chan interface{})
|
||||
err = dispatcher.unsubscribe(someID, randomChan)
|
||||
if err == nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = dispatcher.unsubscribe(id, randomChan)
|
||||
if err == nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
close(randomChan)
|
||||
err = dispatcher.unsubscribe(id, randomChan)
|
||||
if err == nil {
|
||||
t.Error(err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestMux(t *testing.T) {
|
||||
mux = nil
|
||||
_, err := mux.Subscribe(uuid.UUID{})
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = mux.Unsubscribe(uuid.UUID{}, nil)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
err = mux.Publish(nil, nil)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
_, err = mux.GetID()
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
mux = cpyMux
|
||||
|
||||
err = mux.Publish(nil, nil)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
payload := "string"
|
||||
id, err := uuid.NewV4()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
err = mux.Publish([]uuid.UUID{id}, &payload)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
_, err = mux.Subscribe(uuid.UUID{})
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
|
||||
_, err = mux.Subscribe(id)
|
||||
if err == nil {
|
||||
t.Error("error cannot be nil")
|
||||
}
|
||||
}
|
||||
|
||||
func TestSubscribe(t *testing.T) {
|
||||
itemID, err := mux.GetID()
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
var pipes []Pipe
|
||||
for i := 0; i < 1000; i++ {
|
||||
newPipe, err := mux.Subscribe(itemID)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
pipes = append(pipes, newPipe)
|
||||
}
|
||||
|
||||
for i := range pipes {
|
||||
err := pipes[i].Release()
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestPublish(t *testing.T) {
|
||||
itemID, err := mux.GetID()
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
pipe, err := mux.Subscribe(itemID)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(1)
|
||||
go func(wg *sync.WaitGroup) {
|
||||
wg.Done()
|
||||
for {
|
||||
_, ok := <-pipe.C
|
||||
if !ok {
|
||||
pErr := pipe.Release()
|
||||
if pErr != nil {
|
||||
t.Error(pErr)
|
||||
}
|
||||
wg.Done()
|
||||
return
|
||||
}
|
||||
}
|
||||
}(&wg)
|
||||
wg.Wait()
|
||||
wg.Add(1)
|
||||
mainPayload := "PAYLOAD"
|
||||
for i := 0; i < 100; i++ {
|
||||
errMux := mux.Publish([]uuid.UUID{itemID}, &mainPayload)
|
||||
if errMux != nil {
|
||||
t.Error(errMux)
|
||||
}
|
||||
}
|
||||
|
||||
// Shut down dispatch system
|
||||
err = Stop()
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
wg.Wait()
|
||||
}
|
||||
|
||||
func BenchmarkSubscribe(b *testing.B) {
|
||||
newID, err := mux.GetID()
|
||||
if err != nil {
|
||||
b.Error(err)
|
||||
}
|
||||
|
||||
for n := 0; n < b.N; n++ {
|
||||
_, err := mux.Subscribe(newID)
|
||||
if err != nil {
|
||||
b.Error(err)
|
||||
}
|
||||
}
|
||||
}
|
||||
88
dispatch/dispatch_types.go
Normal file
88
dispatch/dispatch_types.go
Normal file
@@ -0,0 +1,88 @@
|
||||
package dispatch
|
||||
|
||||
import (
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/gofrs/uuid"
|
||||
)
|
||||
|
||||
const (
|
||||
// DefaultJobBuffer defines a maxiumum amount of jobs allowed in channel
|
||||
DefaultJobBuffer = 100
|
||||
|
||||
// DefaultMaxWorkers is the package default worker ceiling amount
|
||||
DefaultMaxWorkers = 10
|
||||
|
||||
// DefaultHandshakeTimeout defines a workers max length of time to wait on a
|
||||
// an unbuffered channel for a receiver before moving on to next route
|
||||
DefaultHandshakeTimeout = 200 * time.Nanosecond
|
||||
|
||||
errNotInitialised = "dispatcher not initialised"
|
||||
errAlreadyStarted = "dispatcher already started"
|
||||
errCannotShutdown = "dispatcher cannot shutdown, already stopped"
|
||||
errShutdownRoutines = "dispatcher did not shutdown properly, routines failed to close"
|
||||
)
|
||||
|
||||
// dispatcher is our main in memory instance with a stop/start mtx below
|
||||
var dispatcher *Dispatcher
|
||||
var mtx sync.Mutex
|
||||
|
||||
// Dispatcher defines an internal subsystem communication/change state publisher
|
||||
type Dispatcher struct {
|
||||
// routes refers to a subystem uuid ticket map with associated publish
|
||||
// channels, a relayer will be given a unique id through its job channel,
|
||||
// then publish the data across the full registered channels for that uuid.
|
||||
// See relayer() method below.
|
||||
routes map[uuid.UUID][]chan interface{}
|
||||
|
||||
// rMtx protects the routes variable ensuring acceptable read/write access
|
||||
rMtx sync.RWMutex
|
||||
|
||||
// Persistent buffered job queue for relayers
|
||||
jobs chan *job
|
||||
|
||||
// Dynamic channel pool; returns an unbuffered channel for routes map
|
||||
outbound sync.Pool
|
||||
|
||||
// MaxWorkers defines max worker ceiling
|
||||
maxWorkers int64
|
||||
|
||||
// Atomic values -----------------------
|
||||
// Worker counter
|
||||
count int64
|
||||
// Dispatch status
|
||||
running uint32
|
||||
|
||||
// Unbufferd shutdown chan, sync wg for ensuring concurrency when only
|
||||
// dropping a single relayer routine
|
||||
shutdown chan *sync.WaitGroup
|
||||
|
||||
// Relayer shutdown tracking
|
||||
wg sync.WaitGroup
|
||||
}
|
||||
|
||||
// job defines a relaying job associated with a ticket which allows routing to
|
||||
// routines that require specific data
|
||||
type job struct {
|
||||
Data interface{}
|
||||
ID uuid.UUID
|
||||
}
|
||||
|
||||
// Mux defines a new multiplexor for the dispatch system, these a generated
|
||||
// per subsystem
|
||||
type Mux struct {
|
||||
// Reference to the main running dispatch service
|
||||
d *Dispatcher
|
||||
sync.RWMutex
|
||||
}
|
||||
|
||||
// Pipe defines an outbound object to the desired routine
|
||||
type Pipe struct {
|
||||
// Channel to get all our lovely informations
|
||||
C chan interface{}
|
||||
// ID to tracked system
|
||||
id uuid.UUID
|
||||
// Reference to multiplexor
|
||||
m *Mux
|
||||
}
|
||||
77
dispatch/mux.go
Normal file
77
dispatch/mux.go
Normal file
@@ -0,0 +1,77 @@
|
||||
package dispatch
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"reflect"
|
||||
|
||||
"github.com/gofrs/uuid"
|
||||
)
|
||||
|
||||
// GetNewMux returns a new multiplexor to track subsystem updates
|
||||
func GetNewMux() *Mux {
|
||||
return &Mux{d: dispatcher}
|
||||
}
|
||||
|
||||
// Subscribe takes in a package defined signature element pointing to an ID set
|
||||
// and returns the associated pipe
|
||||
func (m *Mux) Subscribe(id uuid.UUID) (Pipe, error) {
|
||||
if m == nil {
|
||||
return Pipe{}, errors.New("mux is nil")
|
||||
}
|
||||
|
||||
if id == (uuid.UUID{}) {
|
||||
return Pipe{}, errors.New("id not set")
|
||||
}
|
||||
|
||||
ch, err := m.d.subscribe(id)
|
||||
if err != nil {
|
||||
return Pipe{}, err
|
||||
}
|
||||
|
||||
return Pipe{C: ch, id: id, m: m}, nil
|
||||
}
|
||||
|
||||
// Unsubscribe returns channel to the pool for the full signature set
|
||||
func (m *Mux) Unsubscribe(id uuid.UUID, ch chan interface{}) error {
|
||||
if m == nil {
|
||||
return errors.New("mux is nil")
|
||||
}
|
||||
return m.d.unsubscribe(id, ch)
|
||||
}
|
||||
|
||||
// Publish takes in a persistent memory address and dispatches changes to
|
||||
// required pipes. Data should be of *type.
|
||||
func (m *Mux) Publish(ids []uuid.UUID, data interface{}) error {
|
||||
if m == nil {
|
||||
return errors.New("mux is nil")
|
||||
}
|
||||
|
||||
if data == nil {
|
||||
return errors.New("data payload is nil")
|
||||
}
|
||||
|
||||
cpy := reflect.ValueOf(data).Elem().Interface()
|
||||
|
||||
for i := range ids {
|
||||
// Create copy to not interfere with stored value
|
||||
err := m.d.publish(ids[i], &cpy)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// GetID gets a lovely new ID
|
||||
func (m *Mux) GetID() (uuid.UUID, error) {
|
||||
if m == nil {
|
||||
return uuid.UUID{}, errors.New("mux is nil")
|
||||
}
|
||||
|
||||
return m.d.getNewID()
|
||||
}
|
||||
|
||||
// Release returns the channel to the communications pool to be reused
|
||||
func (p *Pipe) Release() error {
|
||||
return p.m.Unsubscribe(p.id, p.C)
|
||||
}
|
||||
Reference in New Issue
Block a user