Compare commits
No commits in common. "13ebb273359f09639e8d73fab8f006fe7dd082b5" and "4665c90c2cf0b832249b7ce793ad1ff06a2b1db6" have entirely different histories.
13ebb27335
...
4665c90c2c
10 changed files with 93 additions and 87 deletions
|
|
@ -1,24 +1,17 @@
|
||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
|
||||||
"os"
|
|
||||||
"os/signal"
|
|
||||||
"syscall"
|
|
||||||
"task-processor/config"
|
"task-processor/config"
|
||||||
"task-processor/internal/app"
|
"task-processor/internal/app"
|
||||||
"task-processor/internal/logging"
|
"task-processor/internal/logging"
|
||||||
)
|
)
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
ctx, cancel := signal.NotifyContext(context.Background(), os.Interrupt, syscall.SIGTERM)
|
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
c := config.NewConfig()
|
c := config.NewConfig()
|
||||||
|
|
||||||
logging.LogSetup(c.LogLevel)
|
logging.LogSetup(c.LogLevel)
|
||||||
|
|
||||||
appl := app.New(c)
|
appl := app.New(c)
|
||||||
|
|
||||||
appl.Run(ctx)
|
appl.Run()
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -3,9 +3,11 @@ package app
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
"google.golang.org/grpc"
|
|
||||||
"net"
|
"net"
|
||||||
|
"os"
|
||||||
|
"os/signal"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
"syscall"
|
||||||
"task-processor/config"
|
"task-processor/config"
|
||||||
"task-processor/internal/appState"
|
"task-processor/internal/appState"
|
||||||
"task-processor/internal/parsers"
|
"task-processor/internal/parsers"
|
||||||
|
|
@ -26,8 +28,6 @@ type App struct {
|
||||||
state *appState.State
|
state *appState.State
|
||||||
network *remote.Network
|
network *remote.Network
|
||||||
numCPUs int
|
numCPUs int
|
||||||
metricsSrv *router.Handler
|
|
||||||
taskApiSrv *grpc.Server
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(c *config.Config) *App {
|
func New(c *config.Config) *App {
|
||||||
|
|
@ -38,14 +38,6 @@ func New(c *config.Config) *App {
|
||||||
|
|
||||||
st := appState.NewState(numCPUs, c.CheckPeriod, c.TasksCfg.RetryCount, c.TasksCfg.RetryMinutes)
|
st := appState.NewState(numCPUs, c.CheckPeriod, c.TasksCfg.RetryCount, c.TasksCfg.RetryMinutes)
|
||||||
|
|
||||||
server := newServer(st)
|
|
||||||
|
|
||||||
//metrics
|
|
||||||
mSrv := router.NewHandler(router.Deps{
|
|
||||||
Addr: net.JoinHostPort(c.Metrics.Host, c.Metrics.Port),
|
|
||||||
GinMode: c.Metrics.GinMode,
|
|
||||||
})
|
|
||||||
|
|
||||||
return &App{
|
return &App{
|
||||||
config: c,
|
config: c,
|
||||||
checkPeriod: time.Duration(c.CheckPeriod),
|
checkPeriod: time.Duration(c.CheckPeriod),
|
||||||
|
|
@ -55,31 +47,33 @@ func New(c *config.Config) *App {
|
||||||
state: st,
|
state: st,
|
||||||
network: remote.NewHandler(),
|
network: remote.NewHandler(),
|
||||||
numCPUs: numCPUs,
|
numCPUs: numCPUs,
|
||||||
metricsSrv: mSrv,
|
|
||||||
taskApiSrv: server,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (app *App) Run(ctx context.Context) {
|
func (app *App) Run() {
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
log.Info("Application start")
|
log.Info("Application start")
|
||||||
|
|
||||||
addr := net.JoinHostPort(app.config.GrpcCfg.ServerHost, app.config.GrpcCfg.ServerPort)
|
|
||||||
|
|
||||||
log.WithFields(log.Fields{
|
log.WithFields(log.Fields{
|
||||||
"Service address": addr,
|
"Service address": app.config.GrpcCfg.ServerHost + ":" + app.config.GrpcCfg.ServerPort,
|
||||||
"Number of CPUs": app.numCPUs,
|
"Number of CPUs": app.numCPUs,
|
||||||
}).Debug("App settings")
|
}).Debug("App settings")
|
||||||
|
|
||||||
errChan := make(chan error, 16)
|
//metrics
|
||||||
|
mSrv := router.NewHandler(router.Deps{
|
||||||
|
Addr: net.JoinHostPort(app.config.Metrics.Host, app.config.Metrics.Port),
|
||||||
|
GinMode: app.config.Metrics.GinMode,
|
||||||
|
})
|
||||||
|
|
||||||
//main
|
//main
|
||||||
|
server := newServer(app)
|
||||||
apiClient := newApiClient(app.config.GrpcCfg.ApiClientHost + ":" + app.config.GrpcCfg.ApiClientPort)
|
apiClient := newApiClient(app.config.GrpcCfg.ApiClientHost + ":" + app.config.GrpcCfg.ApiClientPort)
|
||||||
|
|
||||||
period := time.NewTicker(app.checkPeriod * time.Hour)
|
period := time.NewTicker(app.checkPeriod * time.Hour)
|
||||||
defer period.Stop()
|
defer period.Stop()
|
||||||
|
|
||||||
sender := make(chan shared.TaskResult, app.numCPUs*10)
|
sender := make(chan shared.TaskResult, app.numCPUs*10)
|
||||||
defer close(sender)
|
|
||||||
|
|
||||||
// external scrapper
|
// external scrapper
|
||||||
surugayaScrapper := newSurugayaScrapperClient(app.config.GrpcCfg.SurugayaScrapperHost + ":" + app.config.GrpcCfg.SurugayaScrapperPort)
|
surugayaScrapper := newSurugayaScrapperClient(app.config.GrpcCfg.SurugayaScrapperHost + ":" + app.config.GrpcCfg.SurugayaScrapperPort)
|
||||||
|
|
@ -88,11 +82,11 @@ func (app *App) Run(ctx context.Context) {
|
||||||
handlers := make(map[string]parsers.TaskHandler)
|
handlers := make(map[string]parsers.TaskHandler)
|
||||||
|
|
||||||
if app.config.OriginEnabled.Surugaya {
|
if app.config.OriginEnabled.Surugaya {
|
||||||
handlers[shared.OriginSurugaya] = parsers.NewSurugayaParser(surugayaScrapper)
|
handlers[shared.OriginSurugaya] = parsers.NewSurugayaParser(ctx, surugayaScrapper)
|
||||||
}
|
}
|
||||||
|
|
||||||
if app.config.OriginEnabled.Mandarake {
|
if app.config.OriginEnabled.Mandarake {
|
||||||
handlers[shared.OriginMandarake] = mandarake.NewParser(mandarake.Deps{
|
handlers[shared.OriginMandarake] = mandarake.NewParser(mandarake.ParserDeps{
|
||||||
Enabled: app.config.OriginEnabled.Mandarake,
|
Enabled: app.config.OriginEnabled.Mandarake,
|
||||||
ExternalBrowser: app.config.ExternalBrowser,
|
ExternalBrowser: app.config.ExternalBrowser,
|
||||||
GoroutinesNumber: app.numCPUs,
|
GoroutinesNumber: app.numCPUs,
|
||||||
|
|
@ -114,7 +108,7 @@ func (app *App) Run(ctx context.Context) {
|
||||||
receivedTasks := app.network.RequestTasks(ctx, apiClient)
|
receivedTasks := app.network.RequestTasks(ctx, apiClient)
|
||||||
log.WithField("length", len(receivedTasks)).Debug("End receiving")
|
log.WithField("length", len(receivedTasks)).Debug("End receiving")
|
||||||
|
|
||||||
taskProcessor.StartWork(ctx, receivedTasks)
|
taskProcessor.StartWork(receivedTasks)
|
||||||
}
|
}
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
|
|
@ -149,38 +143,37 @@ func (app *App) Run(ctx context.Context) {
|
||||||
|
|
||||||
//start metrics server
|
//start metrics server
|
||||||
go func() {
|
go func() {
|
||||||
if err := app.metricsSrv.Run(); err != nil {
|
if err := mSrv.Run(); err != nil {
|
||||||
errChan <- err
|
log.WithError(err).Error("Metrics server run failed")
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
//gRPC Server for status response
|
//gRPC Server for status response
|
||||||
go func() {
|
go func() {
|
||||||
listener, err := net.Listen("tcp", addr)
|
listener, err := net.Listen("tcp", app.config.GrpcCfg.ServerHost+":"+app.config.GrpcCfg.ServerPort)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
log.Fatalf("failed to listen: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Infof("gRPC Server listening at %v", addr)
|
log.Infof("gRPC Server listening at %v", app.config.GrpcCfg.ServerHost+":"+app.config.GrpcCfg.ServerPort)
|
||||||
if err = app.taskApiSrv.Serve(listener); err != nil {
|
if err := server.Serve(listener); err != nil {
|
||||||
errChan <- err
|
log.Fatalf("failed to serve: %v", err)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
select {
|
go func() {
|
||||||
case <-ctx.Done():
|
sigint := make(chan os.Signal, 1)
|
||||||
app.shutdown(ctx)
|
signal.Notify(sigint, os.Interrupt, syscall.SIGTERM)
|
||||||
case err := <-errChan:
|
<-sigint
|
||||||
log.WithError(err).Fatal("Application run error")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (app *App) shutdown(ctx context.Context) {
|
|
||||||
log.Info("Shutting down...")
|
log.Info("Shutting down...")
|
||||||
|
|
||||||
app.taskApiSrv.GracefulStop()
|
period.Stop()
|
||||||
|
server.GracefulStop()
|
||||||
if err := app.metricsSrv.Shutdown(ctx); err != nil {
|
cancel()
|
||||||
|
if err := mSrv.Shutdown(ctx); err != nil {
|
||||||
log.WithError(err).Error("Failed to shutdown server")
|
log.WithError(err).Error("Failed to shutdown server")
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
<-ctx.Done()
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -12,10 +12,10 @@ type Server struct {
|
||||||
state *appState.State
|
state *appState.State
|
||||||
}
|
}
|
||||||
|
|
||||||
func newServer(state *appState.State) *grpc.Server {
|
func newServer(app *App) *grpc.Server {
|
||||||
s := grpc.NewServer()
|
s := grpc.NewServer()
|
||||||
srv := &Server{
|
srv := &Server{
|
||||||
state: state,
|
state: app.state,
|
||||||
}
|
}
|
||||||
pb.RegisterTaskProcessorServer(s, srv)
|
pb.RegisterTaskProcessorServer(s, srv)
|
||||||
return s
|
return s
|
||||||
|
|
|
||||||
|
|
@ -1,11 +1,10 @@
|
||||||
package parsers
|
package parsers
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
|
||||||
"task-processor/internal/appState"
|
"task-processor/internal/appState"
|
||||||
"task-processor/internal/shared"
|
"task-processor/internal/shared"
|
||||||
)
|
)
|
||||||
|
|
||||||
type TaskHandler interface {
|
type TaskHandler interface {
|
||||||
HandleTasks(ctx context.Context, tasks []shared.Task, sender chan shared.TaskResult, state *appState.State)
|
HandleTasks(tasks []shared.Task, sender chan shared.TaskResult, state *appState.State)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -4,20 +4,15 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/chromedp/chromedp"
|
"github.com/chromedp/chromedp"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
"runtime"
|
|
||||||
"sync"
|
"sync"
|
||||||
"task-processor/internal/appState"
|
"task-processor/internal/appState"
|
||||||
"task-processor/internal/shared"
|
"task-processor/internal/shared"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (s *Parser) HandleTasks(ctx context.Context, tasks []shared.Task, sender chan shared.TaskResult, state *appState.State) {
|
func (s *Parser) HandleTasks(tasks []shared.Task, sender chan shared.TaskResult, state *appState.State) {
|
||||||
log.Infof("%v %v handling tasks", logHeader, logWorker)
|
log.Debug(logHeader + logWorker + "handling tasks")
|
||||||
|
|
||||||
allocCtx, allocCancel := chromedp.NewRemoteAllocator(ctx, s.externalBrowser)
|
allocCtx, allocCancel := chromedp.NewRemoteAllocator(s.baseCtx, s.externalBrowser)
|
||||||
defer allocCancel()
|
|
||||||
|
|
||||||
sessionCtx, sessionCancel := chromedp.NewContext(allocCtx /* chromedp.WithLogf(log.Printf) */, chromedp.WithLogf(func(string, ...any) {}))
|
|
||||||
defer sessionCancel()
|
|
||||||
|
|
||||||
receiver := make(chan shared.Task, len(tasks))
|
receiver := make(chan shared.Task, len(tasks))
|
||||||
for _, task := range tasks {
|
for _, task := range tasks {
|
||||||
|
|
@ -25,27 +20,41 @@ func (s *Parser) HandleTasks(ctx context.Context, tasks []shared.Task, sender ch
|
||||||
}
|
}
|
||||||
close(receiver)
|
close(receiver)
|
||||||
|
|
||||||
log.Debugf("%v gorutines before wait group: %v", logHeader, runtime.NumGoroutine())
|
|
||||||
wg := sync.WaitGroup{}
|
wg := sync.WaitGroup{}
|
||||||
for i := 0; i < s.goroutinesNumber; i++ {
|
for i := 0; i < s.goroutinesNumber; i++ {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
s.worker(sessionCtx, receiver, sender)
|
s.worker(allocCtx, receiver, sender, state)
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
allocCancel()
|
||||||
log.Debugf("%v gorutines after wait group: %v", logHeader, runtime.NumGoroutine())
|
log.Debug(logHeader + logWorker + "finished handling tasks")
|
||||||
log.Infof(logHeader + logWorker + "finished handling tasks")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Parser) worker(ctx context.Context, receiver chan shared.Task, sender chan shared.TaskResult) {
|
func (s *Parser) worker(ctx context.Context, receiver chan shared.Task, sender chan shared.TaskResult, state *appState.State) {
|
||||||
for task := range receiver {
|
for task := range receiver {
|
||||||
log.WithField("task_uuid", task.MerchUuid).Infof("%v %v processing task", logHeader, logWorker)
|
log.WithField("task_uuid", task.MerchUuid).Debug(logHeader + logWorker + "processing task")
|
||||||
|
|
||||||
|
//pageCtx, pageCancel := chromedp.NewContext(ctx, chromedp.WithLogf(func(string, ...any) {}))
|
||||||
|
//
|
||||||
|
//price, err := s.getPrice(pageCtx, task)
|
||||||
|
//pageCancel()
|
||||||
|
|
||||||
|
//price, err := s.getMinimalPrice(task)
|
||||||
|
//if err != nil {
|
||||||
|
// log.WithField("task_uuid", task.MerchUuid).Warn(logHeader + logWorker + logTaskWarning + "failed to process, zero price")
|
||||||
|
// sender <- shared.TaskResult{
|
||||||
|
// MerchUuid: task.MerchUuid,
|
||||||
|
// Origin: task.Origin,
|
||||||
|
// Price: zeroPrice,
|
||||||
|
// }
|
||||||
|
// continue
|
||||||
|
//}
|
||||||
|
|
||||||
//price will be zeroPrice value in case of any error or if price not found
|
//price will be zeroPrice value in case of any error or if price not found
|
||||||
price := s.getMinimalPrice(ctx, task)
|
price := s.getMinimalPrice(task)
|
||||||
sender <- shared.TaskResult{
|
sender <- shared.TaskResult{
|
||||||
MerchUuid: task.MerchUuid,
|
MerchUuid: task.MerchUuid,
|
||||||
Origin: task.Origin,
|
Origin: task.Origin,
|
||||||
|
|
|
||||||
|
|
@ -1,36 +1,39 @@
|
||||||
package mandarake
|
package mandarake
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
zeroPrice int32 = 0
|
zeroPrice int32 = 0
|
||||||
taxMultiplier float64 = 1.1
|
taxMultiplier float64 = 1.1
|
||||||
logHeader = "Mandarake parser |"
|
logHeader = "Mandarake parser | "
|
||||||
logWorker = "worker:"
|
logWorker = "worker: "
|
||||||
logTaskWarning = "task warning:"
|
logTaskWarning = "task warning: "
|
||||||
logGetPrice = "get price:"
|
logGetPrice = "get price: "
|
||||||
)
|
)
|
||||||
|
|
||||||
type Parser struct {
|
type Parser struct {
|
||||||
|
baseCtx context.Context
|
||||||
externalBrowser string
|
externalBrowser string
|
||||||
goroutinesNumber int
|
goroutinesNumber int
|
||||||
}
|
}
|
||||||
|
|
||||||
type Deps struct {
|
type ParserDeps struct {
|
||||||
Enabled bool
|
Enabled bool
|
||||||
ExternalBrowser string
|
ExternalBrowser string
|
||||||
GoroutinesNumber int
|
GoroutinesNumber int
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewParser(deps Deps) *Parser {
|
func NewParser(deps ParserDeps) *Parser {
|
||||||
if !deps.Enabled {
|
if !deps.Enabled {
|
||||||
log.Info(logHeader + "disabled")
|
log.Info(logHeader + "disabled")
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return &Parser{
|
return &Parser{
|
||||||
|
baseCtx: context.Background(),
|
||||||
externalBrowser: deps.ExternalBrowser,
|
externalBrowser: deps.ExternalBrowser,
|
||||||
goroutinesNumber: deps.GoroutinesNumber,
|
goroutinesNumber: deps.GoroutinesNumber,
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -41,7 +41,14 @@ func (s *Parser) getPrice(ctx context.Context, task shared.Task) (int32, error)
|
||||||
return minimal, nil
|
return minimal, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Parser) getMinimalPrice(sessionCtx context.Context, task shared.Task) int32 {
|
func (s *Parser) getMinimalPrice(task shared.Task) int32 {
|
||||||
|
ctx := context.Background()
|
||||||
|
allocCtx, allocCancel := chromedp.NewRemoteAllocator(ctx, s.externalBrowser)
|
||||||
|
defer allocCancel()
|
||||||
|
|
||||||
|
sessionCtx, sessionCancel := chromedp.NewContext(allocCtx, chromedp.WithLogf(log.Printf))
|
||||||
|
defer sessionCancel()
|
||||||
|
|
||||||
var (
|
var (
|
||||||
singlePrice string
|
singlePrice string
|
||||||
rangedPrice string
|
rangedPrice string
|
||||||
|
|
|
||||||
|
|
@ -65,6 +65,7 @@ func TestParser_processPrices(t *testing.T) {
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
s := &Parser{
|
s := &Parser{
|
||||||
|
baseCtx: tt.fields.baseCtx,
|
||||||
externalBrowser: tt.fields.externalBrowser,
|
externalBrowser: tt.fields.externalBrowser,
|
||||||
goroutinesNumber: tt.fields.goroutinesNumber,
|
goroutinesNumber: tt.fields.goroutinesNumber,
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -11,19 +11,21 @@ import (
|
||||||
|
|
||||||
type SurugayaParser struct {
|
type SurugayaParser struct {
|
||||||
scrapper sc.SurugayaScrapperClient
|
scrapper sc.SurugayaScrapperClient
|
||||||
|
ctx context.Context
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewSurugayaParser(scrapper sc.SurugayaScrapperClient) *SurugayaParser {
|
func NewSurugayaParser(ctx context.Context, scrapper sc.SurugayaScrapperClient) *SurugayaParser {
|
||||||
log.Debug("Surugaya parser init")
|
log.Debug("Surugaya parser init")
|
||||||
return &SurugayaParser{
|
return &SurugayaParser{
|
||||||
scrapper: scrapper,
|
scrapper: scrapper,
|
||||||
|
ctx: ctx,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *SurugayaParser) HandleTasks(ctx context.Context, tasks []shared.Task, sender chan shared.TaskResult, state *appState.State) {
|
func (s *SurugayaParser) HandleTasks(tasks []shared.Task, sender chan shared.TaskResult, state *appState.State) {
|
||||||
log.WithField("count", len(tasks)).Debug("Handling Surugaya Tasks")
|
log.WithField("count", len(tasks)).Debug("Handling Surugaya Tasks")
|
||||||
|
|
||||||
stream, err := s.scrapper.ProcessTasks(ctx)
|
stream, err := s.scrapper.ProcessTasks(s.ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.WithField("err", err).Error("Error creating stream")
|
log.WithField("err", err).Error("Error creating stream")
|
||||||
return
|
return
|
||||||
|
|
|
||||||
|
|
@ -1,14 +1,13 @@
|
||||||
package processor
|
package processor
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
"sync"
|
"sync"
|
||||||
"task-processor/internal/appState"
|
"task-processor/internal/appState"
|
||||||
"task-processor/internal/shared"
|
"task-processor/internal/shared"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (p *Processor) StartWork(ctx context.Context, receivedTasks []shared.TaskResponse) {
|
func (p *Processor) StartWork(receivedTasks []shared.TaskResponse) {
|
||||||
log.Info("Starting work...")
|
log.Info("Starting work...")
|
||||||
p.state.ResetCounters()
|
p.state.ResetCounters()
|
||||||
|
|
||||||
|
|
@ -28,7 +27,7 @@ func (p *Processor) StartWork(ctx context.Context, receivedTasks []shared.TaskRe
|
||||||
go func(origin string, tasks []shared.Task) {
|
go func(origin string, tasks []shared.Task) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
log.Info("Running task handler for origin: ", origin)
|
log.Info("Running task handler for origin: ", origin)
|
||||||
p.handlers[origin].HandleTasks(ctx, tasks, p.out, p.state)
|
p.handlers[origin].HandleTasks(tasks, p.out, p.state)
|
||||||
}(origin, tasks)
|
}(origin, tasks)
|
||||||
}
|
}
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue