fix: limit the different worker pools to available CPU cores. Should resolve #3189
This commit is contained in:
parent
b1381170c1
commit
d0376cdc75
@ -1,14 +1,14 @@
|
||||
package inbox
|
||||
|
||||
import (
|
||||
"runtime"
|
||||
|
||||
"github.com/owncast/owncast/activitypub/apmodels"
|
||||
log "github.com/sirupsen/logrus"
|
||||
)
|
||||
|
||||
const (
|
||||
// InboxWorkerPoolSize defines the number of concurrent ActivityPub handlers.
|
||||
InboxWorkerPoolSize = 10
|
||||
)
|
||||
// workerPoolSize defines the number of concurrent ActivityPub handlers.
|
||||
var workerPoolSize = runtime.GOMAXPROCS(0)
|
||||
|
||||
// Job struct bundling the ActivityPub and the payload in one struct.
|
||||
type Job struct {
|
||||
@ -22,7 +22,7 @@ func InitInboxWorkerPool() {
|
||||
queue = make(chan Job)
|
||||
|
||||
// start workers
|
||||
for i := 1; i <= InboxWorkerPoolSize; i++ {
|
||||
for i := 1; i <= workerPoolSize; i++ {
|
||||
go worker(i, queue)
|
||||
}
|
||||
}
|
||||
|
@ -2,14 +2,13 @@ package workerpool
|
||||
|
||||
import (
|
||||
"net/http"
|
||||
"runtime"
|
||||
|
||||
log "github.com/sirupsen/logrus"
|
||||
)
|
||||
|
||||
const (
|
||||
// ActivityPubWorkerPoolSize defines the number of concurrent HTTP ActivityPub requests.
|
||||
ActivityPubWorkerPoolSize = 10
|
||||
)
|
||||
// workerPoolSize defines the number of concurrent HTTP ActivityPub requests.
|
||||
var workerPoolSize = runtime.GOMAXPROCS(0)
|
||||
|
||||
// Job struct bundling the ActivityPub and the payload in one struct.
|
||||
type Job struct {
|
||||
@ -23,7 +22,7 @@ func InitOutboundWorkerPool() {
|
||||
queue = make(chan Job)
|
||||
|
||||
// start workers
|
||||
for i := 1; i <= ActivityPubWorkerPoolSize; i++ {
|
||||
for i := 1; i <= workerPoolSize; i++ {
|
||||
go worker(i, queue)
|
||||
}
|
||||
}
|
||||
|
@ -51,7 +51,7 @@ func TestMain(m *testing.M) {
|
||||
// this test ensures that `SendToWebhooks` without a `WaitGroup` doesn't panic.
|
||||
func TestPublicSend(t *testing.T) {
|
||||
// Send enough events to be sure at least one worker delivers a second event.
|
||||
const eventsCount = webhookWorkerPoolSize + 1
|
||||
eventsCount := webhookWorkerPoolSize + 1
|
||||
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(eventsCount)
|
||||
@ -267,7 +267,7 @@ func TestParallel(t *testing.T) {
|
||||
myId := atomic.AddUint32(&calls, 1)
|
||||
|
||||
// We made it to the pool size + 1 event, so we're done with the test.
|
||||
if myId == webhookWorkerPoolSize+1 {
|
||||
if myId == uint32(webhookWorkerPoolSize)+1 {
|
||||
close(finished)
|
||||
return
|
||||
}
|
||||
|
@ -4,6 +4,7 @@ import (
|
||||
"bytes"
|
||||
"encoding/json"
|
||||
"net/http"
|
||||
"runtime"
|
||||
"sync"
|
||||
|
||||
log "github.com/sirupsen/logrus"
|
||||
@ -12,10 +13,8 @@ import (
|
||||
"github.com/owncast/owncast/models"
|
||||
)
|
||||
|
||||
const (
|
||||
// webhookWorkerPoolSize defines the number of concurrent HTTP webhook requests.
|
||||
webhookWorkerPoolSize = 10
|
||||
)
|
||||
// webhookWorkerPoolSize defines the number of concurrent HTTP webhook requests.
|
||||
var webhookWorkerPoolSize = runtime.GOMAXPROCS(0)
|
||||
|
||||
// Job struct bundling the webhook and the payload in one struct.
|
||||
type Job struct {
|
||||
|
Loading…
Reference in New Issue
Block a user