mirror of
https://github.com/safing/portmaster
synced 2025-04-20 10:59:10 +00:00
* Move portbase into monorepo * Add new simple module mgr * [WIP] Switch to new simple module mgr * Add StateMgr and more worker variants * [WIP] Switch more modules * [WIP] Switch more modules * [WIP] swtich more modules * [WIP] switch all SPN modules * [WIP] switch all service modules * [WIP] Convert all workers to the new module system * [WIP] add new task system to module manager * [WIP] Add second take for scheduling workers * [WIP] Add FIXME for bugs in new scheduler * [WIP] Add minor improvements to scheduler * [WIP] Add new worker scheduler * [WIP] Fix more bug related to new module system * [WIP] Fix start handing of the new module system * [WIP] Improve startup process * [WIP] Fix minor issues * [WIP] Fix missing subsystem in settings * [WIP] Initialize managers in constructor * [WIP] Move module event initialization to constrictors * [WIP] Fix setting for enabling and disabling the SPN module * [WIP] Move API registeration into module construction * [WIP] Update states mgr for all modules * [WIP] Add CmdLine operation support * Add state helper methods to module group and instance * Add notification and module status handling to status package * Fix starting issues * Remove pilot widget and update security lock to new status data * Remove debug logs * Improve http server shutdown * Add workaround for cleanly shutting down firewall+netquery * Improve logging * Add syncing states with notifications for new module system * Improve starting, stopping, shutdown; resolve FIXMEs/TODOs * [WIP] Fix most unit tests * Review new module system and fix minor issues * Push shutdown and restart events again via API * Set sleep mode via interface * Update example/template module * [WIP] Fix spn/cabin unit test * Remove deprecated UI elements * Make log output more similar for the logging transition phase * Switch spn hub and observer cmds to new module system * Fix log sources * Make worker mgr less error prone * Fix tests and minor issues * Fix observation hub * Improve shutdown and restart handling * Split up big connection.go source file * Move varint and dsd packages to structures repo * Improve expansion test * Fix linter warnings * Fix interception module on windows * Fix linter errors --------- Co-authored-by: Vladimir Stoilov <vladimir@safing.io>
107 lines
2.4 KiB
Go
107 lines
2.4 KiB
Go
package unit
|
|
|
|
import (
|
|
"fmt"
|
|
"math"
|
|
"math/rand"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/safing/portmaster/service/mgr"
|
|
)
|
|
|
|
func TestUnit(t *testing.T) { //nolint:paralleltest
|
|
// Ignore deprectation, as the given alternative is not safe for concurrent use.
|
|
// The global rand methods use a locked seed, which is not available from outside.
|
|
rand.Seed(time.Now().UnixNano()) //nolint
|
|
|
|
size := 1000000
|
|
workers := 100
|
|
|
|
m := mgr.New("unit-test")
|
|
// Create and start scheduler.
|
|
s := NewScheduler(&SchedulerConfig{})
|
|
s.StartDebugLog()
|
|
// ctx, cancel := context.WithCancel(context.Background())
|
|
m.Go("test", func(w *mgr.WorkerCtx) error {
|
|
err := s.SlotScheduler(w)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return nil
|
|
})
|
|
defer m.Cancel()
|
|
|
|
// Create 10 workers.
|
|
var wg sync.WaitGroup
|
|
wg.Add(workers)
|
|
sizePerWorker := size / workers
|
|
for range workers {
|
|
go func() {
|
|
for range sizePerWorker {
|
|
u := s.NewUnit()
|
|
|
|
// Make 1% high priority.
|
|
if rand.Int()%100 == 0 { //nolint:gosec // This is a test.
|
|
u.MakeHighPriority()
|
|
}
|
|
|
|
u.WaitForSlot()
|
|
time.Sleep(10 * time.Microsecond)
|
|
u.Finish()
|
|
}
|
|
wg.Done()
|
|
}()
|
|
}
|
|
|
|
// Wait for workers to finish.
|
|
wg.Wait()
|
|
|
|
// Wait for two slot durations for values to update.
|
|
time.Sleep(s.config.SlotDuration * 2)
|
|
|
|
// Print current state.
|
|
s.cycleStats()
|
|
fmt.Printf(`scheduler state:
|
|
currentUnitID = %d
|
|
slotPace = %d
|
|
clearanceUpTo = %d
|
|
finished = %d
|
|
maxPace = %d
|
|
maxLeveledPace = %d
|
|
avgPace = %d
|
|
avgUnitLife = %s
|
|
avgWorkSlot = %s
|
|
avgCatchUpSlot = %s
|
|
`,
|
|
s.currentUnitID.Load(),
|
|
s.slotPace.Load(),
|
|
s.clearanceUpTo.Load(),
|
|
s.finished.Load(),
|
|
s.GetMaxSlotPace(),
|
|
s.GetMaxLeveledSlotPace(),
|
|
s.GetAvgSlotPace(),
|
|
time.Duration(s.GetAvgUnitLife()),
|
|
time.Duration(s.GetAvgWorkSlotDuration()),
|
|
time.Duration(s.GetAvgCatchUpSlotDuration()),
|
|
)
|
|
|
|
// Check if everything seems good.
|
|
assert.Equal(t, size, int(s.currentUnitID.Load()), "currentUnitID must match size")
|
|
assert.GreaterOrEqual(
|
|
t,
|
|
int(s.clearanceUpTo.Load()),
|
|
size+int(float64(s.config.MinSlotPace)*s.config.SlotChangeRatePerStreak),
|
|
"clearanceUpTo must be at least size+minSlotPace",
|
|
)
|
|
|
|
// Shutdown
|
|
m.Cancel()
|
|
time.Sleep(s.config.SlotDuration * 10)
|
|
|
|
// Check if scheduler shut down correctly.
|
|
assert.Equal(t, math.MaxInt64-math.MaxInt32, int(s.clearanceUpTo.Load()), "clearance must be near MaxInt64")
|
|
}
|