expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: true,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: true,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: expect.any(Number),
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
max * maxMultiplier
)
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThanOrEqual(
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThanOrEqual(0)
+ expect(workerNode.workerUsage.waitTime.aggregate).toBeGreaterThanOrEqual(
0
)
}
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: expect.any(Number),
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
max * maxMultiplier
)
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.waitTime.aggregate).toBeGreaterThan(0)
}
// We need to clean up the resources after our test
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: true
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: true
})
// We need to clean up the resources after our test
await pool.destroy()
})
+ it('Verify LEAST_ELU strategy can be run in a fixed pool', async () => {
+ const pool = new FixedThreadPool(
+ max,
+ './tests/worker-files/thread/testWorker.js',
+ { workerChoiceStrategy: WorkerChoiceStrategies.LEAST_ELU }
+ )
+ // TODO: Create a better test to cover `LeastEluWorkerChoiceStrategy#choose`
+ const promises = new Set()
+ const maxMultiplier = 2
+ for (let i = 0; i < max * maxMultiplier; i++) {
+ promises.add(pool.execute())
+ }
+ await Promise.all(promises)
+ for (const workerNode of pool.workerNodes) {
+ const expectedWorkerUsage = {
+ tasks: {
+ executed: expect.any(Number),
+ executing: 0,
+ queued: 0,
+ failed: 0
+ },
+ runTime: {
+ aggregate: 0,
+ average: 0,
+ median: 0,
+ history: expect.any(CircularArray)
+ },
+ waitTime: {
+ aggregate: 0,
+ average: 0,
+ median: 0,
+ history: expect.any(CircularArray)
+ }
+ }
+ if (workerNode.workerUsage.elu === undefined) {
+ expect(workerNode.workerUsage).toStrictEqual({
+ ...expectedWorkerUsage,
+ elu: undefined
+ })
+ } else {
+ expect(workerNode.workerUsage).toStrictEqual({
+ ...expectedWorkerUsage,
+ elu: {
+ active: expect.any(Number),
+ idle: 0,
+ utilization: 1
+ }
+ })
+ }
+ expect(workerNode.workerUsage.tasks.executed).toBeGreaterThanOrEqual(0)
+ expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
+ max * maxMultiplier
+ )
+ }
+ // We need to clean up the resources after our test
+ await pool.destroy()
+ })
+
+ it('Verify LEAST_ELU strategy can be run in a dynamic pool', async () => {
+ const pool = new DynamicThreadPool(
+ min,
+ max,
+ './tests/worker-files/thread/testWorker.js',
+ { workerChoiceStrategy: WorkerChoiceStrategies.LEAST_ELU }
+ )
+ // TODO: Create a better test to cover `LeastEluWorkerChoiceStrategy#choose`
+ const promises = new Set()
+ const maxMultiplier = 2
+ for (let i = 0; i < max * maxMultiplier; i++) {
+ promises.add(pool.execute())
+ }
+ await Promise.all(promises)
+ for (const workerNode of pool.workerNodes) {
+ const expectedWorkerUsage = {
+ tasks: {
+ executed: expect.any(Number),
+ executing: 0,
+ queued: 0,
+ failed: 0
+ },
+ runTime: {
+ aggregate: 0,
+ average: 0,
+ median: 0,
+ history: expect.any(CircularArray)
+ },
+ waitTime: {
+ aggregate: 0,
+ average: 0,
+ median: 0,
+ history: expect.any(CircularArray)
+ }
+ }
+ if (workerNode.workerUsage.elu === undefined) {
+ expect(workerNode.workerUsage).toStrictEqual({
+ ...expectedWorkerUsage,
+ elu: undefined
+ })
+ } else {
+ expect(workerNode.workerUsage).toStrictEqual({
+ ...expectedWorkerUsage,
+ elu: {
+ active: expect.any(Number),
+ idle: 0,
+ utilization: 1
+ }
+ })
+ }
+ expect(workerNode.workerUsage.tasks.executed).toBeGreaterThanOrEqual(0)
+ expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
+ max * maxMultiplier
+ )
+ }
+ // We need to clean up the resources after our test
+ await pool.destroy()
+ })
+
it('Verify FAIR_SHARE strategy default tasks usage statistics requirements', async () => {
const workerChoiceStrategy = WorkerChoiceStrategies.FAIR_SHARE
let pool = new FixedThreadPool(
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: true,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: true,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: true,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: true,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: expect.any(Number),
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
elu: undefined
})
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
expect(workerNode.workerUsage.runTime.average).toBeGreaterThan(0)
}
expect(
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: expect.any(Number),
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
elu: undefined
})
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
expect(workerNode.workerUsage.runTime.average).toBeGreaterThan(0)
}
expect(
{
workerChoiceStrategy: WorkerChoiceStrategies.FAIR_SHARE,
workerChoiceStrategyOptions: {
- medRunTime: true
+ runTime: { median: true }
}
}
)
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: 0,
median: expect.any(Number),
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
elu: undefined
})
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
expect(workerNode.workerUsage.runTime.median).toBeGreaterThan(0)
}
expect(
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: true,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: true,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: true,
- avgRunTime: true,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: true,
+ average: true,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: expect.any(Number),
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
max * maxMultiplier
)
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThanOrEqual(
- 0
- )
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThanOrEqual(0)
expect(workerNode.workerUsage.runTime.average).toBeGreaterThanOrEqual(0)
}
expect(
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: expect.any(Number),
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
max * maxMultiplier
)
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
expect(workerNode.workerUsage.runTime.average).toBeGreaterThan(0)
}
expect(
{
workerChoiceStrategy: WorkerChoiceStrategies.WEIGHTED_ROUND_ROBIN,
workerChoiceStrategyOptions: {
- medRunTime: true
+ runTime: { median: true }
}
}
)
failed: 0
},
runTime: {
- aggregation: expect.any(Number),
+ aggregate: expect.any(Number),
average: 0,
median: expect.any(Number),
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
expect(workerNode.workerUsage.tasks.executed).toBeLessThanOrEqual(
max * maxMultiplier
)
- expect(workerNode.workerUsage.runTime.aggregation).toBeGreaterThan(0)
+ expect(workerNode.workerUsage.runTime.aggregate).toBeGreaterThan(0)
expect(workerNode.workerUsage.runTime.median).toBeGreaterThan(0)
}
expect(
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
await pool.destroy()
expect(
pool.workerChoiceStrategyContext.getTaskStatisticsRequirements()
).toStrictEqual({
- runTime: false,
- avgRunTime: false,
- medRunTime: false,
- waitTime: false,
- avgWaitTime: false,
- medWaitTime: false,
+ runTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
+ waitTime: {
+ aggregate: false,
+ average: false,
+ median: false
+ },
elu: false
})
// We need to clean up the resources after our test
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
failed: 0
},
runTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)
},
waitTime: {
- aggregation: 0,
+ aggregate: 0,
average: 0,
median: 0,
history: expect.any(CircularArray)