les: fix and slim the unit tests of les (#20247)
* les: loose restriction of unit tests * les: update unit tests * les, light: slim the unit tests
This commit is contained in:
parent
fc3661f89c
commit
b9bac1f384
@ -459,6 +459,9 @@ func (f *clientPool) addBalance(id enode.ID, amount uint64, setTotal bool) {
|
|||||||
defer func() {
|
defer func() {
|
||||||
c.balanceTracker.setBalance(pb.value, negBalance)
|
c.balanceTracker.setBalance(pb.value, negBalance)
|
||||||
if !c.priority && pb.value > 0 {
|
if !c.priority && pb.value > 0 {
|
||||||
|
// The capacity should be adjusted based on the requirement,
|
||||||
|
// but we have no idea about the new capacity, need a second
|
||||||
|
// call to udpate it.
|
||||||
c.priority = true
|
c.priority = true
|
||||||
c.balanceTracker.addCallback(balanceCallbackZero, 0, func() { f.balanceExhausted(id) })
|
c.balanceTracker.addCallback(balanceCallbackZero, 0, func() { f.balanceExhausted(id) })
|
||||||
}
|
}
|
||||||
|
@ -68,6 +68,14 @@ func (i poolTestPeer) freeClientId() string {
|
|||||||
|
|
||||||
func (i poolTestPeer) updateCapacity(uint64) {}
|
func (i poolTestPeer) updateCapacity(uint64) {}
|
||||||
|
|
||||||
|
type poolTestPeerWithCap struct {
|
||||||
|
poolTestPeer
|
||||||
|
|
||||||
|
cap uint64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (i *poolTestPeerWithCap) updateCapacity(cap uint64) { i.cap = cap }
|
||||||
|
|
||||||
func testClientPool(t *testing.T, connLimit, clientCount, paidCount int, randomDisconnect bool) {
|
func testClientPool(t *testing.T, connLimit, clientCount, paidCount int, randomDisconnect bool) {
|
||||||
rand.Seed(time.Now().UnixNano())
|
rand.Seed(time.Now().UnixNano())
|
||||||
var (
|
var (
|
||||||
@ -308,9 +316,9 @@ func TestFreeClientKickedOut(t *testing.T) {
|
|||||||
|
|
||||||
for i := 0; i < 10; i++ {
|
for i := 0; i < 10; i++ {
|
||||||
pool.connect(poolTestPeer(i), 1)
|
pool.connect(poolTestPeer(i), 1)
|
||||||
clock.Run(100 * time.Millisecond)
|
clock.Run(time.Millisecond)
|
||||||
}
|
}
|
||||||
if pool.connect(poolTestPeer(11), 1) {
|
if pool.connect(poolTestPeer(10), 1) {
|
||||||
t.Fatalf("New free client should be rejected")
|
t.Fatalf("New free client should be rejected")
|
||||||
}
|
}
|
||||||
clock.Run(5 * time.Minute)
|
clock.Run(5 * time.Minute)
|
||||||
@ -320,8 +328,8 @@ func TestFreeClientKickedOut(t *testing.T) {
|
|||||||
for i := 0; i < 10; i++ {
|
for i := 0; i < 10; i++ {
|
||||||
select {
|
select {
|
||||||
case id := <-kicked:
|
case id := <-kicked:
|
||||||
if id != i {
|
if id >= 10 {
|
||||||
t.Fatalf("Kicked client mismatch, want %v, got %v", i, id)
|
t.Fatalf("Old client should be kicked, now got: %d", id)
|
||||||
}
|
}
|
||||||
case <-time.NewTimer(time.Second).C:
|
case <-time.NewTimer(time.Second).C:
|
||||||
t.Fatalf("timeout")
|
t.Fatalf("timeout")
|
||||||
@ -364,11 +372,20 @@ func TestDowngradePriorityClient(t *testing.T) {
|
|||||||
pool.setLimits(10, uint64(10)) // Total capacity limit is 10
|
pool.setLimits(10, uint64(10)) // Total capacity limit is 10
|
||||||
pool.setPriceFactors(priceFactors{1, 0, 1}, priceFactors{1, 0, 1})
|
pool.setPriceFactors(priceFactors{1, 0, 1}, priceFactors{1, 0, 1})
|
||||||
|
|
||||||
pool.addBalance(poolTestPeer(0).ID(), uint64(time.Minute), false)
|
p := &poolTestPeerWithCap{
|
||||||
pool.connect(poolTestPeer(0), 10)
|
poolTestPeer: poolTestPeer(0),
|
||||||
|
}
|
||||||
|
pool.addBalance(p.ID(), uint64(time.Minute), false)
|
||||||
|
pool.connect(p, 10)
|
||||||
|
if p.cap != 10 {
|
||||||
|
t.Fatalf("The capcacity of priority peer hasn't been updated, got: %d", p.cap)
|
||||||
|
}
|
||||||
|
|
||||||
clock.Run(time.Minute) // All positive balance should be used up.
|
clock.Run(time.Minute) // All positive balance should be used up.
|
||||||
time.Sleep(300 * time.Millisecond) // Ensure the callback is called
|
time.Sleep(300 * time.Millisecond) // Ensure the callback is called
|
||||||
|
if p.cap != 1 {
|
||||||
|
t.Fatalf("The capcacity of peer should be downgraded, got: %d", p.cap)
|
||||||
|
}
|
||||||
pb := pool.ndb.getOrNewPB(poolTestPeer(0).ID())
|
pb := pool.ndb.getOrNewPB(poolTestPeer(0).ID())
|
||||||
if pb.value != 0 {
|
if pb.value != 0 {
|
||||||
t.Fatalf("Positive balance mismatch, want %v, got %v", 0, pb.value)
|
t.Fatalf("Positive balance mismatch, want %v, got %v", 0, pb.value)
|
||||||
|
@ -110,13 +110,15 @@ func (d *requestDistributor) registerTestPeer(p distPeer) {
|
|||||||
d.peerLock.Unlock()
|
d.peerLock.Unlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
// distMaxWait is the maximum waiting time after which further necessary waiting
|
var (
|
||||||
// times are recalculated based on new feedback from the servers
|
// distMaxWait is the maximum waiting time after which further necessary waiting
|
||||||
const distMaxWait = time.Millisecond * 50
|
// times are recalculated based on new feedback from the servers
|
||||||
|
distMaxWait = time.Millisecond * 50
|
||||||
|
|
||||||
// waitForPeers is the time window in which a request does not fail even if it
|
// waitForPeers is the time window in which a request does not fail even if it
|
||||||
// has no suitable peers to send to at the moment
|
// has no suitable peers to send to at the moment
|
||||||
const waitForPeers = time.Second * 3
|
waitForPeers = time.Second * 3
|
||||||
|
)
|
||||||
|
|
||||||
// main event loop
|
// main event loop
|
||||||
func (d *requestDistributor) loop() {
|
func (d *requestDistributor) loop() {
|
||||||
|
@ -86,8 +86,8 @@ func (p *testDistPeer) worker(t *testing.T, checkOrder bool, stop chan struct{})
|
|||||||
const (
|
const (
|
||||||
testDistBufLimit = 10000000
|
testDistBufLimit = 10000000
|
||||||
testDistMaxCost = 1000000
|
testDistMaxCost = 1000000
|
||||||
testDistPeerCount = 5
|
testDistPeerCount = 2
|
||||||
testDistReqCount = 5000
|
testDistReqCount = 10
|
||||||
testDistMaxResendCount = 3
|
testDistMaxResendCount = 3
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -128,6 +128,9 @@ func testRequestDistributor(t *testing.T, resend bool) {
|
|||||||
go peers[i].worker(t, !resend, stop)
|
go peers[i].worker(t, !resend, stop)
|
||||||
dist.registerTestPeer(peers[i])
|
dist.registerTestPeer(peers[i])
|
||||||
}
|
}
|
||||||
|
// Disable the mechanism that we will wait a few time for request
|
||||||
|
// even there is no suitable peer to send right now.
|
||||||
|
waitForPeers = 0
|
||||||
|
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
|
|
||||||
|
@ -193,6 +193,9 @@ func testOdr(t *testing.T, protocol int, expFail uint64, checkCached bool, fn od
|
|||||||
if clientHead.Number.Uint64() != 4 {
|
if clientHead.Number.Uint64() != 4 {
|
||||||
t.Fatalf("Failed to sync the chain with server, head: %v", clientHead.Number.Uint64())
|
t.Fatalf("Failed to sync the chain with server, head: %v", clientHead.Number.Uint64())
|
||||||
}
|
}
|
||||||
|
// Disable the mechanism that we will wait a few time for request
|
||||||
|
// even there is no suitable peer to send right now.
|
||||||
|
waitForPeers = 0
|
||||||
|
|
||||||
test := func(expFail uint64) {
|
test := func(expFail uint64) {
|
||||||
// Mark this as a helper to put the failures at the correct lines
|
// Mark this as a helper to put the failures at the correct lines
|
||||||
@ -202,7 +205,9 @@ func testOdr(t *testing.T, protocol int, expFail uint64, checkCached bool, fn od
|
|||||||
bhash := rawdb.ReadCanonicalHash(server.db, i)
|
bhash := rawdb.ReadCanonicalHash(server.db, i)
|
||||||
b1 := fn(light.NoOdr, server.db, server.handler.server.chainConfig, server.handler.blockchain, nil, bhash)
|
b1 := fn(light.NoOdr, server.db, server.handler.server.chainConfig, server.handler.blockchain, nil, bhash)
|
||||||
|
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), 200*time.Millisecond)
|
// Set the timeout as 1 second here, ensure there is enough time
|
||||||
|
// for travis to make the action.
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second)
|
||||||
b2 := fn(ctx, client.db, client.handler.backend.chainConfig, nil, client.handler.backend.blockchain, bhash)
|
b2 := fn(ctx, client.db, client.handler.backend.chainConfig, nil, client.handler.backend.blockchain, bhash)
|
||||||
cancel()
|
cancel()
|
||||||
|
|
||||||
|
@ -89,7 +89,7 @@ func testCheckpointSyncing(t *testing.T, protocol int, syncMode int) {
|
|||||||
for {
|
for {
|
||||||
_, hash, _, err := server.handler.server.oracle.contract.Contract().GetLatestCheckpoint(nil)
|
_, hash, _, err := server.handler.server.oracle.contract.Contract().GetLatestCheckpoint(nil)
|
||||||
if err != nil || hash == [32]byte{} {
|
if err != nil || hash == [32]byte{} {
|
||||||
time.Sleep(100 * time.Millisecond)
|
time.Sleep(10 * time.Millisecond)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
break
|
break
|
||||||
|
@ -71,10 +71,10 @@ var (
|
|||||||
|
|
||||||
var (
|
var (
|
||||||
// The block frequency for creating checkpoint(only used in test)
|
// The block frequency for creating checkpoint(only used in test)
|
||||||
sectionSize = big.NewInt(512)
|
sectionSize = big.NewInt(128)
|
||||||
|
|
||||||
// The number of confirmations needed to generate a checkpoint(only used in test).
|
// The number of confirmations needed to generate a checkpoint(only used in test).
|
||||||
processConfirms = big.NewInt(4)
|
processConfirms = big.NewInt(1)
|
||||||
|
|
||||||
// The token bucket buffer limit for testing purpose.
|
// The token bucket buffer limit for testing purpose.
|
||||||
testBufLimit = uint64(1000000)
|
testBufLimit = uint64(1000000)
|
||||||
|
@ -79,21 +79,21 @@ var (
|
|||||||
}
|
}
|
||||||
// TestServerIndexerConfig wraps a set of configs as a test indexer config for server side.
|
// TestServerIndexerConfig wraps a set of configs as a test indexer config for server side.
|
||||||
TestServerIndexerConfig = &IndexerConfig{
|
TestServerIndexerConfig = &IndexerConfig{
|
||||||
ChtSize: 512,
|
ChtSize: 128,
|
||||||
ChtConfirms: 4,
|
ChtConfirms: 1,
|
||||||
BloomSize: 64,
|
BloomSize: 16,
|
||||||
BloomConfirms: 4,
|
BloomConfirms: 1,
|
||||||
BloomTrieSize: 512,
|
BloomTrieSize: 128,
|
||||||
BloomTrieConfirms: 4,
|
BloomTrieConfirms: 1,
|
||||||
}
|
}
|
||||||
// TestClientIndexerConfig wraps a set of configs as a test indexer config for client side.
|
// TestClientIndexerConfig wraps a set of configs as a test indexer config for client side.
|
||||||
TestClientIndexerConfig = &IndexerConfig{
|
TestClientIndexerConfig = &IndexerConfig{
|
||||||
ChtSize: 512,
|
ChtSize: 128,
|
||||||
ChtConfirms: 32,
|
ChtConfirms: 8,
|
||||||
BloomSize: 512,
|
BloomSize: 128,
|
||||||
BloomConfirms: 32,
|
BloomConfirms: 8,
|
||||||
BloomTrieSize: 512,
|
BloomTrieSize: 128,
|
||||||
BloomTrieConfirms: 32,
|
BloomTrieConfirms: 8,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user