forked from cerc-io/plugeth
11d0ff6578
* swarm/network/stream: introduced visualized snapshot sync test * swarm/network/stream: non-existing hash visualization sim * swarm/network/stream: fixed retrieval tests; new backend for visualization * swarm/network/stream: cleanup of visualized_snapshot_sync_sim_test.go * swarm/network/stream: rebased PR on master * swarm/network/stream: fixed loop logic in retrieval tests * swarm/network/stream: fixed iterations for snapshot tests * swarm/network/stream: address PR comments * swarm/network/stream: addressed PR comments
226 lines
6.4 KiB
Go
226 lines
6.4 KiB
Go
// Copyright 2018 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
// +build withserver
|
|
|
|
package stream
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
"github.com/ethereum/go-ethereum/p2p/discover"
|
|
"github.com/ethereum/go-ethereum/p2p/simulations"
|
|
"github.com/ethereum/go-ethereum/swarm/log"
|
|
"github.com/ethereum/go-ethereum/swarm/network/simulation"
|
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
|
)
|
|
|
|
/*
|
|
The tests in this file need to be executed with
|
|
|
|
-tags=withserver
|
|
|
|
Also, they will stall if executed stand-alone, because they wait
|
|
for the visualization frontend to send a POST /runsim message.
|
|
*/
|
|
|
|
//setup the sim, evaluate nodeCount and chunkCount and create the sim
|
|
func setupSim(serviceMap map[string]simulation.ServiceFunc) (int, int, *simulation.Simulation) {
|
|
nodeCount := *nodes
|
|
chunkCount := *chunks
|
|
|
|
if nodeCount == 0 || chunkCount == 0 {
|
|
nodeCount = 32
|
|
chunkCount = 1
|
|
}
|
|
|
|
//setup the simulation with server, which means the sim won't run
|
|
//until it receives a POST /runsim from the frontend
|
|
sim := simulation.New(serviceMap).WithServer(":8888")
|
|
return nodeCount, chunkCount, sim
|
|
}
|
|
|
|
//watch for disconnections and wait for healthy
|
|
func watchSim(sim *simulation.Simulation) (context.Context, context.CancelFunc) {
|
|
ctx, cancelSimRun := context.WithTimeout(context.Background(), 1*time.Minute)
|
|
|
|
if _, err := sim.WaitTillHealthy(ctx, 2); err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
disconnections := sim.PeerEvents(
|
|
context.Background(),
|
|
sim.NodeIDs(),
|
|
simulation.NewPeerEventsFilter().Type(p2p.PeerEventTypeDrop),
|
|
)
|
|
|
|
go func() {
|
|
for d := range disconnections {
|
|
log.Error("peer drop", "node", d.NodeID, "peer", d.Event.Peer)
|
|
panic("unexpected disconnect")
|
|
cancelSimRun()
|
|
}
|
|
}()
|
|
|
|
return ctx, cancelSimRun
|
|
}
|
|
|
|
//This test requests bogus hashes into the network
|
|
func TestNonExistingHashesWithServer(t *testing.T) {
|
|
nodeCount, _, sim := setupSim(retrievalSimServiceMap)
|
|
defer sim.Close()
|
|
|
|
err := sim.UploadSnapshot(fmt.Sprintf("testing/snapshot_%d.json", nodeCount))
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
ctx, cancelSimRun := watchSim(sim)
|
|
defer cancelSimRun()
|
|
|
|
//in order to get some meaningful visualization, it is beneficial
|
|
//to define a minimum duration of this test
|
|
testDuration := 20 * time.Second
|
|
|
|
result := sim.Run(ctx, func(ctx context.Context, sim *simulation.Simulation) error {
|
|
//check on the node's FileStore (netstore)
|
|
id := sim.RandomUpNode().ID
|
|
item, ok := sim.NodeItem(id, bucketKeyFileStore)
|
|
if !ok {
|
|
t.Fatalf("No filestore")
|
|
}
|
|
fileStore := item.(*storage.FileStore)
|
|
//create a bogus hash
|
|
fakeHash := storage.GenerateRandomChunk(1000).Address()
|
|
//try to retrieve it - will propagate RetrieveRequestMsg into the network
|
|
reader, _ := fileStore.Retrieve(context.TODO(), fakeHash)
|
|
if _, err := reader.Size(ctx, nil); err != nil {
|
|
log.Debug("expected error for non-existing chunk")
|
|
}
|
|
//sleep so that the frontend can have something to display
|
|
time.Sleep(testDuration)
|
|
|
|
return nil
|
|
})
|
|
if result.Error != nil {
|
|
sendSimTerminatedEvent(sim)
|
|
t.Fatal(result.Error)
|
|
}
|
|
|
|
sendSimTerminatedEvent(sim)
|
|
|
|
}
|
|
|
|
//send a termination event to the frontend
|
|
func sendSimTerminatedEvent(sim *simulation.Simulation) {
|
|
evt := &simulations.Event{
|
|
Type: EventTypeSimTerminated,
|
|
Control: false,
|
|
}
|
|
sim.Net.Events().Send(evt)
|
|
}
|
|
|
|
//This test is the same as the snapshot sync test,
|
|
//but with a HTTP server
|
|
//It also sends some custom events so that the frontend
|
|
//can visualize messages like SendOfferedMsg, WantedHashesMsg, DeliveryMsg
|
|
func TestSnapshotSyncWithServer(t *testing.T) {
|
|
|
|
nodeCount, chunkCount, sim := setupSim(simServiceMap)
|
|
defer sim.Close()
|
|
|
|
log.Info("Initializing test config")
|
|
|
|
conf := &synctestConfig{}
|
|
//map of discover ID to indexes of chunks expected at that ID
|
|
conf.idToChunksMap = make(map[discover.NodeID][]int)
|
|
//map of overlay address to discover ID
|
|
conf.addrToIDMap = make(map[string]discover.NodeID)
|
|
//array where the generated chunk hashes will be stored
|
|
conf.hashes = make([]storage.Address, 0)
|
|
|
|
err := sim.UploadSnapshot(fmt.Sprintf("testing/snapshot_%d.json", nodeCount))
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
ctx, cancelSimRun := watchSim(sim)
|
|
defer cancelSimRun()
|
|
|
|
//setup filters in the event feed
|
|
offeredHashesFilter := simulation.NewPeerEventsFilter().Type(p2p.PeerEventTypeMsgRecv).Protocol("stream").MsgCode(1)
|
|
wantedFilter := simulation.NewPeerEventsFilter().Type(p2p.PeerEventTypeMsgRecv).Protocol("stream").MsgCode(2)
|
|
deliveryFilter := simulation.NewPeerEventsFilter().Type(p2p.PeerEventTypeMsgRecv).Protocol("stream").MsgCode(6)
|
|
eventC := sim.PeerEvents(ctx, sim.UpNodeIDs(), offeredHashesFilter, wantedFilter, deliveryFilter)
|
|
|
|
quit := make(chan struct{})
|
|
|
|
go func() {
|
|
for e := range eventC {
|
|
select {
|
|
case <-quit:
|
|
fmt.Println("quitting event loop")
|
|
return
|
|
default:
|
|
}
|
|
if e.Error != nil {
|
|
t.Fatal(e.Error)
|
|
}
|
|
if *e.Event.MsgCode == uint64(1) {
|
|
evt := &simulations.Event{
|
|
Type: EventTypeChunkOffered,
|
|
Node: sim.Net.GetNode(e.NodeID),
|
|
Control: false,
|
|
}
|
|
sim.Net.Events().Send(evt)
|
|
} else if *e.Event.MsgCode == uint64(2) {
|
|
evt := &simulations.Event{
|
|
Type: EventTypeChunkWanted,
|
|
Node: sim.Net.GetNode(e.NodeID),
|
|
Control: false,
|
|
}
|
|
sim.Net.Events().Send(evt)
|
|
} else if *e.Event.MsgCode == uint64(6) {
|
|
evt := &simulations.Event{
|
|
Type: EventTypeChunkDelivered,
|
|
Node: sim.Net.GetNode(e.NodeID),
|
|
Control: false,
|
|
}
|
|
sim.Net.Events().Send(evt)
|
|
}
|
|
}
|
|
}()
|
|
//run the sim
|
|
result := runSim(conf, ctx, sim, chunkCount)
|
|
|
|
//send terminated event
|
|
evt := &simulations.Event{
|
|
Type: EventTypeSimTerminated,
|
|
Control: false,
|
|
}
|
|
sim.Net.Events().Send(evt)
|
|
|
|
if result.Error != nil {
|
|
panic(result.Error)
|
|
}
|
|
close(quit)
|
|
log.Info("Simulation ended")
|
|
}
|