forked from cerc-io/plugeth
swarm/storage/localstore: fix export db.Put signature
cmd/swarm/swarm-smoke: improve smoke tests (#1337) swarm/network: remove dead code (#1339) swarm/network: remove FetchStore and SyncChunkStore in favor of NetStore (#1342)
This commit is contained in:
parent
996755c4a8
commit
993b145f25
@ -26,11 +26,11 @@ const (
|
|||||||
feedRandomDataLength = 8
|
feedRandomDataLength = 8
|
||||||
)
|
)
|
||||||
|
|
||||||
func feedUploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
func feedUploadAndSyncCmd(ctx *cli.Context) error {
|
||||||
errc := make(chan error)
|
errc := make(chan error)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
errc <- feedUploadAndSync(ctx, tuid)
|
errc <- feedUploadAndSync(ctx)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
select {
|
select {
|
||||||
@ -46,7 +46,7 @@ func feedUploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func feedUploadAndSync(c *cli.Context, tuid string) error {
|
func feedUploadAndSync(c *cli.Context) error {
|
||||||
log.Info("generating and uploading feeds to " + httpEndpoint(hosts[0]) + " and syncing")
|
log.Info("generating and uploading feeds to " + httpEndpoint(hosts[0]) + " and syncing")
|
||||||
|
|
||||||
// create a random private key to sign updates with and derive the address
|
// create a random private key to sign updates with and derive the address
|
||||||
@ -272,7 +272,7 @@ func feedUploadAndSync(c *cli.Context, tuid string) error {
|
|||||||
ruid := uuid.New()[:8]
|
ruid := uuid.New()[:8]
|
||||||
go func(url string, endpoint string, ruid string) {
|
go func(url string, endpoint string, ruid string) {
|
||||||
for {
|
for {
|
||||||
err := fetch(url, endpoint, fileHash, ruid, "")
|
err := fetch(url, endpoint, fileHash, ruid)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
@ -40,14 +40,14 @@ var (
|
|||||||
allhosts string
|
allhosts string
|
||||||
hosts []string
|
hosts []string
|
||||||
filesize int
|
filesize int
|
||||||
inputSeed int
|
|
||||||
syncDelay int
|
syncDelay int
|
||||||
|
inputSeed int
|
||||||
httpPort int
|
httpPort int
|
||||||
wsPort int
|
wsPort int
|
||||||
verbosity int
|
verbosity int
|
||||||
timeout int
|
timeout int
|
||||||
single bool
|
single bool
|
||||||
trackTimeout int
|
onlyUpload bool
|
||||||
)
|
)
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
@ -101,7 +101,7 @@ func main() {
|
|||||||
},
|
},
|
||||||
cli.IntFlag{
|
cli.IntFlag{
|
||||||
Name: "timeout",
|
Name: "timeout",
|
||||||
Value: 120,
|
Value: 180,
|
||||||
Usage: "timeout in seconds after which kill the process",
|
Usage: "timeout in seconds after which kill the process",
|
||||||
Destination: &timeout,
|
Destination: &timeout,
|
||||||
},
|
},
|
||||||
@ -110,11 +110,10 @@ func main() {
|
|||||||
Usage: "whether to fetch content from a single node or from all nodes",
|
Usage: "whether to fetch content from a single node or from all nodes",
|
||||||
Destination: &single,
|
Destination: &single,
|
||||||
},
|
},
|
||||||
cli.IntFlag{
|
cli.BoolFlag{
|
||||||
Name: "track-timeout",
|
Name: "only-upload",
|
||||||
Value: 5,
|
Usage: "whether to only upload content to a single node without fetching",
|
||||||
Usage: "timeout in seconds to wait for GetAllReferences to return",
|
Destination: &onlyUpload,
|
||||||
Destination: &trackTimeout,
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -35,11 +35,11 @@ type uploadResult struct {
|
|||||||
digest []byte
|
digest []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
func slidingWindowCmd(ctx *cli.Context, tuid string) error {
|
func slidingWindowCmd(ctx *cli.Context) error {
|
||||||
errc := make(chan error)
|
errc := make(chan error)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
errc <- slidingWindow(ctx, tuid)
|
errc <- slidingWindow(ctx)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
err := <-errc
|
err := <-errc
|
||||||
@ -49,10 +49,10 @@ func slidingWindowCmd(ctx *cli.Context, tuid string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func slidingWindow(ctx *cli.Context, tuid string) error {
|
func slidingWindow(ctx *cli.Context) error {
|
||||||
var hashes []uploadResult //swarm hashes of the uploads
|
var hashes []uploadResult //swarm hashes of the uploads
|
||||||
nodes := len(hosts)
|
nodes := len(hosts)
|
||||||
log.Info("sliding window test started", "tuid", tuid, "nodes", nodes, "filesize(kb)", filesize, "timeout", timeout)
|
log.Info("sliding window test started", "nodes", nodes, "filesize(kb)", filesize, "timeout", timeout)
|
||||||
uploadedBytes := 0
|
uploadedBytes := 0
|
||||||
networkDepth := 0
|
networkDepth := 0
|
||||||
errored := false
|
errored := false
|
||||||
@ -107,7 +107,7 @@ outer:
|
|||||||
start = time.Now()
|
start = time.Now()
|
||||||
// fetch hangs when swarm dies out, so we have to jump through a bit more hoops to actually
|
// fetch hangs when swarm dies out, so we have to jump through a bit more hoops to actually
|
||||||
// catch the timeout, but also allow this retry logic
|
// catch the timeout, but also allow this retry logic
|
||||||
err := fetch(v.hash, httpEndpoint(hosts[idx]), v.digest, ruid, "")
|
err := fetch(v.hash, httpEndpoint(hosts[idx]), v.digest, ruid)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("error fetching hash", "err", err)
|
log.Error("error fetching hash", "err", err)
|
||||||
continue
|
continue
|
||||||
|
@ -23,22 +23,20 @@ import (
|
|||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
"github.com/ethereum/go-ethereum/swarm/api"
|
|
||||||
"github.com/ethereum/go-ethereum/swarm/storage"
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
"github.com/ethereum/go-ethereum/swarm/testutil"
|
"github.com/ethereum/go-ethereum/swarm/testutil"
|
||||||
"github.com/pborman/uuid"
|
|
||||||
|
|
||||||
cli "gopkg.in/urfave/cli.v1"
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
func uploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
func uploadAndSyncCmd(ctx *cli.Context) error {
|
||||||
// use input seed if it has been set
|
// use input seed if it has been set
|
||||||
if inputSeed != 0 {
|
if inputSeed != 0 {
|
||||||
seed = inputSeed
|
seed = inputSeed
|
||||||
@ -49,7 +47,7 @@ func uploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
|||||||
errc := make(chan error)
|
errc := make(chan error)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
errc <- uploadAndSync(ctx, randomBytes, tuid)
|
errc <- uploadAndSync(ctx, randomBytes)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
@ -65,7 +63,7 @@ func uploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// trigger debug functionality on randomBytes
|
// trigger debug functionality on randomBytes
|
||||||
e := trackChunks(randomBytes[:])
|
e := trackChunks(randomBytes[:], true)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
log.Error(e.Error())
|
log.Error(e.Error())
|
||||||
}
|
}
|
||||||
@ -73,50 +71,84 @@ func uploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func trackChunks(testData []byte) error {
|
func trackChunks(testData []byte, submitMetrics bool) error {
|
||||||
addrs, err := getAllRefs(testData)
|
addrs, err := getAllRefs(testData)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
for i, ref := range addrs {
|
for i, ref := range addrs {
|
||||||
log.Trace(fmt.Sprintf("ref %d", i), "ref", ref)
|
log.Debug(fmt.Sprintf("ref %d", i), "ref", ref)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var globalYes, globalNo int
|
||||||
|
var globalMu sync.Mutex
|
||||||
|
var hasErr bool
|
||||||
|
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
wg.Add(len(hosts))
|
||||||
|
|
||||||
for _, host := range hosts {
|
for _, host := range hosts {
|
||||||
|
host := host
|
||||||
|
go func() {
|
||||||
|
defer wg.Done()
|
||||||
httpHost := fmt.Sprintf("ws://%s:%d", host, 8546)
|
httpHost := fmt.Sprintf("ws://%s:%d", host, 8546)
|
||||||
|
|
||||||
hostChunks := []string{}
|
ctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)
|
||||||
|
defer cancel()
|
||||||
rpcClient, err := rpc.Dial(httpHost)
|
rpcClient, err := rpc.DialContext(ctx, httpHost)
|
||||||
|
if rpcClient != nil {
|
||||||
|
defer rpcClient.Close()
|
||||||
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("error dialing host", "err", err, "host", httpHost)
|
log.Error("error dialing host", "err", err, "host", httpHost)
|
||||||
continue
|
hasErr = true
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
var hasInfo []api.HasInfo
|
var hostChunks string
|
||||||
err = rpcClient.Call(&hasInfo, "bzz_has", addrs)
|
err = rpcClient.Call(&hostChunks, "bzz_has", addrs)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("error calling rpc client", "err", err, "host", httpHost)
|
log.Error("error calling rpc client", "err", err, "host", httpHost)
|
||||||
continue
|
hasErr = true
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
count := 0
|
yes, no := 0, 0
|
||||||
for _, info := range hasInfo {
|
for _, val := range hostChunks {
|
||||||
if info.Has {
|
if val == '1' {
|
||||||
hostChunks = append(hostChunks, "1")
|
yes++
|
||||||
} else {
|
} else {
|
||||||
hostChunks = append(hostChunks, "0")
|
no++
|
||||||
count++
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if count == 0 {
|
if no == 0 {
|
||||||
log.Info("host reported to have all chunks", "host", host)
|
log.Info("host reported to have all chunks", "host", host)
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Trace("chunks", "chunks", strings.Join(hostChunks, ""), "host", host)
|
log.Debug("chunks", "chunks", hostChunks, "yes", yes, "no", no, "host", host)
|
||||||
|
|
||||||
|
if submitMetrics {
|
||||||
|
globalMu.Lock()
|
||||||
|
globalYes += yes
|
||||||
|
globalNo += no
|
||||||
|
globalMu.Unlock()
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
|
if !hasErr && submitMetrics {
|
||||||
|
// remove the chunks stored on the uploader node
|
||||||
|
globalYes -= len(addrs)
|
||||||
|
|
||||||
|
metrics.GetOrRegisterCounter("deployment.chunks.yes", nil).Inc(int64(globalYes))
|
||||||
|
metrics.GetOrRegisterCounter("deployment.chunks.no", nil).Inc(int64(globalNo))
|
||||||
|
metrics.GetOrRegisterCounter("deployment.chunks.refs", nil).Inc(int64(len(addrs)))
|
||||||
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -130,15 +162,13 @@ func getAllRefs(testData []byte) (storage.AddressCollection, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), time.Duration(trackTimeout)*time.Second)
|
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
reader := bytes.NewReader(testData)
|
reader := bytes.NewReader(testData)
|
||||||
return fileStore.GetAllReferences(ctx, reader, false)
|
return fileStore.GetAllReferences(context.Background(), reader, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
func uploadAndSync(c *cli.Context, randomBytes []byte, tuid string) error {
|
func uploadAndSync(c *cli.Context, randomBytes []byte) error {
|
||||||
log.Info("uploading to "+httpEndpoint(hosts[0])+" and syncing", "tuid", tuid, "seed", seed)
|
log.Info("uploading to "+httpEndpoint(hosts[0])+" and syncing", "seed", seed)
|
||||||
|
|
||||||
t1 := time.Now()
|
t1 := time.Now()
|
||||||
hash, err := upload(randomBytes, httpEndpoint(hosts[0]))
|
hash, err := upload(randomBytes, httpEndpoint(hosts[0]))
|
||||||
@ -155,53 +185,91 @@ func uploadAndSync(c *cli.Context, randomBytes []byte, tuid string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Info("uploaded successfully", "tuid", tuid, "hash", hash, "took", t2, "digest", fmt.Sprintf("%x", fhash))
|
log.Info("uploaded successfully", "hash", hash, "took", t2, "digest", fmt.Sprintf("%x", fhash))
|
||||||
|
|
||||||
time.Sleep(time.Duration(syncDelay) * time.Second)
|
waitToSync()
|
||||||
|
|
||||||
|
log.Debug("chunks before fetch attempt", "hash", hash)
|
||||||
|
|
||||||
|
err = trackChunks(randomBytes, false)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
if onlyUpload {
|
||||||
|
log.Debug("only-upload is true, stoppping test", "hash", hash)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
wg := sync.WaitGroup{}
|
|
||||||
if single {
|
|
||||||
randIndex := 1 + rand.Intn(len(hosts)-1)
|
randIndex := 1 + rand.Intn(len(hosts)-1)
|
||||||
ruid := uuid.New()[:8]
|
|
||||||
wg.Add(1)
|
|
||||||
go func(endpoint string, ruid string) {
|
|
||||||
for {
|
for {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
err := fetch(hash, endpoint, fhash, ruid, tuid)
|
err := fetch(hash, httpEndpoint(hosts[randIndex]), fhash, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
time.Sleep(2 * time.Second)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
ended := time.Since(start)
|
ended := time.Since(start)
|
||||||
|
|
||||||
metrics.GetOrRegisterResettingTimer("upload-and-sync.single.fetch-time", nil).Update(ended)
|
metrics.GetOrRegisterResettingTimer("upload-and-sync.single.fetch-time", nil).Update(ended)
|
||||||
log.Info("fetch successful", "tuid", tuid, "ruid", ruid, "took", ended, "endpoint", endpoint)
|
log.Info("fetch successful", "took", ended, "endpoint", httpEndpoint(hosts[randIndex]))
|
||||||
wg.Done()
|
break
|
||||||
return
|
|
||||||
}
|
}
|
||||||
}(httpEndpoint(hosts[randIndex]), ruid)
|
|
||||||
} else {
|
|
||||||
for _, endpoint := range hosts[1:] {
|
|
||||||
ruid := uuid.New()[:8]
|
|
||||||
wg.Add(1)
|
|
||||||
go func(endpoint string, ruid string) {
|
|
||||||
for {
|
|
||||||
start := time.Now()
|
|
||||||
err := fetch(hash, endpoint, fhash, ruid, tuid)
|
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
ended := time.Since(start)
|
|
||||||
|
|
||||||
metrics.GetOrRegisterResettingTimer("upload-and-sync.each.fetch-time", nil).Update(ended)
|
|
||||||
log.Info("fetch successful", "tuid", tuid, "ruid", ruid, "took", ended, "endpoint", endpoint)
|
|
||||||
wg.Done()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}(httpEndpoint(endpoint), ruid)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
wg.Wait()
|
|
||||||
log.Info("all hosts synced random file successfully")
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func isSyncing(wsHost string) (bool, error) {
|
||||||
|
rpcClient, err := rpc.Dial(wsHost)
|
||||||
|
if rpcClient != nil {
|
||||||
|
defer rpcClient.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
log.Error("error dialing host", "err", err)
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var isSyncing bool
|
||||||
|
err = rpcClient.Call(&isSyncing, "bzz_isSyncing")
|
||||||
|
if err != nil {
|
||||||
|
log.Error("error calling host for isSyncing", "err", err)
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Debug("isSyncing result", "host", wsHost, "isSyncing", isSyncing)
|
||||||
|
|
||||||
|
return isSyncing, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func waitToSync() {
|
||||||
|
t1 := time.Now()
|
||||||
|
|
||||||
|
ns := uint64(1)
|
||||||
|
|
||||||
|
for ns > 0 {
|
||||||
|
time.Sleep(3 * time.Second)
|
||||||
|
|
||||||
|
notSynced := uint64(0)
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
wg.Add(len(hosts))
|
||||||
|
for i := 0; i < len(hosts); i++ {
|
||||||
|
i := i
|
||||||
|
go func(idx int) {
|
||||||
|
stillSyncing, err := isSyncing(wsEndpoint(hosts[idx]))
|
||||||
|
|
||||||
|
if stillSyncing || err != nil {
|
||||||
|
atomic.AddUint64(¬Synced, 1)
|
||||||
|
}
|
||||||
|
wg.Done()
|
||||||
|
}(i)
|
||||||
|
}
|
||||||
|
wg.Wait()
|
||||||
|
|
||||||
|
ns = atomic.LoadUint64(¬Synced)
|
||||||
|
}
|
||||||
|
|
||||||
|
t2 := time.Since(t1)
|
||||||
|
metrics.GetOrRegisterResettingTimer("upload-and-sync.single.wait-for-sync.deployment", nil).Update(t2)
|
||||||
|
}
|
||||||
|
@ -28,14 +28,14 @@ import (
|
|||||||
cli "gopkg.in/urfave/cli.v1"
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
func uploadSpeedCmd(ctx *cli.Context, tuid string) error {
|
func uploadSpeedCmd(ctx *cli.Context) error {
|
||||||
log.Info("uploading to "+hosts[0], "tuid", tuid, "seed", seed)
|
log.Info("uploading to "+hosts[0], "seed", seed)
|
||||||
randomBytes := testutil.RandomBytes(seed, filesize*1000)
|
randomBytes := testutil.RandomBytes(seed, filesize*1000)
|
||||||
|
|
||||||
errc := make(chan error)
|
errc := make(chan error)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
errc <- uploadSpeed(ctx, tuid, randomBytes)
|
errc <- uploadSpeed(ctx, randomBytes)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
select {
|
select {
|
||||||
@ -53,7 +53,7 @@ func uploadSpeedCmd(ctx *cli.Context, tuid string) error {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func uploadSpeed(c *cli.Context, tuid string, data []byte) error {
|
func uploadSpeed(c *cli.Context, data []byte) error {
|
||||||
t1 := time.Now()
|
t1 := time.Now()
|
||||||
hash, err := upload(data, hosts[0])
|
hash, err := upload(data, hosts[0])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -38,7 +38,6 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/swarm/api/client"
|
"github.com/ethereum/go-ethereum/swarm/api/client"
|
||||||
"github.com/ethereum/go-ethereum/swarm/spancontext"
|
"github.com/ethereum/go-ethereum/swarm/spancontext"
|
||||||
opentracing "github.com/opentracing/opentracing-go"
|
opentracing "github.com/opentracing/opentracing-go"
|
||||||
"github.com/pborman/uuid"
|
|
||||||
cli "gopkg.in/urfave/cli.v1"
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -59,28 +58,25 @@ func wsEndpoint(host string) string {
|
|||||||
return fmt.Sprintf("ws://%s:%d", host, wsPort)
|
return fmt.Sprintf("ws://%s:%d", host, wsPort)
|
||||||
}
|
}
|
||||||
|
|
||||||
func wrapCliCommand(name string, command func(*cli.Context, string) error) func(*cli.Context) error {
|
func wrapCliCommand(name string, command func(*cli.Context) error) func(*cli.Context) error {
|
||||||
return func(ctx *cli.Context) error {
|
return func(ctx *cli.Context) error {
|
||||||
log.PrintOrigins(true)
|
log.PrintOrigins(true)
|
||||||
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(verbosity), log.StreamHandler(os.Stdout, log.TerminalFormat(false))))
|
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(verbosity), log.StreamHandler(os.Stdout, log.TerminalFormat(false))))
|
||||||
|
|
||||||
// test uuid
|
|
||||||
tuid := uuid.New()[:8]
|
|
||||||
|
|
||||||
commandName = name
|
commandName = name
|
||||||
|
|
||||||
hosts = strings.Split(allhosts, ",")
|
hosts = strings.Split(allhosts, ",")
|
||||||
|
|
||||||
defer func(now time.Time) {
|
defer func(now time.Time) {
|
||||||
totalTime := time.Since(now)
|
totalTime := time.Since(now)
|
||||||
log.Info("total time", "tuid", tuid, "time", totalTime, "kb", filesize)
|
log.Info("total time", "time", totalTime, "kb", filesize)
|
||||||
metrics.GetOrRegisterResettingTimer(name+".total-time", nil).Update(totalTime)
|
metrics.GetOrRegisterResettingTimer(name+".total-time", nil).Update(totalTime)
|
||||||
}(time.Now())
|
}(time.Now())
|
||||||
|
|
||||||
log.Info("smoke test starting", "tuid", tuid, "task", name, "timeout", timeout)
|
log.Info("smoke test starting", "task", name, "timeout", timeout)
|
||||||
metrics.GetOrRegisterCounter(name, nil).Inc(1)
|
metrics.GetOrRegisterCounter(name, nil).Inc(1)
|
||||||
|
|
||||||
return command(ctx, tuid)
|
return command(ctx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -142,11 +138,11 @@ func fetchFeed(topic string, user string, endpoint string, original []byte, ruid
|
|||||||
}
|
}
|
||||||
|
|
||||||
// fetch is getting the requested `hash` from the `endpoint` and compares it with the `original` file
|
// fetch is getting the requested `hash` from the `endpoint` and compares it with the `original` file
|
||||||
func fetch(hash string, endpoint string, original []byte, ruid string, tuid string) error {
|
func fetch(hash string, endpoint string, original []byte, ruid string) error {
|
||||||
ctx, sp := spancontext.StartSpan(context.Background(), "upload-and-sync.fetch")
|
ctx, sp := spancontext.StartSpan(context.Background(), "upload-and-sync.fetch")
|
||||||
defer sp.Finish()
|
defer sp.Finish()
|
||||||
|
|
||||||
log.Info("http get request", "tuid", tuid, "ruid", ruid, "endpoint", endpoint, "hash", hash)
|
log.Info("http get request", "ruid", ruid, "endpoint", endpoint, "hash", hash)
|
||||||
|
|
||||||
var tn time.Time
|
var tn time.Time
|
||||||
reqUri := endpoint + "/bzz:/" + hash + "/"
|
reqUri := endpoint + "/bzz:/" + hash + "/"
|
||||||
@ -170,7 +166,7 @@ func fetch(hash string, endpoint string, original []byte, ruid string, tuid stri
|
|||||||
log.Error(err.Error(), "ruid", ruid)
|
log.Error(err.Error(), "ruid", ruid)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
log.Info("http get response", "tuid", tuid, "ruid", ruid, "endpoint", endpoint, "hash", hash, "code", res.StatusCode, "len", res.ContentLength)
|
log.Info("http get response", "ruid", ruid, "endpoint", endpoint, "hash", hash, "code", res.StatusCode, "len", res.ContentLength)
|
||||||
|
|
||||||
if res.StatusCode != 200 {
|
if res.StatusCode != 200 {
|
||||||
err := fmt.Errorf("expected status code %d, got %v", 200, res.StatusCode)
|
err := fmt.Errorf("expected status code %d, got %v", 200, res.StatusCode)
|
||||||
|
@ -243,7 +243,7 @@ func (p *Peer) Run(handler func(ctx context.Context, msg interface{}) error) err
|
|||||||
// Drop disconnects a peer.
|
// Drop disconnects a peer.
|
||||||
// TODO: may need to implement protocol drop only? don't want to kick off the peer
|
// TODO: may need to implement protocol drop only? don't want to kick off the peer
|
||||||
// if they are useful for other protocols
|
// if they are useful for other protocols
|
||||||
func (p *Peer) Drop(err error) {
|
func (p *Peer) Drop() {
|
||||||
p.Disconnect(p2p.DiscSubprotocolError)
|
p.Disconnect(p2p.DiscSubprotocolError)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -291,7 +291,7 @@ func (p *Peer) Send(ctx context.Context, msg interface{}) error {
|
|||||||
if p.spec.Hook != nil {
|
if p.spec.Hook != nil {
|
||||||
err := p.spec.Hook.Send(p, wmsg.Size, msg)
|
err := p.spec.Hook.Send(p, wmsg.Size, msg)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
p.Drop(err)
|
p.Drop()
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -126,7 +126,7 @@ func newProtocol(pp *p2ptest.TestPeerPool) func(*p2p.Peer, p2p.MsgReadWriter) er
|
|||||||
case *kill:
|
case *kill:
|
||||||
// demonstrates use of peerPool, killing another peer connection as a response to a message
|
// demonstrates use of peerPool, killing another peer connection as a response to a message
|
||||||
id := msg.C
|
id := msg.C
|
||||||
pp.Get(id).Drop(errors.New("killed"))
|
pp.Get(id).Drop()
|
||||||
return nil
|
return nil
|
||||||
|
|
||||||
case *drop:
|
case *drop:
|
||||||
|
@ -26,7 +26,7 @@ import (
|
|||||||
|
|
||||||
type TestPeer interface {
|
type TestPeer interface {
|
||||||
ID() enode.ID
|
ID() enode.ID
|
||||||
Drop(error)
|
Drop()
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestPeerPool is an example peerPool to demonstrate registration of peer connections
|
// TestPeerPool is an example peerPool to demonstrate registration of peer connections
|
||||||
|
@ -19,7 +19,11 @@ package api
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/log"
|
||||||
"github.com/ethereum/go-ethereum/swarm/network"
|
"github.com/ethereum/go-ethereum/swarm/network"
|
||||||
"github.com/ethereum/go-ethereum/swarm/storage"
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
)
|
)
|
||||||
@ -47,25 +51,34 @@ func (inspector *Inspector) ListKnown() []string {
|
|||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
||||||
type HasInfo struct {
|
func (inspector *Inspector) IsSyncing() bool {
|
||||||
Addr string `json:"address"`
|
lastReceivedChunksMsg := metrics.GetOrRegisterGauge("network.stream.received_chunks", nil)
|
||||||
Has bool `json:"has"`
|
|
||||||
|
// last received chunks msg time
|
||||||
|
lrct := time.Unix(0, lastReceivedChunksMsg.Value())
|
||||||
|
|
||||||
|
// if last received chunks msg time is after now-15sec. (i.e. within the last 15sec.) then we say that the node is still syncing
|
||||||
|
// technically this is not correct, because this might have been a retrieve request, but for the time being it works for our purposes
|
||||||
|
// because we know we are not making retrieve requests on the node while checking this
|
||||||
|
return lrct.After(time.Now().Add(-15 * time.Second))
|
||||||
}
|
}
|
||||||
|
|
||||||
// Has checks whether each chunk address is present in the underlying datastore,
|
// Has checks whether each chunk address is present in the underlying datastore,
|
||||||
// the bool in the returned structs indicates if the underlying datastore has
|
// the bool in the returned structs indicates if the underlying datastore has
|
||||||
// the chunk stored with the given address (true), or not (false)
|
// the chunk stored with the given address (true), or not (false)
|
||||||
func (inspector *Inspector) Has(chunkAddresses []storage.Address) []HasInfo {
|
func (inspector *Inspector) Has(chunkAddresses []storage.Address) string {
|
||||||
results := make([]HasInfo, 0)
|
hostChunks := []string{}
|
||||||
for _, addr := range chunkAddresses {
|
for _, addr := range chunkAddresses {
|
||||||
res := HasInfo{}
|
|
||||||
res.Addr = addr.String()
|
|
||||||
has, err := inspector.netStore.Has(context.Background(), addr)
|
has, err := inspector.netStore.Has(context.Background(), addr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
has = false
|
log.Error(err.Error())
|
||||||
}
|
}
|
||||||
res.Has = has
|
if has {
|
||||||
results = append(results, res)
|
hostChunks = append(hostChunks, "1")
|
||||||
|
} else {
|
||||||
|
hostChunks = append(hostChunks, "0")
|
||||||
}
|
}
|
||||||
return results
|
}
|
||||||
|
|
||||||
|
return strings.Join(hostChunks, "")
|
||||||
}
|
}
|
||||||
|
@ -172,12 +172,6 @@ type Store interface {
|
|||||||
Close() (err error)
|
Close() (err error)
|
||||||
}
|
}
|
||||||
|
|
||||||
// FetchStore is a Store which supports syncing
|
|
||||||
type FetchStore interface {
|
|
||||||
Store
|
|
||||||
FetchFunc(ctx context.Context, addr Address) func(context.Context) error
|
|
||||||
}
|
|
||||||
|
|
||||||
// Validator validates a chunk.
|
// Validator validates a chunk.
|
||||||
type Validator interface {
|
type Validator interface {
|
||||||
Validate(ch Chunk) bool
|
Validate(ch Chunk) bool
|
||||||
|
@ -116,7 +116,7 @@ func (h *Hive) Stop() error {
|
|||||||
log.Info(fmt.Sprintf("%08x hive stopped, dropping peers", h.BaseAddr()[:4]))
|
log.Info(fmt.Sprintf("%08x hive stopped, dropping peers", h.BaseAddr()[:4]))
|
||||||
h.EachConn(nil, 255, func(p *Peer, _ int) bool {
|
h.EachConn(nil, 255, func(p *Peer, _ int) bool {
|
||||||
log.Info(fmt.Sprintf("%08x dropping peer %08x", h.BaseAddr()[:4], p.Address()[:4]))
|
log.Info(fmt.Sprintf("%08x dropping peer %08x", h.BaseAddr()[:4], p.Address()[:4]))
|
||||||
p.Drop(nil)
|
p.Drop()
|
||||||
return true
|
return true
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -25,6 +25,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
"github.com/ethereum/go-ethereum/swarm/log"
|
"github.com/ethereum/go-ethereum/swarm/log"
|
||||||
"github.com/ethereum/go-ethereum/swarm/pot"
|
"github.com/ethereum/go-ethereum/swarm/pot"
|
||||||
sv "github.com/ethereum/go-ethereum/swarm/version"
|
sv "github.com/ethereum/go-ethereum/swarm/version"
|
||||||
@ -138,6 +139,9 @@ func (e *entry) Hex() string {
|
|||||||
func (k *Kademlia) Register(peers ...*BzzAddr) error {
|
func (k *Kademlia) Register(peers ...*BzzAddr) error {
|
||||||
k.lock.Lock()
|
k.lock.Lock()
|
||||||
defer k.lock.Unlock()
|
defer k.lock.Unlock()
|
||||||
|
|
||||||
|
metrics.GetOrRegisterCounter("kad.register", nil).Inc(1)
|
||||||
|
|
||||||
var known, size int
|
var known, size int
|
||||||
for _, p := range peers {
|
for _, p := range peers {
|
||||||
log.Trace("kademlia trying to register", "addr", p)
|
log.Trace("kademlia trying to register", "addr", p)
|
||||||
@ -164,8 +168,6 @@ func (k *Kademlia) Register(peers ...*BzzAddr) error {
|
|||||||
return newEntry(p)
|
return newEntry(p)
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Trace("found among known peers, underlay addr is same, do nothing", "new", p, "old", e.BzzAddr)
|
|
||||||
|
|
||||||
return v
|
return v
|
||||||
})
|
})
|
||||||
if found {
|
if found {
|
||||||
@ -186,6 +188,9 @@ func (k *Kademlia) Register(peers ...*BzzAddr) error {
|
|||||||
func (k *Kademlia) SuggestPeer() (suggestedPeer *BzzAddr, saturationDepth int, changed bool) {
|
func (k *Kademlia) SuggestPeer() (suggestedPeer *BzzAddr, saturationDepth int, changed bool) {
|
||||||
k.lock.Lock()
|
k.lock.Lock()
|
||||||
defer k.lock.Unlock()
|
defer k.lock.Unlock()
|
||||||
|
|
||||||
|
metrics.GetOrRegisterCounter("kad.suggestpeer", nil).Inc(1)
|
||||||
|
|
||||||
radius := neighbourhoodRadiusForPot(k.conns, k.NeighbourhoodSize, k.base)
|
radius := neighbourhoodRadiusForPot(k.conns, k.NeighbourhoodSize, k.base)
|
||||||
// collect undersaturated bins in ascending order of number of connected peers
|
// collect undersaturated bins in ascending order of number of connected peers
|
||||||
// and from shallow to deep (ascending order of PO)
|
// and from shallow to deep (ascending order of PO)
|
||||||
@ -297,6 +302,9 @@ func (k *Kademlia) SuggestPeer() (suggestedPeer *BzzAddr, saturationDepth int, c
|
|||||||
func (k *Kademlia) On(p *Peer) (uint8, bool) {
|
func (k *Kademlia) On(p *Peer) (uint8, bool) {
|
||||||
k.lock.Lock()
|
k.lock.Lock()
|
||||||
defer k.lock.Unlock()
|
defer k.lock.Unlock()
|
||||||
|
|
||||||
|
metrics.GetOrRegisterCounter("kad.on", nil).Inc(1)
|
||||||
|
|
||||||
var ins bool
|
var ins bool
|
||||||
k.conns, _, _, _ = pot.Swap(k.conns, p, Pof, func(v pot.Val) pot.Val {
|
k.conns, _, _, _ = pot.Swap(k.conns, p, Pof, func(v pot.Val) pot.Val {
|
||||||
// if not found live
|
// if not found live
|
||||||
@ -320,7 +328,6 @@ func (k *Kademlia) On(p *Peer) (uint8, bool) {
|
|||||||
k.addrCountC <- k.addrs.Size()
|
k.addrCountC <- k.addrs.Size()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
log.Trace(k.string())
|
|
||||||
// calculate if depth of saturation changed
|
// calculate if depth of saturation changed
|
||||||
depth := uint8(k.saturation())
|
depth := uint8(k.saturation())
|
||||||
var changed bool
|
var changed bool
|
||||||
@ -608,7 +615,7 @@ func (k *Kademlia) string() string {
|
|||||||
if len(sv.GitCommit) > 0 {
|
if len(sv.GitCommit) > 0 {
|
||||||
rows = append(rows, fmt.Sprintf("commit hash: %s", sv.GitCommit))
|
rows = append(rows, fmt.Sprintf("commit hash: %s", sv.GitCommit))
|
||||||
}
|
}
|
||||||
rows = append(rows, fmt.Sprintf("%v KΛÐΞMLIΛ hive: queen's address: %x", time.Now().UTC().Format(time.UnixDate), k.BaseAddr()[:3]))
|
rows = append(rows, fmt.Sprintf("%v KΛÐΞMLIΛ hive: queen's address: %x", time.Now().UTC().Format(time.UnixDate), k.BaseAddr()))
|
||||||
rows = append(rows, fmt.Sprintf("population: %d (%d), NeighbourhoodSize: %d, MinBinSize: %d, MaxBinSize: %d", k.conns.Size(), k.addrs.Size(), k.NeighbourhoodSize, k.MinBinSize, k.MaxBinSize))
|
rows = append(rows, fmt.Sprintf("population: %d (%d), NeighbourhoodSize: %d, MinBinSize: %d, MaxBinSize: %d", k.conns.Size(), k.addrs.Size(), k.NeighbourhoodSize, k.MinBinSize, k.MaxBinSize))
|
||||||
|
|
||||||
liverows := make([]string, k.MaxProxDisplay)
|
liverows := make([]string, k.MaxProxDisplay)
|
||||||
|
@ -541,7 +541,7 @@ func TestKademliaHiveString(t *testing.T) {
|
|||||||
tk.Register("10000000", "10000001")
|
tk.Register("10000000", "10000001")
|
||||||
tk.MaxProxDisplay = 8
|
tk.MaxProxDisplay = 8
|
||||||
h := tk.String()
|
h := tk.String()
|
||||||
expH := "\n=========================================================================\nMon Feb 27 12:10:28 UTC 2017 KΛÐΞMLIΛ hive: queen's address: 000000\npopulation: 2 (4), NeighbourhoodSize: 2, MinBinSize: 2, MaxBinSize: 4\n============ DEPTH: 0 ==========================================\n000 0 | 2 8100 (0) 8000 (0)\n001 1 4000 | 1 4000 (0)\n002 1 2000 | 1 2000 (0)\n003 0 | 0\n004 0 | 0\n005 0 | 0\n006 0 | 0\n007 0 | 0\n========================================================================="
|
expH := "\n=========================================================================\nMon Feb 27 12:10:28 UTC 2017 KΛÐΞMLIΛ hive: queen's address: 0000000000000000000000000000000000000000000000000000000000000000\npopulation: 2 (4), NeighbourhoodSize: 2, MinBinSize: 2, MaxBinSize: 4\n============ DEPTH: 0 ==========================================\n000 0 | 2 8100 (0) 8000 (0)\n001 1 4000 | 1 4000 (0)\n002 1 2000 | 1 2000 (0)\n003 0 | 0\n004 0 | 0\n005 0 | 0\n006 0 | 0\n007 0 | 0\n========================================================================="
|
||||||
if expH[104:] != h[104:] {
|
if expH[104:] != h[104:] {
|
||||||
t.Fatalf("incorrect hive output. expected %v, got %v", expH, h)
|
t.Fatalf("incorrect hive output. expected %v, got %v", expH, h)
|
||||||
}
|
}
|
||||||
|
@ -20,6 +20,7 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
"github.com/ethereum/go-ethereum/p2p/enode"
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
||||||
@ -33,11 +34,6 @@ import (
|
|||||||
olog "github.com/opentracing/opentracing-go/log"
|
olog "github.com/opentracing/opentracing-go/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
swarmChunkServerStreamName = "RETRIEVE_REQUEST"
|
|
||||||
deliveryCap = 32
|
|
||||||
)
|
|
||||||
|
|
||||||
var (
|
var (
|
||||||
processReceivedChunksCount = metrics.NewRegisteredCounter("network.stream.received_chunks.count", nil)
|
processReceivedChunksCount = metrics.NewRegisteredCounter("network.stream.received_chunks.count", nil)
|
||||||
handleRetrieveRequestMsgCount = metrics.NewRegisteredCounter("network.stream.handle_retrieve_request_msg.count", nil)
|
handleRetrieveRequestMsgCount = metrics.NewRegisteredCounter("network.stream.handle_retrieve_request_msg.count", nil)
|
||||||
@ -45,91 +41,23 @@ var (
|
|||||||
|
|
||||||
requestFromPeersCount = metrics.NewRegisteredCounter("network.stream.request_from_peers.count", nil)
|
requestFromPeersCount = metrics.NewRegisteredCounter("network.stream.request_from_peers.count", nil)
|
||||||
requestFromPeersEachCount = metrics.NewRegisteredCounter("network.stream.request_from_peers_each.count", nil)
|
requestFromPeersEachCount = metrics.NewRegisteredCounter("network.stream.request_from_peers_each.count", nil)
|
||||||
|
|
||||||
|
lastReceivedChunksMsg = metrics.GetOrRegisterGauge("network.stream.received_chunks", nil)
|
||||||
)
|
)
|
||||||
|
|
||||||
type Delivery struct {
|
type Delivery struct {
|
||||||
chunkStore chunk.FetchStore
|
netStore *storage.NetStore
|
||||||
kad *network.Kademlia
|
kad *network.Kademlia
|
||||||
getPeer func(enode.ID) *Peer
|
getPeer func(enode.ID) *Peer
|
||||||
}
|
|
||||||
|
|
||||||
func NewDelivery(kad *network.Kademlia, chunkStore chunk.FetchStore) *Delivery {
|
|
||||||
return &Delivery{
|
|
||||||
chunkStore: chunkStore,
|
|
||||||
kad: kad,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// SwarmChunkServer implements Server
|
|
||||||
type SwarmChunkServer struct {
|
|
||||||
deliveryC chan []byte
|
|
||||||
batchC chan []byte
|
|
||||||
chunkStore storage.ChunkStore
|
|
||||||
currentLen uint64
|
|
||||||
quit chan struct{}
|
quit chan struct{}
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewSwarmChunkServer is SwarmChunkServer constructor
|
func NewDelivery(kad *network.Kademlia, netStore *storage.NetStore) *Delivery {
|
||||||
func NewSwarmChunkServer(chunkStore storage.ChunkStore) *SwarmChunkServer {
|
return &Delivery{
|
||||||
s := &SwarmChunkServer{
|
netStore: netStore,
|
||||||
deliveryC: make(chan []byte, deliveryCap),
|
kad: kad,
|
||||||
batchC: make(chan []byte),
|
|
||||||
chunkStore: chunkStore,
|
|
||||||
quit: make(chan struct{}),
|
quit: make(chan struct{}),
|
||||||
}
|
}
|
||||||
go s.processDeliveries()
|
|
||||||
return s
|
|
||||||
}
|
|
||||||
|
|
||||||
// processDeliveries handles delivered chunk hashes
|
|
||||||
func (s *SwarmChunkServer) processDeliveries() {
|
|
||||||
var hashes []byte
|
|
||||||
var batchC chan []byte
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-s.quit:
|
|
||||||
return
|
|
||||||
case hash := <-s.deliveryC:
|
|
||||||
hashes = append(hashes, hash...)
|
|
||||||
batchC = s.batchC
|
|
||||||
case batchC <- hashes:
|
|
||||||
hashes = nil
|
|
||||||
batchC = nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// SessionIndex returns zero in all cases for SwarmChunkServer.
|
|
||||||
func (s *SwarmChunkServer) SessionIndex() (uint64, error) {
|
|
||||||
return 0, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// SetNextBatch
|
|
||||||
func (s *SwarmChunkServer) SetNextBatch(_, _ uint64) (hashes []byte, from uint64, to uint64, proof *HandoverProof, err error) {
|
|
||||||
select {
|
|
||||||
case hashes = <-s.batchC:
|
|
||||||
case <-s.quit:
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
from = s.currentLen
|
|
||||||
s.currentLen += uint64(len(hashes))
|
|
||||||
to = s.currentLen
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close needs to be called on a stream server
|
|
||||||
func (s *SwarmChunkServer) Close() {
|
|
||||||
close(s.quit)
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetData retrieves chunk data from db store
|
|
||||||
func (s *SwarmChunkServer) GetData(ctx context.Context, key []byte) ([]byte, error) {
|
|
||||||
ch, err := s.chunkStore.Get(ctx, chunk.ModeGetRequest, storage.Address(key))
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
return ch.Data(), nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// RetrieveRequestMsg is the protocol msg for chunk retrieve requests
|
// RetrieveRequestMsg is the protocol msg for chunk retrieve requests
|
||||||
@ -150,12 +78,6 @@ func (d *Delivery) handleRetrieveRequestMsg(ctx context.Context, sp *Peer, req *
|
|||||||
|
|
||||||
osp.LogFields(olog.String("ref", req.Addr.String()))
|
osp.LogFields(olog.String("ref", req.Addr.String()))
|
||||||
|
|
||||||
s, err := sp.getServer(NewStream(swarmChunkServerStreamName, "", true))
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
streamer := s.Server.(*SwarmChunkServer)
|
|
||||||
|
|
||||||
var cancel func()
|
var cancel func()
|
||||||
// TODO: do something with this hardcoded timeout, maybe use TTL in the future
|
// TODO: do something with this hardcoded timeout, maybe use TTL in the future
|
||||||
ctx = context.WithValue(ctx, "peer", sp.ID().String())
|
ctx = context.WithValue(ctx, "peer", sp.ID().String())
|
||||||
@ -165,36 +87,26 @@ func (d *Delivery) handleRetrieveRequestMsg(ctx context.Context, sp *Peer, req *
|
|||||||
go func() {
|
go func() {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
case <-streamer.quit:
|
case <-d.quit:
|
||||||
}
|
}
|
||||||
cancel()
|
cancel()
|
||||||
}()
|
}()
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
defer osp.Finish()
|
defer osp.Finish()
|
||||||
ch, err := d.chunkStore.Get(ctx, chunk.ModeGetRequest, req.Addr)
|
ch, err := d.netStore.Get(ctx, chunk.ModeGetRequest, req.Addr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
retrieveChunkFail.Inc(1)
|
retrieveChunkFail.Inc(1)
|
||||||
log.Debug("ChunkStore.Get can not retrieve chunk", "peer", sp.ID().String(), "addr", req.Addr, "hopcount", req.HopCount, "err", err)
|
log.Debug("ChunkStore.Get can not retrieve chunk", "peer", sp.ID().String(), "addr", req.Addr, "hopcount", req.HopCount, "err", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if req.SkipCheck {
|
|
||||||
syncing := false
|
syncing := false
|
||||||
osp.LogFields(olog.Bool("skipCheck", true))
|
|
||||||
|
|
||||||
err = sp.Deliver(ctx, ch, s.priority, syncing)
|
err = sp.Deliver(ctx, ch, Top, syncing)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Warn("ERROR in handleRetrieveRequestMsg", "err", err)
|
log.Warn("ERROR in handleRetrieveRequestMsg", "err", err)
|
||||||
}
|
}
|
||||||
osp.LogFields(olog.Bool("delivered", true))
|
osp.LogFields(olog.Bool("delivered", true))
|
||||||
return
|
|
||||||
}
|
|
||||||
osp.LogFields(olog.Bool("skipCheck", false))
|
|
||||||
select {
|
|
||||||
case streamer.deliveryC <- ch.Address()[:]:
|
|
||||||
case <-streamer.quit:
|
|
||||||
}
|
|
||||||
|
|
||||||
}()
|
}()
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
@ -225,6 +137,9 @@ func (d *Delivery) handleChunkDeliveryMsg(ctx context.Context, sp *Peer, req int
|
|||||||
|
|
||||||
processReceivedChunksCount.Inc(1)
|
processReceivedChunksCount.Inc(1)
|
||||||
|
|
||||||
|
// record the last time we received a chunk delivery message
|
||||||
|
lastReceivedChunksMsg.Update(time.Now().UnixNano())
|
||||||
|
|
||||||
var msg *ChunkDeliveryMsg
|
var msg *ChunkDeliveryMsg
|
||||||
var mode chunk.ModePut
|
var mode chunk.ModePut
|
||||||
switch r := req.(type) {
|
switch r := req.(type) {
|
||||||
@ -244,31 +159,25 @@ func (d *Delivery) handleChunkDeliveryMsg(ctx context.Context, sp *Peer, req int
|
|||||||
case *ChunkDeliveryMsgSyncing:
|
case *ChunkDeliveryMsgSyncing:
|
||||||
msg = (*ChunkDeliveryMsg)(r)
|
msg = (*ChunkDeliveryMsg)(r)
|
||||||
mode = chunk.ModePutSync
|
mode = chunk.ModePutSync
|
||||||
|
case *ChunkDeliveryMsg:
|
||||||
|
msg = r
|
||||||
|
mode = chunk.ModePutSync
|
||||||
}
|
}
|
||||||
|
|
||||||
// retrieve the span for the originating retrieverequest
|
|
||||||
spanID := fmt.Sprintf("stream.send.request.%v.%v", sp.ID(), msg.Addr)
|
|
||||||
span := tracing.ShiftSpanByKey(spanID)
|
|
||||||
|
|
||||||
log.Trace("handle.chunk.delivery", "ref", msg.Addr, "from peer", sp.ID())
|
log.Trace("handle.chunk.delivery", "ref", msg.Addr, "from peer", sp.ID())
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
defer osp.Finish()
|
defer osp.Finish()
|
||||||
|
|
||||||
if span != nil {
|
|
||||||
span.LogFields(olog.String("finish", "from handleChunkDeliveryMsg"))
|
|
||||||
defer span.Finish()
|
|
||||||
}
|
|
||||||
|
|
||||||
msg.peer = sp
|
msg.peer = sp
|
||||||
log.Trace("handle.chunk.delivery", "put", msg.Addr)
|
log.Trace("handle.chunk.delivery", "put", msg.Addr)
|
||||||
_, err := d.chunkStore.Put(ctx, mode, storage.NewChunk(msg.Addr, msg.SData))
|
_, err := d.netStore.Put(ctx, mode, storage.NewChunk(msg.Addr, msg.SData))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if err == storage.ErrChunkInvalid {
|
if err == storage.ErrChunkInvalid {
|
||||||
// we removed this log because it spams the logs
|
// we removed this log because it spams the logs
|
||||||
// TODO: Enable this log line
|
// TODO: Enable this log line
|
||||||
// log.Warn("invalid chunk delivered", "peer", sp.ID(), "chunk", msg.Addr, )
|
// log.Warn("invalid chunk delivered", "peer", sp.ID(), "chunk", msg.Addr, )
|
||||||
msg.peer.Drop(err)
|
msg.peer.Drop()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
log.Trace("handle.chunk.delivery", "done put", msg.Addr, "err", err)
|
log.Trace("handle.chunk.delivery", "done put", msg.Addr, "err", err)
|
||||||
@ -276,6 +185,12 @@ func (d *Delivery) handleChunkDeliveryMsg(ctx context.Context, sp *Peer, req int
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (d *Delivery) Close() {
|
||||||
|
d.kad.CloseNeighbourhoodDepthC()
|
||||||
|
d.kad.CloseAddrCountC()
|
||||||
|
close(d.quit)
|
||||||
|
}
|
||||||
|
|
||||||
// RequestFromPeers sends a chunk retrieve request to a peer
|
// RequestFromPeers sends a chunk retrieve request to a peer
|
||||||
// The most eligible peer that hasn't already been sent to is chosen
|
// The most eligible peer that hasn't already been sent to is chosen
|
||||||
// TODO: define "eligible"
|
// TODO: define "eligible"
|
||||||
|
@ -41,63 +41,10 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/swarm/testutil"
|
"github.com/ethereum/go-ethereum/swarm/testutil"
|
||||||
)
|
)
|
||||||
|
|
||||||
//Tests initializing a retrieve request
|
|
||||||
func TestStreamerRetrieveRequest(t *testing.T) {
|
|
||||||
regOpts := &RegistryOptions{
|
|
||||||
Retrieval: RetrievalClientOnly,
|
|
||||||
Syncing: SyncingDisabled,
|
|
||||||
}
|
|
||||||
tester, streamer, _, teardown, err := newStreamerTester(regOpts)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
defer teardown()
|
|
||||||
|
|
||||||
node := tester.Nodes[0]
|
|
||||||
|
|
||||||
ctx := context.Background()
|
|
||||||
req := network.NewRequest(
|
|
||||||
storage.Address(hash0[:]),
|
|
||||||
true,
|
|
||||||
&sync.Map{},
|
|
||||||
)
|
|
||||||
streamer.delivery.RequestFromPeers(ctx, req)
|
|
||||||
|
|
||||||
stream := NewStream(swarmChunkServerStreamName, "", true)
|
|
||||||
|
|
||||||
err = tester.TestExchanges(p2ptest.Exchange{
|
|
||||||
Label: "RetrieveRequestMsg",
|
|
||||||
Expects: []p2ptest.Expect{
|
|
||||||
{ //start expecting a subscription for RETRIEVE_REQUEST due to `RetrievalClientOnly`
|
|
||||||
Code: 4,
|
|
||||||
Msg: &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
History: nil,
|
|
||||||
Priority: Top,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
{ //expect a retrieve request message for the given hash
|
|
||||||
Code: 5,
|
|
||||||
Msg: &RetrieveRequestMsg{
|
|
||||||
Addr: hash0[:],
|
|
||||||
SkipCheck: true,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Expected no error, got %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//Test requesting a chunk from a peer then issuing a "empty" OfferedHashesMsg (no hashes available yet)
|
//Test requesting a chunk from a peer then issuing a "empty" OfferedHashesMsg (no hashes available yet)
|
||||||
//Should time out as the peer does not have the chunk (no syncing happened previously)
|
//Should time out as the peer does not have the chunk (no syncing happened previously)
|
||||||
func TestStreamerUpstreamRetrieveRequestMsgExchangeWithoutStore(t *testing.T) {
|
func TestStreamerUpstreamRetrieveRequestMsgExchangeWithoutStore(t *testing.T) {
|
||||||
tester, streamer, _, teardown, err := newStreamerTester(&RegistryOptions{
|
tester, _, _, teardown, err := newStreamerTester(&RegistryOptions{
|
||||||
Retrieval: RetrievalEnabled,
|
|
||||||
Syncing: SyncingDisabled, //do no syncing
|
Syncing: SyncingDisabled, //do no syncing
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -109,30 +56,8 @@ func TestStreamerUpstreamRetrieveRequestMsgExchangeWithoutStore(t *testing.T) {
|
|||||||
|
|
||||||
chunk := storage.NewChunk(storage.Address(hash0[:]), nil)
|
chunk := storage.NewChunk(storage.Address(hash0[:]), nil)
|
||||||
|
|
||||||
peer := streamer.getPeer(node.ID())
|
|
||||||
|
|
||||||
stream := NewStream(swarmChunkServerStreamName, "", true)
|
|
||||||
//simulate pre-subscription to RETRIEVE_REQUEST stream on peer
|
|
||||||
peer.handleSubscribeMsg(context.TODO(), &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
History: nil,
|
|
||||||
Priority: Top,
|
|
||||||
})
|
|
||||||
|
|
||||||
//test the exchange
|
//test the exchange
|
||||||
err = tester.TestExchanges(p2ptest.Exchange{
|
err = tester.TestExchanges(p2ptest.Exchange{
|
||||||
Expects: []p2ptest.Expect{
|
|
||||||
{ //first expect a subscription to the RETRIEVE_REQUEST stream
|
|
||||||
Code: 4,
|
|
||||||
Msg: &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
History: nil,
|
|
||||||
Priority: Top,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}, p2ptest.Exchange{
|
|
||||||
Label: "RetrieveRequestMsg",
|
Label: "RetrieveRequestMsg",
|
||||||
Triggers: []p2ptest.Trigger{
|
Triggers: []p2ptest.Trigger{
|
||||||
{ //then the actual RETRIEVE_REQUEST....
|
{ //then the actual RETRIEVE_REQUEST....
|
||||||
@ -159,7 +84,7 @@ func TestStreamerUpstreamRetrieveRequestMsgExchangeWithoutStore(t *testing.T) {
|
|||||||
|
|
||||||
//should fail with a timeout as the peer we are requesting
|
//should fail with a timeout as the peer we are requesting
|
||||||
//the chunk from does not have the chunk
|
//the chunk from does not have the chunk
|
||||||
expectedError := `exchange #1 "RetrieveRequestMsg": timed out`
|
expectedError := `exchange #0 "RetrieveRequestMsg": timed out`
|
||||||
if err == nil || err.Error() != expectedError {
|
if err == nil || err.Error() != expectedError {
|
||||||
t.Fatalf("Expected error %v, got %v", expectedError, err)
|
t.Fatalf("Expected error %v, got %v", expectedError, err)
|
||||||
}
|
}
|
||||||
@ -168,8 +93,7 @@ func TestStreamerUpstreamRetrieveRequestMsgExchangeWithoutStore(t *testing.T) {
|
|||||||
// upstream request server receives a retrieve Request and responds with
|
// upstream request server receives a retrieve Request and responds with
|
||||||
// offered hashes or delivery if skipHash is set to true
|
// offered hashes or delivery if skipHash is set to true
|
||||||
func TestStreamerUpstreamRetrieveRequestMsgExchange(t *testing.T) {
|
func TestStreamerUpstreamRetrieveRequestMsgExchange(t *testing.T) {
|
||||||
tester, streamer, localStore, teardown, err := newStreamerTester(&RegistryOptions{
|
tester, _, localStore, teardown, err := newStreamerTester(&RegistryOptions{
|
||||||
Retrieval: RetrievalEnabled,
|
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -179,70 +103,8 @@ func TestStreamerUpstreamRetrieveRequestMsgExchange(t *testing.T) {
|
|||||||
|
|
||||||
node := tester.Nodes[0]
|
node := tester.Nodes[0]
|
||||||
|
|
||||||
peer := streamer.getPeer(node.ID())
|
hash := storage.Address(hash1[:])
|
||||||
|
ch := storage.NewChunk(hash, hash1[:])
|
||||||
stream := NewStream(swarmChunkServerStreamName, "", true)
|
|
||||||
|
|
||||||
peer.handleSubscribeMsg(context.TODO(), &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
History: nil,
|
|
||||||
Priority: Top,
|
|
||||||
})
|
|
||||||
|
|
||||||
hash := storage.Address(hash0[:])
|
|
||||||
ch := storage.NewChunk(hash, hash)
|
|
||||||
_, err = localStore.Put(context.TODO(), chunk.ModePutUpload, ch)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Expected no err got %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
err = tester.TestExchanges(p2ptest.Exchange{
|
|
||||||
Expects: []p2ptest.Expect{
|
|
||||||
{
|
|
||||||
Code: 4,
|
|
||||||
Msg: &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
History: nil,
|
|
||||||
Priority: Top,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}, p2ptest.Exchange{
|
|
||||||
Label: "RetrieveRequestMsg",
|
|
||||||
Triggers: []p2ptest.Trigger{
|
|
||||||
{
|
|
||||||
Code: 5,
|
|
||||||
Msg: &RetrieveRequestMsg{
|
|
||||||
Addr: hash,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Expects: []p2ptest.Expect{
|
|
||||||
{
|
|
||||||
Code: 1,
|
|
||||||
Msg: &OfferedHashesMsg{
|
|
||||||
HandoverProof: &HandoverProof{
|
|
||||||
Handover: &Handover{},
|
|
||||||
},
|
|
||||||
Hashes: hash,
|
|
||||||
From: 0,
|
|
||||||
// TODO: why is this 32???
|
|
||||||
To: 32,
|
|
||||||
Stream: stream,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
hash = storage.Address(hash1[:])
|
|
||||||
ch = storage.NewChunk(hash, hash1[:])
|
|
||||||
_, err = localStore.Put(context.TODO(), chunk.ModePutUpload, ch)
|
_, err = localStore.Put(context.TODO(), chunk.ModePutUpload, ch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Expected no err got %v", err)
|
t.Fatalf("Expected no err got %v", err)
|
||||||
@ -255,7 +117,6 @@ func TestStreamerUpstreamRetrieveRequestMsgExchange(t *testing.T) {
|
|||||||
Code: 5,
|
Code: 5,
|
||||||
Msg: &RetrieveRequestMsg{
|
Msg: &RetrieveRequestMsg{
|
||||||
Addr: hash,
|
Addr: hash,
|
||||||
SkipCheck: true,
|
|
||||||
},
|
},
|
||||||
Peer: node.ID(),
|
Peer: node.ID(),
|
||||||
},
|
},
|
||||||
@ -264,8 +125,8 @@ func TestStreamerUpstreamRetrieveRequestMsgExchange(t *testing.T) {
|
|||||||
{
|
{
|
||||||
Code: 6,
|
Code: 6,
|
||||||
Msg: &ChunkDeliveryMsg{
|
Msg: &ChunkDeliveryMsg{
|
||||||
Addr: hash,
|
Addr: ch.Address(),
|
||||||
SData: hash,
|
SData: ch.Data(),
|
||||||
},
|
},
|
||||||
Peer: node.ID(),
|
Peer: node.ID(),
|
||||||
},
|
},
|
||||||
@ -359,7 +220,6 @@ func TestRequestFromPeersWithLightNode(t *testing.T) {
|
|||||||
|
|
||||||
func TestStreamerDownstreamChunkDeliveryMsgExchange(t *testing.T) {
|
func TestStreamerDownstreamChunkDeliveryMsgExchange(t *testing.T) {
|
||||||
tester, streamer, localStore, teardown, err := newStreamerTester(&RegistryOptions{
|
tester, streamer, localStore, teardown, err := newStreamerTester(&RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -472,7 +332,6 @@ func testDeliveryFromNodes(t *testing.T, nodes, chunkCount int, skipCheck bool)
|
|||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
SkipCheck: skipCheck,
|
SkipCheck: skipCheck,
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
Retrieval: RetrievalEnabled,
|
|
||||||
}, nil)
|
}, nil)
|
||||||
bucket.Store(bucketKeyRegistry, r)
|
bucket.Store(bucketKeyRegistry, r)
|
||||||
|
|
||||||
@ -623,7 +482,6 @@ func benchmarkDeliveryFromNodes(b *testing.B, nodes, chunkCount int, skipCheck b
|
|||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
SkipCheck: skipCheck,
|
SkipCheck: skipCheck,
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
SyncUpdateDelay: 0,
|
SyncUpdateDelay: 0,
|
||||||
}, nil)
|
}, nil)
|
||||||
bucket.Store(bucketKeyRegistry, r)
|
bucket.Store(bucketKeyRegistry, r)
|
||||||
|
@ -29,7 +29,6 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/node"
|
"github.com/ethereum/go-ethereum/node"
|
||||||
"github.com/ethereum/go-ethereum/p2p/enode"
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
||||||
"github.com/ethereum/go-ethereum/p2p/simulations/adapters"
|
"github.com/ethereum/go-ethereum/p2p/simulations/adapters"
|
||||||
"github.com/ethereum/go-ethereum/swarm/chunk"
|
|
||||||
"github.com/ethereum/go-ethereum/swarm/network/simulation"
|
"github.com/ethereum/go-ethereum/swarm/network/simulation"
|
||||||
"github.com/ethereum/go-ethereum/swarm/state"
|
"github.com/ethereum/go-ethereum/swarm/state"
|
||||||
"github.com/ethereum/go-ethereum/swarm/storage"
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
@ -67,7 +66,6 @@ func testIntervals(t *testing.T, live bool, history *Range, skipCheck bool) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingRegisterOnly,
|
Syncing: SyncingRegisterOnly,
|
||||||
SkipCheck: skipCheck,
|
SkipCheck: skipCheck,
|
||||||
}, nil)
|
}, nil)
|
||||||
@ -288,20 +286,20 @@ func enableNotifications(r *Registry, peerID enode.ID, s Stream) error {
|
|||||||
|
|
||||||
type testExternalClient struct {
|
type testExternalClient struct {
|
||||||
hashes chan []byte
|
hashes chan []byte
|
||||||
store chunk.FetchStore
|
netStore *storage.NetStore
|
||||||
enableNotificationsC chan struct{}
|
enableNotificationsC chan struct{}
|
||||||
}
|
}
|
||||||
|
|
||||||
func newTestExternalClient(store chunk.FetchStore) *testExternalClient {
|
func newTestExternalClient(netStore *storage.NetStore) *testExternalClient {
|
||||||
return &testExternalClient{
|
return &testExternalClient{
|
||||||
hashes: make(chan []byte),
|
hashes: make(chan []byte),
|
||||||
store: store,
|
netStore: netStore,
|
||||||
enableNotificationsC: make(chan struct{}),
|
enableNotificationsC: make(chan struct{}),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *testExternalClient) NeedData(ctx context.Context, hash []byte) func(context.Context) error {
|
func (c *testExternalClient) NeedData(ctx context.Context, hash []byte) func(context.Context) error {
|
||||||
wait := c.store.FetchFunc(ctx, storage.Address(hash))
|
wait := c.netStore.FetchFunc(ctx, storage.Address(hash))
|
||||||
if wait == nil {
|
if wait == nil {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -21,94 +21,10 @@ import (
|
|||||||
p2ptest "github.com/ethereum/go-ethereum/p2p/testing"
|
p2ptest "github.com/ethereum/go-ethereum/p2p/testing"
|
||||||
)
|
)
|
||||||
|
|
||||||
// This test checks the default behavior of the server, that is
|
|
||||||
// when it is serving Retrieve requests.
|
|
||||||
func TestLigthnodeRetrieveRequestWithRetrieve(t *testing.T) {
|
|
||||||
registryOptions := &RegistryOptions{
|
|
||||||
Retrieval: RetrievalClientOnly,
|
|
||||||
Syncing: SyncingDisabled,
|
|
||||||
}
|
|
||||||
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
defer teardown()
|
|
||||||
|
|
||||||
node := tester.Nodes[0]
|
|
||||||
|
|
||||||
stream := NewStream(swarmChunkServerStreamName, "", false)
|
|
||||||
|
|
||||||
err = tester.TestExchanges(p2ptest.Exchange{
|
|
||||||
Label: "SubscribeMsg",
|
|
||||||
Triggers: []p2ptest.Trigger{
|
|
||||||
{
|
|
||||||
Code: 4,
|
|
||||||
Msg: &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Got %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
err = tester.TestDisconnected(&p2ptest.Disconnect{Peer: node.ID()})
|
|
||||||
if err == nil || err.Error() != "timed out waiting for peers to disconnect" {
|
|
||||||
t.Fatalf("Expected no disconnect, got %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// This test checks the Lightnode behavior of server, when serving Retrieve
|
|
||||||
// requests are disabled
|
|
||||||
func TestLigthnodeRetrieveRequestWithoutRetrieve(t *testing.T) {
|
|
||||||
registryOptions := &RegistryOptions{
|
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingDisabled,
|
|
||||||
}
|
|
||||||
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
defer teardown()
|
|
||||||
|
|
||||||
node := tester.Nodes[0]
|
|
||||||
|
|
||||||
stream := NewStream(swarmChunkServerStreamName, "", false)
|
|
||||||
|
|
||||||
err = tester.TestExchanges(
|
|
||||||
p2ptest.Exchange{
|
|
||||||
Label: "SubscribeMsg",
|
|
||||||
Triggers: []p2ptest.Trigger{
|
|
||||||
{
|
|
||||||
Code: 4,
|
|
||||||
Msg: &SubscribeMsg{
|
|
||||||
Stream: stream,
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
Expects: []p2ptest.Expect{
|
|
||||||
{
|
|
||||||
Code: 7,
|
|
||||||
Msg: &SubscribeErrorMsg{
|
|
||||||
Error: "stream RETRIEVE_REQUEST not registered",
|
|
||||||
},
|
|
||||||
Peer: node.ID(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Got %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// This test checks the default behavior of the server, that is
|
// This test checks the default behavior of the server, that is
|
||||||
// when syncing is enabled.
|
// when syncing is enabled.
|
||||||
func TestLigthnodeRequestSubscriptionWithSync(t *testing.T) {
|
func TestLigthnodeRequestSubscriptionWithSync(t *testing.T) {
|
||||||
registryOptions := &RegistryOptions{
|
registryOptions := &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingRegisterOnly,
|
Syncing: SyncingRegisterOnly,
|
||||||
}
|
}
|
||||||
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
||||||
@ -153,7 +69,6 @@ func TestLigthnodeRequestSubscriptionWithSync(t *testing.T) {
|
|||||||
// when syncing is disabled.
|
// when syncing is disabled.
|
||||||
func TestLigthnodeRequestSubscriptionWithoutSync(t *testing.T) {
|
func TestLigthnodeRequestSubscriptionWithoutSync(t *testing.T) {
|
||||||
registryOptions := &RegistryOptions{
|
registryOptions := &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
}
|
}
|
||||||
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
tester, _, _, teardown, err := newStreamerTester(registryOptions)
|
||||||
|
@ -247,7 +247,7 @@ func (p *Peer) handleOfferedHashesMsg(ctx context.Context, req *OfferedHashesMsg
|
|||||||
case err := <-errC:
|
case err := <-errC:
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Debug("client.handleOfferedHashesMsg() error waiting for chunk, dropping peer", "peer", p.ID(), "err", err)
|
log.Debug("client.handleOfferedHashesMsg() error waiting for chunk, dropping peer", "peer", p.ID(), "err", err)
|
||||||
p.Drop(err)
|
p.Drop()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
@ -289,7 +289,7 @@ func (p *Peer) handleOfferedHashesMsg(ctx context.Context, req *OfferedHashesMsg
|
|||||||
case err := <-c.next:
|
case err := <-c.next:
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Warn("c.next error dropping peer", "err", err)
|
log.Warn("c.next error dropping peer", "err", err)
|
||||||
p.Drop(err)
|
p.Drop()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
case <-c.quit:
|
case <-c.quit:
|
||||||
|
@ -90,7 +90,7 @@ func NewPeer(peer *protocols.Peer, streamer *Registry) *Peer {
|
|||||||
err := p.Send(wmsg.Context, wmsg.Msg)
|
err := p.Send(wmsg.Context, wmsg.Msg)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("Message send error, dropping peer", "peer", p.ID(), "err", err)
|
log.Error("Message send error, dropping peer", "peer", p.ID(), "err", err)
|
||||||
p.Drop(err)
|
p.Drop()
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -119,7 +119,6 @@ var retrievalSimServiceMap = map[string]simulation.ServiceFunc{
|
|||||||
}
|
}
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
Retrieval: RetrievalEnabled,
|
|
||||||
Syncing: SyncingAutoSubscribe,
|
Syncing: SyncingAutoSubscribe,
|
||||||
SyncUpdateDelay: syncUpdateDelay,
|
SyncUpdateDelay: syncUpdateDelay,
|
||||||
}, nil)
|
}, nil)
|
||||||
|
@ -118,7 +118,6 @@ var simServiceMap = map[string]simulation.ServiceFunc{
|
|||||||
store := state.NewInmemoryStore()
|
store := state.NewInmemoryStore()
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, store, &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, store, &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingAutoSubscribe,
|
Syncing: SyncingAutoSubscribe,
|
||||||
SyncUpdateDelay: 3 * time.Second,
|
SyncUpdateDelay: 3 * time.Second,
|
||||||
}, nil)
|
}, nil)
|
||||||
|
@ -18,7 +18,6 @@ package stream
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"math"
|
"math"
|
||||||
"reflect"
|
"reflect"
|
||||||
@ -30,11 +29,11 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/p2p/enode"
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
||||||
"github.com/ethereum/go-ethereum/p2p/protocols"
|
"github.com/ethereum/go-ethereum/p2p/protocols"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
"github.com/ethereum/go-ethereum/swarm/chunk"
|
|
||||||
"github.com/ethereum/go-ethereum/swarm/log"
|
"github.com/ethereum/go-ethereum/swarm/log"
|
||||||
"github.com/ethereum/go-ethereum/swarm/network"
|
"github.com/ethereum/go-ethereum/swarm/network"
|
||||||
"github.com/ethereum/go-ethereum/swarm/network/stream/intervals"
|
"github.com/ethereum/go-ethereum/swarm/network/stream/intervals"
|
||||||
"github.com/ethereum/go-ethereum/swarm/state"
|
"github.com/ethereum/go-ethereum/swarm/state"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -49,7 +48,6 @@ const (
|
|||||||
|
|
||||||
// Enumerate options for syncing and retrieval
|
// Enumerate options for syncing and retrieval
|
||||||
type SyncingOption int
|
type SyncingOption int
|
||||||
type RetrievalOption int
|
|
||||||
|
|
||||||
// Syncing options
|
// Syncing options
|
||||||
const (
|
const (
|
||||||
@ -61,17 +59,6 @@ const (
|
|||||||
SyncingAutoSubscribe
|
SyncingAutoSubscribe
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
// Retrieval disabled. Used mostly for tests to isolate syncing features (i.e. syncing only)
|
|
||||||
RetrievalDisabled RetrievalOption = iota
|
|
||||||
// Only the client side of the retrieve request is registered.
|
|
||||||
// (light nodes do not serve retrieve requests)
|
|
||||||
// once the client is registered, subscription to retrieve request stream is always sent
|
|
||||||
RetrievalClientOnly
|
|
||||||
// Both client and server funcs are registered, subscribe sent automatically
|
|
||||||
RetrievalEnabled
|
|
||||||
)
|
|
||||||
|
|
||||||
// subscriptionFunc is used to determine what to do in order to perform subscriptions
|
// subscriptionFunc is used to determine what to do in order to perform subscriptions
|
||||||
// usually we would start to really subscribe to nodes, but for tests other functionality may be needed
|
// usually we would start to really subscribe to nodes, but for tests other functionality may be needed
|
||||||
// (see TestRequestPeerSubscriptions in streamer_test.go)
|
// (see TestRequestPeerSubscriptions in streamer_test.go)
|
||||||
@ -90,7 +77,6 @@ type Registry struct {
|
|||||||
peers map[enode.ID]*Peer
|
peers map[enode.ID]*Peer
|
||||||
delivery *Delivery
|
delivery *Delivery
|
||||||
intervalsStore state.Store
|
intervalsStore state.Store
|
||||||
autoRetrieval bool // automatically subscribe to retrieve request stream
|
|
||||||
maxPeerServers int
|
maxPeerServers int
|
||||||
spec *protocols.Spec //this protocol's spec
|
spec *protocols.Spec //this protocol's spec
|
||||||
balance protocols.Balance //implements protocols.Balance, for accounting
|
balance protocols.Balance //implements protocols.Balance, for accounting
|
||||||
@ -102,21 +88,18 @@ type Registry struct {
|
|||||||
type RegistryOptions struct {
|
type RegistryOptions struct {
|
||||||
SkipCheck bool
|
SkipCheck bool
|
||||||
Syncing SyncingOption // Defines syncing behavior
|
Syncing SyncingOption // Defines syncing behavior
|
||||||
Retrieval RetrievalOption // Defines retrieval behavior
|
|
||||||
SyncUpdateDelay time.Duration
|
SyncUpdateDelay time.Duration
|
||||||
MaxPeerServers int // The limit of servers for each peer in registry
|
MaxPeerServers int // The limit of servers for each peer in registry
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewRegistry is Streamer constructor
|
// NewRegistry is Streamer constructor
|
||||||
func NewRegistry(localID enode.ID, delivery *Delivery, syncChunkStore chunk.FetchStore, intervalsStore state.Store, options *RegistryOptions, balance protocols.Balance) *Registry {
|
func NewRegistry(localID enode.ID, delivery *Delivery, netStore *storage.NetStore, intervalsStore state.Store, options *RegistryOptions, balance protocols.Balance) *Registry {
|
||||||
if options == nil {
|
if options == nil {
|
||||||
options = &RegistryOptions{}
|
options = &RegistryOptions{}
|
||||||
}
|
}
|
||||||
if options.SyncUpdateDelay <= 0 {
|
if options.SyncUpdateDelay <= 0 {
|
||||||
options.SyncUpdateDelay = 15 * time.Second
|
options.SyncUpdateDelay = 15 * time.Second
|
||||||
}
|
}
|
||||||
// check if retrieval has been disabled
|
|
||||||
retrieval := options.Retrieval != RetrievalDisabled
|
|
||||||
|
|
||||||
quit := make(chan struct{})
|
quit := make(chan struct{})
|
||||||
|
|
||||||
@ -128,7 +111,6 @@ func NewRegistry(localID enode.ID, delivery *Delivery, syncChunkStore chunk.Fetc
|
|||||||
peers: make(map[enode.ID]*Peer),
|
peers: make(map[enode.ID]*Peer),
|
||||||
delivery: delivery,
|
delivery: delivery,
|
||||||
intervalsStore: intervalsStore,
|
intervalsStore: intervalsStore,
|
||||||
autoRetrieval: retrieval,
|
|
||||||
maxPeerServers: options.MaxPeerServers,
|
maxPeerServers: options.MaxPeerServers,
|
||||||
balance: balance,
|
balance: balance,
|
||||||
quit: quit,
|
quit: quit,
|
||||||
@ -139,27 +121,10 @@ func NewRegistry(localID enode.ID, delivery *Delivery, syncChunkStore chunk.Fetc
|
|||||||
streamer.api = NewAPI(streamer)
|
streamer.api = NewAPI(streamer)
|
||||||
delivery.getPeer = streamer.getPeer
|
delivery.getPeer = streamer.getPeer
|
||||||
|
|
||||||
// if retrieval is enabled, register the server func, so that retrieve requests will be served (non-light nodes only)
|
|
||||||
if options.Retrieval == RetrievalEnabled {
|
|
||||||
streamer.RegisterServerFunc(swarmChunkServerStreamName, func(_ *Peer, _ string, live bool) (Server, error) {
|
|
||||||
if !live {
|
|
||||||
return nil, errors.New("only live retrieval requests supported")
|
|
||||||
}
|
|
||||||
return NewSwarmChunkServer(delivery.chunkStore), nil
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// if retrieval is not disabled, register the client func (both light nodes and normal nodes can issue retrieve requests)
|
|
||||||
if options.Retrieval != RetrievalDisabled {
|
|
||||||
streamer.RegisterClientFunc(swarmChunkServerStreamName, func(p *Peer, t string, live bool) (Client, error) {
|
|
||||||
return NewSwarmSyncerClient(p, syncChunkStore, NewStream(swarmChunkServerStreamName, t, live))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// If syncing is not disabled, the syncing functions are registered (both client and server)
|
// If syncing is not disabled, the syncing functions are registered (both client and server)
|
||||||
if options.Syncing != SyncingDisabled {
|
if options.Syncing != SyncingDisabled {
|
||||||
RegisterSwarmSyncerServer(streamer, syncChunkStore)
|
RegisterSwarmSyncerServer(streamer, netStore)
|
||||||
RegisterSwarmSyncerClient(streamer, syncChunkStore)
|
RegisterSwarmSyncerClient(streamer, netStore)
|
||||||
}
|
}
|
||||||
|
|
||||||
// if syncing is set to automatically subscribe to the syncing stream, start the subscription process
|
// if syncing is set to automatically subscribe to the syncing stream, start the subscription process
|
||||||
@ -381,7 +346,7 @@ func (r *Registry) Subscribe(peerId enode.ID, s Stream, h *Range, priority uint8
|
|||||||
}
|
}
|
||||||
log.Debug("Subscribe ", "peer", peerId, "stream", s, "history", h)
|
log.Debug("Subscribe ", "peer", peerId, "stream", s, "history", h)
|
||||||
|
|
||||||
return peer.SendPriority(context.TODO(), msg, priority)
|
return peer.Send(context.TODO(), msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *Registry) Unsubscribe(peerId enode.ID, s Stream) error {
|
func (r *Registry) Unsubscribe(peerId enode.ID, s Stream) error {
|
||||||
@ -422,8 +387,7 @@ func (r *Registry) Quit(peerId enode.ID, s Stream) error {
|
|||||||
func (r *Registry) Close() error {
|
func (r *Registry) Close() error {
|
||||||
// Stop sending neighborhood depth change and address count
|
// Stop sending neighborhood depth change and address count
|
||||||
// change from Kademlia that were initiated in NewRegistry constructor.
|
// change from Kademlia that were initiated in NewRegistry constructor.
|
||||||
r.delivery.kad.CloseNeighbourhoodDepthC()
|
r.delivery.Close()
|
||||||
r.delivery.kad.CloseAddrCountC()
|
|
||||||
close(r.quit)
|
close(r.quit)
|
||||||
return r.intervalsStore.Close()
|
return r.intervalsStore.Close()
|
||||||
}
|
}
|
||||||
@ -464,13 +428,6 @@ func (r *Registry) Run(p *network.BzzPeer) error {
|
|||||||
defer close(sp.quit)
|
defer close(sp.quit)
|
||||||
defer sp.close()
|
defer sp.close()
|
||||||
|
|
||||||
if r.autoRetrieval && !p.LightNode {
|
|
||||||
err := r.Subscribe(p.ID(), NewStream(swarmChunkServerStreamName, "", true), nil, Top)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return sp.Run(sp.HandleMsg)
|
return sp.Run(sp.HandleMsg)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -619,19 +576,66 @@ func (p *Peer) HandleMsg(ctx context.Context, msg interface{}) error {
|
|||||||
return p.handleUnsubscribeMsg(msg)
|
return p.handleUnsubscribeMsg(msg)
|
||||||
|
|
||||||
case *OfferedHashesMsg:
|
case *OfferedHashesMsg:
|
||||||
return p.handleOfferedHashesMsg(ctx, msg)
|
go func() {
|
||||||
|
err := p.handleOfferedHashesMsg(ctx, msg)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
case *TakeoverProofMsg:
|
case *TakeoverProofMsg:
|
||||||
return p.handleTakeoverProofMsg(ctx, msg)
|
go func() {
|
||||||
|
err := p.handleTakeoverProofMsg(ctx, msg)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
case *WantedHashesMsg:
|
case *WantedHashesMsg:
|
||||||
return p.handleWantedHashesMsg(ctx, msg)
|
go func() {
|
||||||
|
err := p.handleWantedHashesMsg(ctx, msg)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
case *ChunkDeliveryMsgRetrieval, *ChunkDeliveryMsgSyncing:
|
case *ChunkDeliveryMsgRetrieval:
|
||||||
return p.streamer.delivery.handleChunkDeliveryMsg(ctx, p, msg)
|
// handling chunk delivery is the same for retrieval and syncing, so let's cast the msg
|
||||||
|
go func() {
|
||||||
|
err := p.streamer.delivery.handleChunkDeliveryMsg(ctx, p, ((*ChunkDeliveryMsg)(msg)))
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
|
case *ChunkDeliveryMsgSyncing:
|
||||||
|
// handling chunk delivery is the same for retrieval and syncing, so let's cast the msg
|
||||||
|
go func() {
|
||||||
|
err := p.streamer.delivery.handleChunkDeliveryMsg(ctx, p, ((*ChunkDeliveryMsg)(msg)))
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
case *RetrieveRequestMsg:
|
case *RetrieveRequestMsg:
|
||||||
return p.streamer.delivery.handleRetrieveRequestMsg(ctx, p, msg)
|
go func() {
|
||||||
|
err := p.streamer.delivery.handleRetrieveRequestMsg(ctx, p, msg)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
p.Drop()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
|
||||||
case *RequestSubscriptionMsg:
|
case *RequestSubscriptionMsg:
|
||||||
return p.handleRequestSubscription(ctx, msg)
|
return p.handleRequestSubscription(ctx, msg)
|
||||||
@ -762,7 +766,7 @@ func (c *client) batchDone(p *Peer, req *OfferedHashesMsg, hashes []byte) error
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := p.SendPriority(context.TODO(), tp, c.priority); err != nil {
|
if err := p.Send(context.TODO(), tp); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if c.to > 0 && tp.Takeover.End >= c.to {
|
if c.to > 0 && tp.Takeover.End >= c.to {
|
||||||
@ -964,15 +968,13 @@ func (api *API) UnsubscribeStream(peerId enode.ID, s Stream) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
GetPeerSubscriptions is a API function which allows to query a peer for stream subscriptions it has.
|
GetPeerServerSubscriptions is a API function which allows to query a peer for stream subscriptions it has.
|
||||||
It can be called via RPC.
|
It can be called via RPC.
|
||||||
It returns a map of node IDs with an array of string representations of Stream objects.
|
It returns a map of node IDs with an array of string representations of Stream objects.
|
||||||
*/
|
*/
|
||||||
func (api *API) GetPeerSubscriptions() map[string][]string {
|
func (api *API) GetPeerServerSubscriptions() map[string][]string {
|
||||||
//create the empty map
|
|
||||||
pstreams := make(map[string][]string)
|
pstreams := make(map[string][]string)
|
||||||
|
|
||||||
//iterate all streamer peers
|
|
||||||
api.streamer.peersMu.RLock()
|
api.streamer.peersMu.RLock()
|
||||||
defer api.streamer.peersMu.RUnlock()
|
defer api.streamer.peersMu.RUnlock()
|
||||||
|
|
||||||
|
@ -539,7 +539,7 @@ func TestStreamerDownstreamCorruptHashesMsgExchange(t *testing.T) {
|
|||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
expectedError := errors.New("Message handler error: (msg code 1): error invalid hashes length (len: 40)")
|
expectedError := errors.New("subprotocol error")
|
||||||
if err := tester.TestDisconnected(&p2ptest.Disconnect{Peer: node.ID(), Error: expectedError}); err != nil {
|
if err := tester.TestDisconnected(&p2ptest.Disconnect{Peer: node.ID(), Error: expectedError}); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
@ -779,7 +779,6 @@ func TestStreamerRequestSubscriptionQuitMsgExchange(t *testing.T) {
|
|||||||
func TestMaxPeerServersWithUnsubscribe(t *testing.T) {
|
func TestMaxPeerServersWithUnsubscribe(t *testing.T) {
|
||||||
var maxPeerServers = 6
|
var maxPeerServers = 6
|
||||||
tester, streamer, _, teardown, err := newStreamerTester(&RegistryOptions{
|
tester, streamer, _, teardown, err := newStreamerTester(&RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
MaxPeerServers: maxPeerServers,
|
MaxPeerServers: maxPeerServers,
|
||||||
})
|
})
|
||||||
@ -940,7 +939,6 @@ func TestMaxPeerServersWithoutUnsubscribe(t *testing.T) {
|
|||||||
//`Price` interface implementation
|
//`Price` interface implementation
|
||||||
func TestHasPriceImplementation(t *testing.T) {
|
func TestHasPriceImplementation(t *testing.T) {
|
||||||
_, r, _, teardown, err := newStreamerTester(&RegistryOptions{
|
_, r, _, teardown, err := newStreamerTester(&RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingDisabled,
|
Syncing: SyncingDisabled,
|
||||||
})
|
})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -1123,8 +1121,8 @@ func TestRequestPeerSubscriptions(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestGetSubscriptions is a unit test for the api.GetPeerSubscriptions() function
|
// TestGetServerSubscriptions is a unit test for the api.GetPeerServerSubscriptions() function
|
||||||
func TestGetSubscriptions(t *testing.T) {
|
func TestGetServerSubscriptions(t *testing.T) {
|
||||||
// create an amount of dummy peers
|
// create an amount of dummy peers
|
||||||
testPeerCount := 8
|
testPeerCount := 8
|
||||||
// every peer will have this amount of dummy servers
|
// every peer will have this amount of dummy servers
|
||||||
@ -1135,7 +1133,7 @@ func TestGetSubscriptions(t *testing.T) {
|
|||||||
r := &Registry{}
|
r := &Registry{}
|
||||||
api := NewAPI(r)
|
api := NewAPI(r)
|
||||||
// call once, at this point should be empty
|
// call once, at this point should be empty
|
||||||
regs := api.GetPeerSubscriptions()
|
regs := api.GetPeerServerSubscriptions()
|
||||||
if len(regs) != 0 {
|
if len(regs) != 0 {
|
||||||
t.Fatal("Expected subscription count to be 0, but it is not")
|
t.Fatal("Expected subscription count to be 0, but it is not")
|
||||||
}
|
}
|
||||||
@ -1159,7 +1157,7 @@ func TestGetSubscriptions(t *testing.T) {
|
|||||||
r.peers = peerMap
|
r.peers = peerMap
|
||||||
|
|
||||||
// call the subscriptions again
|
// call the subscriptions again
|
||||||
regs = api.GetPeerSubscriptions()
|
regs = api.GetPeerServerSubscriptions()
|
||||||
// count how many (fake) subscriptions there are
|
// count how many (fake) subscriptions there are
|
||||||
cnt := 0
|
cnt := 0
|
||||||
for _, reg := range regs {
|
for _, reg := range regs {
|
||||||
@ -1175,11 +1173,11 @@ func TestGetSubscriptions(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
TestGetSubscriptionsRPC sets up a simulation network of `nodeCount` nodes,
|
TestGetServerSubscriptionsRPC sets up a simulation network of `nodeCount` nodes,
|
||||||
starts the simulation, waits for SyncUpdateDelay in order to kick off
|
starts the simulation, waits for SyncUpdateDelay in order to kick off
|
||||||
stream registration, then tests that there are subscriptions.
|
stream registration, then tests that there are subscriptions.
|
||||||
*/
|
*/
|
||||||
func TestGetSubscriptionsRPC(t *testing.T) {
|
func TestGetServerSubscriptionsRPC(t *testing.T) {
|
||||||
|
|
||||||
if testutil.RaceEnabled && os.Getenv("TRAVIS") == "true" {
|
if testutil.RaceEnabled && os.Getenv("TRAVIS") == "true" {
|
||||||
t.Skip("flaky with -race on Travis")
|
t.Skip("flaky with -race on Travis")
|
||||||
@ -1226,7 +1224,6 @@ func TestGetSubscriptionsRPC(t *testing.T) {
|
|||||||
|
|
||||||
// configure so that sync registrations actually happen
|
// configure so that sync registrations actually happen
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
Retrieval: RetrievalEnabled,
|
|
||||||
Syncing: SyncingAutoSubscribe, //enable sync registrations
|
Syncing: SyncingAutoSubscribe, //enable sync registrations
|
||||||
SyncUpdateDelay: syncUpdateDelay,
|
SyncUpdateDelay: syncUpdateDelay,
|
||||||
}, nil)
|
}, nil)
|
||||||
@ -1321,7 +1318,7 @@ func TestGetSubscriptionsRPC(t *testing.T) {
|
|||||||
|
|
||||||
//ask it for subscriptions
|
//ask it for subscriptions
|
||||||
pstreams := make(map[string][]string)
|
pstreams := make(map[string][]string)
|
||||||
err = client.Call(&pstreams, "stream_getPeerSubscriptions")
|
err = client.Call(&pstreams, "stream_getPeerServerSubscriptions")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("client call stream_getPeerSubscriptions: %v", err)
|
return fmt.Errorf("client call stream_getPeerSubscriptions: %v", err)
|
||||||
}
|
}
|
||||||
|
@ -35,26 +35,26 @@ const (
|
|||||||
// * (live/non-live historical) chunk syncing per proximity bin
|
// * (live/non-live historical) chunk syncing per proximity bin
|
||||||
type SwarmSyncerServer struct {
|
type SwarmSyncerServer struct {
|
||||||
po uint8
|
po uint8
|
||||||
store chunk.FetchStore
|
netStore *storage.NetStore
|
||||||
quit chan struct{}
|
quit chan struct{}
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewSwarmSyncerServer is constructor for SwarmSyncerServer
|
// NewSwarmSyncerServer is constructor for SwarmSyncerServer
|
||||||
func NewSwarmSyncerServer(po uint8, syncChunkStore chunk.FetchStore) (*SwarmSyncerServer, error) {
|
func NewSwarmSyncerServer(po uint8, netStore *storage.NetStore) (*SwarmSyncerServer, error) {
|
||||||
return &SwarmSyncerServer{
|
return &SwarmSyncerServer{
|
||||||
po: po,
|
po: po,
|
||||||
store: syncChunkStore,
|
netStore: netStore,
|
||||||
quit: make(chan struct{}),
|
quit: make(chan struct{}),
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func RegisterSwarmSyncerServer(streamer *Registry, syncChunkStore chunk.FetchStore) {
|
func RegisterSwarmSyncerServer(streamer *Registry, netStore *storage.NetStore) {
|
||||||
streamer.RegisterServerFunc("SYNC", func(_ *Peer, t string, _ bool) (Server, error) {
|
streamer.RegisterServerFunc("SYNC", func(_ *Peer, t string, _ bool) (Server, error) {
|
||||||
po, err := ParseSyncBinKey(t)
|
po, err := ParseSyncBinKey(t)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return NewSwarmSyncerServer(po, syncChunkStore)
|
return NewSwarmSyncerServer(po, netStore)
|
||||||
})
|
})
|
||||||
// streamer.RegisterServerFunc(stream, func(p *Peer) (Server, error) {
|
// streamer.RegisterServerFunc(stream, func(p *Peer) (Server, error) {
|
||||||
// return NewOutgoingProvableSwarmSyncer(po, db)
|
// return NewOutgoingProvableSwarmSyncer(po, db)
|
||||||
@ -68,7 +68,7 @@ func (s *SwarmSyncerServer) Close() {
|
|||||||
|
|
||||||
// GetData retrieves the actual chunk from netstore
|
// GetData retrieves the actual chunk from netstore
|
||||||
func (s *SwarmSyncerServer) GetData(ctx context.Context, key []byte) ([]byte, error) {
|
func (s *SwarmSyncerServer) GetData(ctx context.Context, key []byte) ([]byte, error) {
|
||||||
ch, err := s.store.Get(ctx, chunk.ModeGetSync, storage.Address(key))
|
ch, err := s.netStore.Get(ctx, chunk.ModeGetSync, storage.Address(key))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -77,7 +77,7 @@ func (s *SwarmSyncerServer) GetData(ctx context.Context, key []byte) ([]byte, er
|
|||||||
|
|
||||||
// SessionIndex returns current storage bin (po) index.
|
// SessionIndex returns current storage bin (po) index.
|
||||||
func (s *SwarmSyncerServer) SessionIndex() (uint64, error) {
|
func (s *SwarmSyncerServer) SessionIndex() (uint64, error) {
|
||||||
return s.store.LastPullSubscriptionBinID(s.po)
|
return s.netStore.LastPullSubscriptionBinID(s.po)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetNextBatch retrieves the next batch of hashes from the localstore.
|
// SetNextBatch retrieves the next batch of hashes from the localstore.
|
||||||
@ -88,7 +88,7 @@ func (s *SwarmSyncerServer) SessionIndex() (uint64, error) {
|
|||||||
// are added in batchTimeout period, the batch will be returned. This function
|
// are added in batchTimeout period, the batch will be returned. This function
|
||||||
// will block until new chunks are received from localstore pull subscription.
|
// will block until new chunks are received from localstore pull subscription.
|
||||||
func (s *SwarmSyncerServer) SetNextBatch(from, to uint64) ([]byte, uint64, uint64, *HandoverProof, error) {
|
func (s *SwarmSyncerServer) SetNextBatch(from, to uint64) ([]byte, uint64, uint64, *HandoverProof, error) {
|
||||||
descriptors, stop := s.store.SubscribePull(context.Background(), s.po, from, to)
|
descriptors, stop := s.netStore.SubscribePull(context.Background(), s.po, from, to)
|
||||||
defer stop()
|
defer stop()
|
||||||
|
|
||||||
const batchTimeout = 2 * time.Second
|
const batchTimeout = 2 * time.Second
|
||||||
@ -118,7 +118,7 @@ func (s *SwarmSyncerServer) SetNextBatch(from, to uint64) ([]byte, uint64, uint6
|
|||||||
// This is the most naive approach to label the chunk as synced
|
// This is the most naive approach to label the chunk as synced
|
||||||
// allowing it to be garbage collected. A proper way requires
|
// allowing it to be garbage collected. A proper way requires
|
||||||
// validating that the chunk is successfully stored by the peer.
|
// validating that the chunk is successfully stored by the peer.
|
||||||
err := s.store.Set(context.Background(), chunk.ModeSetSync, d.Address)
|
err := s.netStore.Set(context.Background(), chunk.ModeSetSync, d.Address)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, 0, 0, nil, err
|
return nil, 0, 0, nil, err
|
||||||
}
|
}
|
||||||
@ -158,67 +158,31 @@ func (s *SwarmSyncerServer) SetNextBatch(from, to uint64) ([]byte, uint64, uint6
|
|||||||
|
|
||||||
// SwarmSyncerClient
|
// SwarmSyncerClient
|
||||||
type SwarmSyncerClient struct {
|
type SwarmSyncerClient struct {
|
||||||
store chunk.FetchStore
|
netStore *storage.NetStore
|
||||||
peer *Peer
|
peer *Peer
|
||||||
stream Stream
|
stream Stream
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewSwarmSyncerClient is a contructor for provable data exchange syncer
|
// NewSwarmSyncerClient is a contructor for provable data exchange syncer
|
||||||
func NewSwarmSyncerClient(p *Peer, store chunk.FetchStore, stream Stream) (*SwarmSyncerClient, error) {
|
func NewSwarmSyncerClient(p *Peer, netStore *storage.NetStore, stream Stream) (*SwarmSyncerClient, error) {
|
||||||
return &SwarmSyncerClient{
|
return &SwarmSyncerClient{
|
||||||
store: store,
|
netStore: netStore,
|
||||||
peer: p,
|
peer: p,
|
||||||
stream: stream,
|
stream: stream,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// // NewIncomingProvableSwarmSyncer is a contructor for provable data exchange syncer
|
|
||||||
// func NewIncomingProvableSwarmSyncer(po int, priority int, index uint64, sessionAt uint64, intervals []uint64, sessionRoot storage.Address, chunker *storage.PyramidChunker, store storage.ChunkStore, p Peer) *SwarmSyncerClient {
|
|
||||||
// retrieveC := make(storage.Chunk, chunksCap)
|
|
||||||
// RunChunkRequestor(p, retrieveC)
|
|
||||||
// storeC := make(storage.Chunk, chunksCap)
|
|
||||||
// RunChunkStorer(store, storeC)
|
|
||||||
// s := &SwarmSyncerClient{
|
|
||||||
// po: po,
|
|
||||||
// priority: priority,
|
|
||||||
// sessionAt: sessionAt,
|
|
||||||
// start: index,
|
|
||||||
// end: index,
|
|
||||||
// nextC: make(chan struct{}, 1),
|
|
||||||
// intervals: intervals,
|
|
||||||
// sessionRoot: sessionRoot,
|
|
||||||
// sessionReader: chunker.Join(sessionRoot, retrieveC),
|
|
||||||
// retrieveC: retrieveC,
|
|
||||||
// storeC: storeC,
|
|
||||||
// }
|
|
||||||
// return s
|
|
||||||
// }
|
|
||||||
|
|
||||||
// // StartSyncing is called on the Peer to start the syncing process
|
|
||||||
// // the idea is that it is called only after kademlia is close to healthy
|
|
||||||
// func StartSyncing(s *Streamer, peerId enode.ID, po uint8, nn bool) {
|
|
||||||
// lastPO := po
|
|
||||||
// if nn {
|
|
||||||
// lastPO = maxPO
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// for i := po; i <= lastPO; i++ {
|
|
||||||
// s.Subscribe(peerId, "SYNC", newSyncLabel("LIVE", po), 0, 0, High, true)
|
|
||||||
// s.Subscribe(peerId, "SYNC", newSyncLabel("HISTORY", po), 0, 0, Mid, false)
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
// RegisterSwarmSyncerClient registers the client constructor function for
|
// RegisterSwarmSyncerClient registers the client constructor function for
|
||||||
// to handle incoming sync streams
|
// to handle incoming sync streams
|
||||||
func RegisterSwarmSyncerClient(streamer *Registry, store chunk.FetchStore) {
|
func RegisterSwarmSyncerClient(streamer *Registry, netStore *storage.NetStore) {
|
||||||
streamer.RegisterClientFunc("SYNC", func(p *Peer, t string, live bool) (Client, error) {
|
streamer.RegisterClientFunc("SYNC", func(p *Peer, t string, live bool) (Client, error) {
|
||||||
return NewSwarmSyncerClient(p, store, NewStream("SYNC", t, live))
|
return NewSwarmSyncerClient(p, netStore, NewStream("SYNC", t, live))
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
// NeedData
|
// NeedData
|
||||||
func (s *SwarmSyncerClient) NeedData(ctx context.Context, key []byte) (wait func(context.Context) error) {
|
func (s *SwarmSyncerClient) NeedData(ctx context.Context, key []byte) (wait func(context.Context) error) {
|
||||||
return s.store.FetchFunc(ctx, key)
|
return s.netStore.FetchFunc(ctx, key)
|
||||||
}
|
}
|
||||||
|
|
||||||
// BatchDone
|
// BatchDone
|
||||||
|
@ -83,7 +83,6 @@ func testSyncBetweenNodes(t *testing.T, nodes, chunkCount int, skipCheck bool, p
|
|||||||
}
|
}
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, store, &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, store, &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingAutoSubscribe,
|
Syncing: SyncingAutoSubscribe,
|
||||||
SkipCheck: skipCheck,
|
SkipCheck: skipCheck,
|
||||||
}, nil)
|
}, nil)
|
||||||
@ -232,7 +231,6 @@ func TestSameVersionID(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingAutoSubscribe,
|
Syncing: SyncingAutoSubscribe,
|
||||||
}, nil)
|
}, nil)
|
||||||
bucket.Store(bucketKeyRegistry, r)
|
bucket.Store(bucketKeyRegistry, r)
|
||||||
@ -296,7 +294,6 @@ func TestDifferentVersionID(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
|
||||||
Retrieval: RetrievalDisabled,
|
|
||||||
Syncing: SyncingAutoSubscribe,
|
Syncing: SyncingAutoSubscribe,
|
||||||
}, nil)
|
}, nil)
|
||||||
bucket.Store(bucketKeyRegistry, r)
|
bucket.Store(bucketKeyRegistry, r)
|
||||||
|
@ -169,7 +169,7 @@ func (db *DB) Import(r io.Reader, legacy bool) (count int64, err error) {
|
|||||||
wg.Done()
|
wg.Done()
|
||||||
<-tokenPool
|
<-tokenPool
|
||||||
default:
|
default:
|
||||||
err := db.Put(ctx, chunk.ModePutUpload, ch)
|
_, err := db.Put(ctx, chunk.ModePutUpload, ch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errC <- err
|
errC <- err
|
||||||
}
|
}
|
||||||
|
@ -204,15 +204,9 @@ func NewSwarm(config *api.Config, mockStore *mock.NodeStore) (self *Swarm, err e
|
|||||||
syncing = stream.SyncingDisabled
|
syncing = stream.SyncingDisabled
|
||||||
}
|
}
|
||||||
|
|
||||||
retrieval := stream.RetrievalEnabled
|
|
||||||
if config.LightNodeEnabled {
|
|
||||||
retrieval = stream.RetrievalClientOnly
|
|
||||||
}
|
|
||||||
|
|
||||||
registryOptions := &stream.RegistryOptions{
|
registryOptions := &stream.RegistryOptions{
|
||||||
SkipCheck: config.DeliverySkipCheck,
|
SkipCheck: config.DeliverySkipCheck,
|
||||||
Syncing: syncing,
|
Syncing: syncing,
|
||||||
Retrieval: retrieval,
|
|
||||||
SyncUpdateDelay: config.SyncUpdateDelay,
|
SyncUpdateDelay: config.SyncUpdateDelay,
|
||||||
MaxPeerServers: config.MaxStreamPeerServers,
|
MaxPeerServers: config.MaxStreamPeerServers,
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user