forked from cerc-io/plugeth
4d300e4dec
This change imports the Swarm protocol codebase. Compared to the 'swarm' branch, a few mostly cosmetic changes had to be made: * The various redundant log message prefixes are gone. * All files now have LGPLv3 license headers. * Minor code changes were needed to please go vet and make the tests pass on Windows. * Further changes were required to adapt to the go-ethereum develop branch and its new Go APIs. Some code has not (yet) been brought over: * swarm/cmd/bzzhash: will reappear as cmd/bzzhash later * swarm/cmd/bzzup.sh: will be reimplemented in cmd/bzzup * swarm/cmd/makegenesis: will reappear somehow * swarm/examples/album: will move to a separate repository * swarm/examples/filemanager: ditto * swarm/examples/files: will not be merged * swarm/test/*: will not be merged * swarm/services/swear: will reappear as contracts/swear when needed
304 lines
9.0 KiB
Go
304 lines
9.0 KiB
Go
// Copyright 2016 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package storage
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"io"
|
|
"runtime"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
)
|
|
|
|
/*
|
|
Tests TreeChunker by splitting and joining a random byte slice
|
|
*/
|
|
|
|
type test interface {
|
|
Fatalf(string, ...interface{})
|
|
Logf(string, ...interface{})
|
|
}
|
|
|
|
type chunkerTester struct {
|
|
inputs map[uint64][]byte
|
|
chunks map[string]*Chunk
|
|
t test
|
|
}
|
|
|
|
func (self *chunkerTester) checkChunks(t *testing.T, want int) {
|
|
l := len(self.chunks)
|
|
if l != want {
|
|
t.Errorf("expected %v chunks, got %v", want, l)
|
|
}
|
|
}
|
|
|
|
func (self *chunkerTester) Split(chunker Splitter, data io.Reader, size int64, chunkC chan *Chunk, swg *sync.WaitGroup) (key Key) {
|
|
// reset
|
|
self.chunks = make(map[string]*Chunk)
|
|
|
|
if self.inputs == nil {
|
|
self.inputs = make(map[uint64][]byte)
|
|
}
|
|
|
|
quitC := make(chan bool)
|
|
timeout := time.After(600 * time.Second)
|
|
if chunkC != nil {
|
|
go func() {
|
|
for {
|
|
select {
|
|
case <-timeout:
|
|
self.t.Fatalf("Join timeout error")
|
|
|
|
case chunk, ok := <-chunkC:
|
|
if !ok {
|
|
// glog.V(logger.Info).Infof("chunkC closed quitting")
|
|
close(quitC)
|
|
return
|
|
}
|
|
// glog.V(logger.Info).Infof("chunk %v received", len(self.chunks))
|
|
// self.chunks = append(self.chunks, chunk)
|
|
self.chunks[chunk.Key.String()] = chunk
|
|
if chunk.wg != nil {
|
|
chunk.wg.Done()
|
|
}
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
key, err := chunker.Split(data, size, chunkC, swg, nil)
|
|
if err != nil {
|
|
self.t.Fatalf("Split error: %v", err)
|
|
}
|
|
if chunkC != nil {
|
|
if swg != nil {
|
|
// glog.V(logger.Info).Infof("Waiting for storage to finish")
|
|
swg.Wait()
|
|
// glog.V(logger.Info).Infof("Storage finished")
|
|
}
|
|
close(chunkC)
|
|
}
|
|
if chunkC != nil {
|
|
// glog.V(logger.Info).Infof("waiting for splitter finished")
|
|
<-quitC
|
|
// glog.V(logger.Info).Infof("Splitter finished")
|
|
}
|
|
return
|
|
}
|
|
|
|
func (self *chunkerTester) Join(chunker Chunker, key Key, c int, chunkC chan *Chunk, quitC chan bool) LazySectionReader {
|
|
// reset but not the chunks
|
|
|
|
// glog.V(logger.Info).Infof("Splitter finished")
|
|
reader := chunker.Join(key, chunkC)
|
|
|
|
timeout := time.After(600 * time.Second)
|
|
// glog.V(logger.Info).Infof("Splitter finished")
|
|
i := 0
|
|
go func() {
|
|
for {
|
|
select {
|
|
case <-timeout:
|
|
self.t.Fatalf("Join timeout error")
|
|
|
|
case chunk, ok := <-chunkC:
|
|
if !ok {
|
|
close(quitC)
|
|
return
|
|
}
|
|
// glog.V(logger.Info).Infof("chunk %v: %v", i, chunk.Key.String())
|
|
// this just mocks the behaviour of a chunk store retrieval
|
|
stored, success := self.chunks[chunk.Key.String()]
|
|
// glog.V(logger.Info).Infof("chunk %v, success: %v", chunk.Key.String(), success)
|
|
if !success {
|
|
self.t.Fatalf("not found")
|
|
return
|
|
}
|
|
// glog.V(logger.Info).Infof("chunk %v: %v", i, chunk.Key.String())
|
|
chunk.SData = stored.SData
|
|
chunk.Size = int64(binary.LittleEndian.Uint64(chunk.SData[0:8]))
|
|
close(chunk.C)
|
|
i++
|
|
}
|
|
}
|
|
}()
|
|
return reader
|
|
}
|
|
|
|
func testRandomData(splitter Splitter, n int, tester *chunkerTester) {
|
|
if tester.inputs == nil {
|
|
tester.inputs = make(map[uint64][]byte)
|
|
}
|
|
input, found := tester.inputs[uint64(n)]
|
|
var data io.Reader
|
|
if !found {
|
|
data, input = testDataReaderAndSlice(n)
|
|
tester.inputs[uint64(n)] = input
|
|
} else {
|
|
data = limitReader(bytes.NewReader(input), n)
|
|
}
|
|
|
|
chunkC := make(chan *Chunk, 1000)
|
|
swg := &sync.WaitGroup{}
|
|
|
|
key := tester.Split(splitter, data, int64(n), chunkC, swg)
|
|
tester.t.Logf(" Key = %v\n", key)
|
|
|
|
chunkC = make(chan *Chunk, 1000)
|
|
quitC := make(chan bool)
|
|
|
|
chunker := NewTreeChunker(NewChunkerParams())
|
|
reader := tester.Join(chunker, key, 0, chunkC, quitC)
|
|
output := make([]byte, n)
|
|
// glog.V(logger.Info).Infof(" Key = %v\n", key)
|
|
r, err := reader.Read(output)
|
|
// glog.V(logger.Info).Infof(" read = %v %v\n", r, err)
|
|
if r != n || err != io.EOF {
|
|
tester.t.Fatalf("read error read: %v n = %v err = %v\n", r, n, err)
|
|
}
|
|
if input != nil {
|
|
if !bytes.Equal(output, input) {
|
|
tester.t.Fatalf("input and output mismatch\n IN: %v\nOUT: %v\n", input, output)
|
|
}
|
|
}
|
|
close(chunkC)
|
|
<-quitC
|
|
}
|
|
|
|
func TestRandomData(t *testing.T) {
|
|
// sizes := []int{123456}
|
|
sizes := []int{1, 60, 83, 179, 253, 1024, 4095, 4096, 4097, 123456}
|
|
tester := &chunkerTester{t: t}
|
|
chunker := NewTreeChunker(NewChunkerParams())
|
|
for _, s := range sizes {
|
|
testRandomData(chunker, s, tester)
|
|
}
|
|
pyramid := NewPyramidChunker(NewChunkerParams())
|
|
for _, s := range sizes {
|
|
testRandomData(pyramid, s, tester)
|
|
}
|
|
}
|
|
|
|
func readAll(reader LazySectionReader, result []byte) {
|
|
size := int64(len(result))
|
|
|
|
var end int64
|
|
for pos := int64(0); pos < size; pos += 1000 {
|
|
if pos+1000 > size {
|
|
end = size
|
|
} else {
|
|
end = pos + 1000
|
|
}
|
|
reader.ReadAt(result[pos:end], pos)
|
|
}
|
|
}
|
|
|
|
func benchReadAll(reader LazySectionReader) {
|
|
size, _ := reader.Size(nil)
|
|
output := make([]byte, 1000)
|
|
for pos := int64(0); pos < size; pos += 1000 {
|
|
reader.ReadAt(output, pos)
|
|
}
|
|
}
|
|
|
|
func benchmarkJoin(n int, t *testing.B) {
|
|
t.ReportAllocs()
|
|
for i := 0; i < t.N; i++ {
|
|
chunker := NewTreeChunker(NewChunkerParams())
|
|
tester := &chunkerTester{t: t}
|
|
data := testDataReader(n)
|
|
|
|
chunkC := make(chan *Chunk, 1000)
|
|
swg := &sync.WaitGroup{}
|
|
|
|
key := tester.Split(chunker, data, int64(n), chunkC, swg)
|
|
// t.StartTimer()
|
|
chunkC = make(chan *Chunk, 1000)
|
|
quitC := make(chan bool)
|
|
reader := tester.Join(chunker, key, i, chunkC, quitC)
|
|
benchReadAll(reader)
|
|
close(chunkC)
|
|
<-quitC
|
|
// t.StopTimer()
|
|
}
|
|
stats := new(runtime.MemStats)
|
|
runtime.ReadMemStats(stats)
|
|
fmt.Println(stats.Sys)
|
|
}
|
|
|
|
func benchmarkSplitTree(n int, t *testing.B) {
|
|
t.ReportAllocs()
|
|
for i := 0; i < t.N; i++ {
|
|
chunker := NewTreeChunker(NewChunkerParams())
|
|
tester := &chunkerTester{t: t}
|
|
data := testDataReader(n)
|
|
// glog.V(logger.Info).Infof("splitting data of length %v", n)
|
|
tester.Split(chunker, data, int64(n), nil, nil)
|
|
}
|
|
stats := new(runtime.MemStats)
|
|
runtime.ReadMemStats(stats)
|
|
fmt.Println(stats.Sys)
|
|
}
|
|
|
|
func benchmarkSplitPyramid(n int, t *testing.B) {
|
|
t.ReportAllocs()
|
|
for i := 0; i < t.N; i++ {
|
|
splitter := NewPyramidChunker(NewChunkerParams())
|
|
tester := &chunkerTester{t: t}
|
|
data := testDataReader(n)
|
|
// glog.V(logger.Info).Infof("splitting data of length %v", n)
|
|
tester.Split(splitter, data, int64(n), nil, nil)
|
|
}
|
|
stats := new(runtime.MemStats)
|
|
runtime.ReadMemStats(stats)
|
|
fmt.Println(stats.Sys)
|
|
}
|
|
|
|
func BenchmarkJoin_2(t *testing.B) { benchmarkJoin(100, t) }
|
|
func BenchmarkJoin_3(t *testing.B) { benchmarkJoin(1000, t) }
|
|
func BenchmarkJoin_4(t *testing.B) { benchmarkJoin(10000, t) }
|
|
func BenchmarkJoin_5(t *testing.B) { benchmarkJoin(100000, t) }
|
|
func BenchmarkJoin_6(t *testing.B) { benchmarkJoin(1000000, t) }
|
|
func BenchmarkJoin_7(t *testing.B) { benchmarkJoin(10000000, t) }
|
|
func BenchmarkJoin_8(t *testing.B) { benchmarkJoin(100000000, t) }
|
|
|
|
func BenchmarkSplitTree_2(t *testing.B) { benchmarkSplitTree(100, t) }
|
|
func BenchmarkSplitTree_2h(t *testing.B) { benchmarkSplitTree(500, t) }
|
|
func BenchmarkSplitTree_3(t *testing.B) { benchmarkSplitTree(1000, t) }
|
|
func BenchmarkSplitTree_3h(t *testing.B) { benchmarkSplitTree(5000, t) }
|
|
func BenchmarkSplitTree_4(t *testing.B) { benchmarkSplitTree(10000, t) }
|
|
func BenchmarkSplitTree_4h(t *testing.B) { benchmarkSplitTree(50000, t) }
|
|
func BenchmarkSplitTree_5(t *testing.B) { benchmarkSplitTree(100000, t) }
|
|
func BenchmarkSplitTree_6(t *testing.B) { benchmarkSplitTree(1000000, t) }
|
|
func BenchmarkSplitTree_7(t *testing.B) { benchmarkSplitTree(10000000, t) }
|
|
func BenchmarkSplitTree_8(t *testing.B) { benchmarkSplitTree(100000000, t) }
|
|
|
|
func BenchmarkSplitPyramid_2(t *testing.B) { benchmarkSplitPyramid(100, t) }
|
|
func BenchmarkSplitPyramid_2h(t *testing.B) { benchmarkSplitPyramid(500, t) }
|
|
func BenchmarkSplitPyramid_3(t *testing.B) { benchmarkSplitPyramid(1000, t) }
|
|
func BenchmarkSplitPyramid_3h(t *testing.B) { benchmarkSplitPyramid(5000, t) }
|
|
func BenchmarkSplitPyramid_4(t *testing.B) { benchmarkSplitPyramid(10000, t) }
|
|
func BenchmarkSplitPyramid_4h(t *testing.B) { benchmarkSplitPyramid(50000, t) }
|
|
func BenchmarkSplitPyramid_5(t *testing.B) { benchmarkSplitPyramid(100000, t) }
|
|
func BenchmarkSplitPyramid_6(t *testing.B) { benchmarkSplitPyramid(1000000, t) }
|
|
func BenchmarkSplitPyramid_7(t *testing.B) { benchmarkSplitPyramid(10000000, t) }
|
|
func BenchmarkSplitPyramid_8(t *testing.B) { benchmarkSplitPyramid(100000000, t) }
|
|
|
|
// godep go test -bench ./swarm/storage -cpuprofile cpu.out -memprofile mem.out
|