forked from cerc-io/plugeth
consensus/ethash: refactor remote sealer (#20335)
The original idea behind this change was to remove a use of the deprecated CancelRequest method. Simply removing it would've been an option, but I couldn't resist and did a bit of a refactoring instead. All remote sealing code was contained in a single giant function. Remote sealing is now extracted into its own object, remoteSealer.
This commit is contained in:
parent
c9f28ca8e5
commit
a91b704b01
@ -729,7 +729,7 @@ func TestConcurrentDiskCacheGeneration(t *testing.T) {
|
|||||||
|
|
||||||
go func(idx int) {
|
go func(idx int) {
|
||||||
defer pend.Done()
|
defer pend.Done()
|
||||||
ethash := New(Config{cachedir, 0, 1, "", 0, 0, ModeNormal}, nil, false)
|
ethash := New(Config{cachedir, 0, 1, "", 0, 0, ModeNormal, nil}, nil, false)
|
||||||
defer ethash.Close()
|
defer ethash.Close()
|
||||||
if err := ethash.VerifySeal(nil, block.Header()); err != nil {
|
if err := ethash.VerifySeal(nil, block.Header()); err != nil {
|
||||||
t.Errorf("proc %d: block verification failed: %v", idx, err)
|
t.Errorf("proc %d: block verification failed: %v", idx, err)
|
||||||
|
@ -28,7 +28,7 @@ var errEthashStopped = errors.New("ethash stopped")
|
|||||||
|
|
||||||
// API exposes ethash related methods for the RPC interface.
|
// API exposes ethash related methods for the RPC interface.
|
||||||
type API struct {
|
type API struct {
|
||||||
ethash *Ethash // Make sure the mode of ethash is normal.
|
ethash *Ethash
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetWork returns a work package for external miner.
|
// GetWork returns a work package for external miner.
|
||||||
@ -39,7 +39,7 @@ type API struct {
|
|||||||
// result[2] - 32 bytes hex encoded boundary condition ("target"), 2^256/difficulty
|
// result[2] - 32 bytes hex encoded boundary condition ("target"), 2^256/difficulty
|
||||||
// result[3] - hex encoded block number
|
// result[3] - hex encoded block number
|
||||||
func (api *API) GetWork() ([4]string, error) {
|
func (api *API) GetWork() ([4]string, error) {
|
||||||
if api.ethash.config.PowMode != ModeNormal && api.ethash.config.PowMode != ModeTest {
|
if api.ethash.remote == nil {
|
||||||
return [4]string{}, errors.New("not supported")
|
return [4]string{}, errors.New("not supported")
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -47,13 +47,11 @@ func (api *API) GetWork() ([4]string, error) {
|
|||||||
workCh = make(chan [4]string, 1)
|
workCh = make(chan [4]string, 1)
|
||||||
errc = make(chan error, 1)
|
errc = make(chan error, 1)
|
||||||
)
|
)
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case api.ethash.fetchWorkCh <- &sealWork{errc: errc, res: workCh}:
|
case api.ethash.remote.fetchWorkCh <- &sealWork{errc: errc, res: workCh}:
|
||||||
case <-api.ethash.exitCh:
|
case <-api.ethash.remote.exitCh:
|
||||||
return [4]string{}, errEthashStopped
|
return [4]string{}, errEthashStopped
|
||||||
}
|
}
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case work := <-workCh:
|
case work := <-workCh:
|
||||||
return work, nil
|
return work, nil
|
||||||
@ -66,23 +64,21 @@ func (api *API) GetWork() ([4]string, error) {
|
|||||||
// It returns an indication if the work was accepted.
|
// It returns an indication if the work was accepted.
|
||||||
// Note either an invalid solution, a stale work a non-existent work will return false.
|
// Note either an invalid solution, a stale work a non-existent work will return false.
|
||||||
func (api *API) SubmitWork(nonce types.BlockNonce, hash, digest common.Hash) bool {
|
func (api *API) SubmitWork(nonce types.BlockNonce, hash, digest common.Hash) bool {
|
||||||
if api.ethash.config.PowMode != ModeNormal && api.ethash.config.PowMode != ModeTest {
|
if api.ethash.remote == nil {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
var errc = make(chan error, 1)
|
var errc = make(chan error, 1)
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case api.ethash.submitWorkCh <- &mineResult{
|
case api.ethash.remote.submitWorkCh <- &mineResult{
|
||||||
nonce: nonce,
|
nonce: nonce,
|
||||||
mixDigest: digest,
|
mixDigest: digest,
|
||||||
hash: hash,
|
hash: hash,
|
||||||
errc: errc,
|
errc: errc,
|
||||||
}:
|
}:
|
||||||
case <-api.ethash.exitCh:
|
case <-api.ethash.remote.exitCh:
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
err := <-errc
|
err := <-errc
|
||||||
return err == nil
|
return err == nil
|
||||||
}
|
}
|
||||||
@ -94,21 +90,19 @@ func (api *API) SubmitWork(nonce types.BlockNonce, hash, digest common.Hash) boo
|
|||||||
// It accepts the miner hash rate and an identifier which must be unique
|
// It accepts the miner hash rate and an identifier which must be unique
|
||||||
// between nodes.
|
// between nodes.
|
||||||
func (api *API) SubmitHashRate(rate hexutil.Uint64, id common.Hash) bool {
|
func (api *API) SubmitHashRate(rate hexutil.Uint64, id common.Hash) bool {
|
||||||
if api.ethash.config.PowMode != ModeNormal && api.ethash.config.PowMode != ModeTest {
|
if api.ethash.remote == nil {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
var done = make(chan struct{}, 1)
|
var done = make(chan struct{}, 1)
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case api.ethash.submitRateCh <- &hashrate{done: done, rate: uint64(rate), id: id}:
|
case api.ethash.remote.submitRateCh <- &hashrate{done: done, rate: uint64(rate), id: id}:
|
||||||
case <-api.ethash.exitCh:
|
case <-api.ethash.remote.exitCh:
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
// Block until hash rate submitted successfully.
|
// Block until hash rate submitted successfully.
|
||||||
<-done
|
<-done
|
||||||
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -34,9 +34,7 @@ import (
|
|||||||
"unsafe"
|
"unsafe"
|
||||||
|
|
||||||
mmap "github.com/edsrzf/mmap-go"
|
mmap "github.com/edsrzf/mmap-go"
|
||||||
"github.com/ethereum/go-ethereum/common"
|
|
||||||
"github.com/ethereum/go-ethereum/consensus"
|
"github.com/ethereum/go-ethereum/consensus"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
@ -50,7 +48,7 @@ var (
|
|||||||
two256 = new(big.Int).Exp(big.NewInt(2), big.NewInt(256), big.NewInt(0))
|
two256 = new(big.Int).Exp(big.NewInt(2), big.NewInt(256), big.NewInt(0))
|
||||||
|
|
||||||
// sharedEthash is a full instance that can be shared between multiple users.
|
// sharedEthash is a full instance that can be shared between multiple users.
|
||||||
sharedEthash = New(Config{"", 3, 0, "", 1, 0, ModeNormal}, nil, false)
|
sharedEthash = New(Config{"", 3, 0, "", 1, 0, ModeNormal, nil}, nil, false)
|
||||||
|
|
||||||
// algorithmRevision is the data structure version used for file naming.
|
// algorithmRevision is the data structure version used for file naming.
|
||||||
algorithmRevision = 23
|
algorithmRevision = 23
|
||||||
@ -403,36 +401,8 @@ type Config struct {
|
|||||||
DatasetsInMem int
|
DatasetsInMem int
|
||||||
DatasetsOnDisk int
|
DatasetsOnDisk int
|
||||||
PowMode Mode
|
PowMode Mode
|
||||||
}
|
|
||||||
|
|
||||||
// sealTask wraps a seal block with relative result channel for remote sealer thread.
|
Log log.Logger `toml:"-"`
|
||||||
type sealTask struct {
|
|
||||||
block *types.Block
|
|
||||||
results chan<- *types.Block
|
|
||||||
}
|
|
||||||
|
|
||||||
// mineResult wraps the pow solution parameters for the specified block.
|
|
||||||
type mineResult struct {
|
|
||||||
nonce types.BlockNonce
|
|
||||||
mixDigest common.Hash
|
|
||||||
hash common.Hash
|
|
||||||
|
|
||||||
errc chan error
|
|
||||||
}
|
|
||||||
|
|
||||||
// hashrate wraps the hash rate submitted by the remote sealer.
|
|
||||||
type hashrate struct {
|
|
||||||
id common.Hash
|
|
||||||
ping time.Time
|
|
||||||
rate uint64
|
|
||||||
|
|
||||||
done chan struct{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// sealWork wraps a seal work package for remote sealer.
|
|
||||||
type sealWork struct {
|
|
||||||
errc chan error
|
|
||||||
res chan [4]string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Ethash is a consensus engine based on proof-of-work implementing the ethash
|
// Ethash is a consensus engine based on proof-of-work implementing the ethash
|
||||||
@ -448,13 +418,7 @@ type Ethash struct {
|
|||||||
threads int // Number of threads to mine on if mining
|
threads int // Number of threads to mine on if mining
|
||||||
update chan struct{} // Notification channel to update mining parameters
|
update chan struct{} // Notification channel to update mining parameters
|
||||||
hashrate metrics.Meter // Meter tracking the average hashrate
|
hashrate metrics.Meter // Meter tracking the average hashrate
|
||||||
|
remote *remoteSealer
|
||||||
// Remote sealer related fields
|
|
||||||
workCh chan *sealTask // Notification channel to push new work and relative result channel to remote sealer
|
|
||||||
fetchWorkCh chan *sealWork // Channel used for remote sealer to fetch mining work
|
|
||||||
submitWorkCh chan *mineResult // Channel used for remote sealer to submit their mining result
|
|
||||||
fetchRateCh chan chan uint64 // Channel used to gather submitted hash rate for local or remote sealer.
|
|
||||||
submitRateCh chan *hashrate // Channel used for remote sealer to submit their mining hashrate
|
|
||||||
|
|
||||||
// The fields below are hooks for testing
|
// The fields below are hooks for testing
|
||||||
shared *Ethash // Shared PoW verifier to avoid cache regeneration
|
shared *Ethash // Shared PoW verifier to avoid cache regeneration
|
||||||
@ -463,22 +427,24 @@ type Ethash struct {
|
|||||||
|
|
||||||
lock sync.Mutex // Ensures thread safety for the in-memory caches and mining fields
|
lock sync.Mutex // Ensures thread safety for the in-memory caches and mining fields
|
||||||
closeOnce sync.Once // Ensures exit channel will not be closed twice.
|
closeOnce sync.Once // Ensures exit channel will not be closed twice.
|
||||||
exitCh chan chan error // Notification channel to exiting backend threads
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// New creates a full sized ethash PoW scheme and starts a background thread for
|
// New creates a full sized ethash PoW scheme and starts a background thread for
|
||||||
// remote mining, also optionally notifying a batch of remote services of new work
|
// remote mining, also optionally notifying a batch of remote services of new work
|
||||||
// packages.
|
// packages.
|
||||||
func New(config Config, notify []string, noverify bool) *Ethash {
|
func New(config Config, notify []string, noverify bool) *Ethash {
|
||||||
|
if config.Log == nil {
|
||||||
|
config.Log = log.Root()
|
||||||
|
}
|
||||||
if config.CachesInMem <= 0 {
|
if config.CachesInMem <= 0 {
|
||||||
log.Warn("One ethash cache must always be in memory", "requested", config.CachesInMem)
|
config.Log.Warn("One ethash cache must always be in memory", "requested", config.CachesInMem)
|
||||||
config.CachesInMem = 1
|
config.CachesInMem = 1
|
||||||
}
|
}
|
||||||
if config.CacheDir != "" && config.CachesOnDisk > 0 {
|
if config.CacheDir != "" && config.CachesOnDisk > 0 {
|
||||||
log.Info("Disk storage enabled for ethash caches", "dir", config.CacheDir, "count", config.CachesOnDisk)
|
config.Log.Info("Disk storage enabled for ethash caches", "dir", config.CacheDir, "count", config.CachesOnDisk)
|
||||||
}
|
}
|
||||||
if config.DatasetDir != "" && config.DatasetsOnDisk > 0 {
|
if config.DatasetDir != "" && config.DatasetsOnDisk > 0 {
|
||||||
log.Info("Disk storage enabled for ethash DAGs", "dir", config.DatasetDir, "count", config.DatasetsOnDisk)
|
config.Log.Info("Disk storage enabled for ethash DAGs", "dir", config.DatasetDir, "count", config.DatasetsOnDisk)
|
||||||
}
|
}
|
||||||
ethash := &Ethash{
|
ethash := &Ethash{
|
||||||
config: config,
|
config: config,
|
||||||
@ -486,14 +452,8 @@ func New(config Config, notify []string, noverify bool) *Ethash {
|
|||||||
datasets: newlru("dataset", config.DatasetsInMem, newDataset),
|
datasets: newlru("dataset", config.DatasetsInMem, newDataset),
|
||||||
update: make(chan struct{}),
|
update: make(chan struct{}),
|
||||||
hashrate: metrics.NewMeterForced(),
|
hashrate: metrics.NewMeterForced(),
|
||||||
workCh: make(chan *sealTask),
|
|
||||||
fetchWorkCh: make(chan *sealWork),
|
|
||||||
submitWorkCh: make(chan *mineResult),
|
|
||||||
fetchRateCh: make(chan chan uint64),
|
|
||||||
submitRateCh: make(chan *hashrate),
|
|
||||||
exitCh: make(chan chan error),
|
|
||||||
}
|
}
|
||||||
go ethash.remote(notify, noverify)
|
ethash.remote = startRemoteSealer(ethash, notify, noverify)
|
||||||
return ethash
|
return ethash
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -501,19 +461,13 @@ func New(config Config, notify []string, noverify bool) *Ethash {
|
|||||||
// purposes.
|
// purposes.
|
||||||
func NewTester(notify []string, noverify bool) *Ethash {
|
func NewTester(notify []string, noverify bool) *Ethash {
|
||||||
ethash := &Ethash{
|
ethash := &Ethash{
|
||||||
config: Config{PowMode: ModeTest},
|
config: Config{PowMode: ModeTest, Log: log.Root()},
|
||||||
caches: newlru("cache", 1, newCache),
|
caches: newlru("cache", 1, newCache),
|
||||||
datasets: newlru("dataset", 1, newDataset),
|
datasets: newlru("dataset", 1, newDataset),
|
||||||
update: make(chan struct{}),
|
update: make(chan struct{}),
|
||||||
hashrate: metrics.NewMeterForced(),
|
hashrate: metrics.NewMeterForced(),
|
||||||
workCh: make(chan *sealTask),
|
|
||||||
fetchWorkCh: make(chan *sealWork),
|
|
||||||
submitWorkCh: make(chan *mineResult),
|
|
||||||
fetchRateCh: make(chan chan uint64),
|
|
||||||
submitRateCh: make(chan *hashrate),
|
|
||||||
exitCh: make(chan chan error),
|
|
||||||
}
|
}
|
||||||
go ethash.remote(notify, noverify)
|
ethash.remote = startRemoteSealer(ethash, notify, noverify)
|
||||||
return ethash
|
return ethash
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -524,6 +478,7 @@ func NewFaker() *Ethash {
|
|||||||
return &Ethash{
|
return &Ethash{
|
||||||
config: Config{
|
config: Config{
|
||||||
PowMode: ModeFake,
|
PowMode: ModeFake,
|
||||||
|
Log: log.Root(),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -535,6 +490,7 @@ func NewFakeFailer(fail uint64) *Ethash {
|
|||||||
return &Ethash{
|
return &Ethash{
|
||||||
config: Config{
|
config: Config{
|
||||||
PowMode: ModeFake,
|
PowMode: ModeFake,
|
||||||
|
Log: log.Root(),
|
||||||
},
|
},
|
||||||
fakeFail: fail,
|
fakeFail: fail,
|
||||||
}
|
}
|
||||||
@ -547,6 +503,7 @@ func NewFakeDelayer(delay time.Duration) *Ethash {
|
|||||||
return &Ethash{
|
return &Ethash{
|
||||||
config: Config{
|
config: Config{
|
||||||
PowMode: ModeFake,
|
PowMode: ModeFake,
|
||||||
|
Log: log.Root(),
|
||||||
},
|
},
|
||||||
fakeDelay: delay,
|
fakeDelay: delay,
|
||||||
}
|
}
|
||||||
@ -558,6 +515,7 @@ func NewFullFaker() *Ethash {
|
|||||||
return &Ethash{
|
return &Ethash{
|
||||||
config: Config{
|
config: Config{
|
||||||
PowMode: ModeFullFake,
|
PowMode: ModeFullFake,
|
||||||
|
Log: log.Root(),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -573,13 +531,11 @@ func (ethash *Ethash) Close() error {
|
|||||||
var err error
|
var err error
|
||||||
ethash.closeOnce.Do(func() {
|
ethash.closeOnce.Do(func() {
|
||||||
// Short circuit if the exit channel is not allocated.
|
// Short circuit if the exit channel is not allocated.
|
||||||
if ethash.exitCh == nil {
|
if ethash.remote == nil {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
errc := make(chan error)
|
close(ethash.remote.requestExit)
|
||||||
ethash.exitCh <- errc
|
<-ethash.remote.exitCh
|
||||||
err = <-errc
|
|
||||||
close(ethash.exitCh)
|
|
||||||
})
|
})
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -680,8 +636,8 @@ func (ethash *Ethash) Hashrate() float64 {
|
|||||||
var res = make(chan uint64, 1)
|
var res = make(chan uint64, 1)
|
||||||
|
|
||||||
select {
|
select {
|
||||||
case ethash.fetchRateCh <- res:
|
case ethash.remote.fetchRateCh <- res:
|
||||||
case <-ethash.exitCh:
|
case <-ethash.remote.exitCh:
|
||||||
// Return local hashrate only if ethash is stopped.
|
// Return local hashrate only if ethash is stopped.
|
||||||
return ethash.hashrate.Rate1()
|
return ethash.hashrate.Rate1()
|
||||||
}
|
}
|
||||||
|
@ -18,6 +18,7 @@ package ethash
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
|
"context"
|
||||||
crand "crypto/rand"
|
crand "crypto/rand"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
@ -33,7 +34,6 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
"github.com/ethereum/go-ethereum/consensus"
|
"github.com/ethereum/go-ethereum/consensus"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -56,7 +56,7 @@ func (ethash *Ethash) Seal(chain consensus.ChainReader, block *types.Block, resu
|
|||||||
select {
|
select {
|
||||||
case results <- block.WithSeal(header):
|
case results <- block.WithSeal(header):
|
||||||
default:
|
default:
|
||||||
log.Warn("Sealing result is not read by miner", "mode", "fake", "sealhash", ethash.SealHash(block.Header()))
|
ethash.config.Log.Warn("Sealing result is not read by miner", "mode", "fake", "sealhash", ethash.SealHash(block.Header()))
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -85,8 +85,8 @@ func (ethash *Ethash) Seal(chain consensus.ChainReader, block *types.Block, resu
|
|||||||
threads = 0 // Allows disabling local mining without extra logic around local/remote
|
threads = 0 // Allows disabling local mining without extra logic around local/remote
|
||||||
}
|
}
|
||||||
// Push new work to remote sealer
|
// Push new work to remote sealer
|
||||||
if ethash.workCh != nil {
|
if ethash.remote != nil {
|
||||||
ethash.workCh <- &sealTask{block: block, results: results}
|
ethash.remote.workCh <- &sealTask{block: block, results: results}
|
||||||
}
|
}
|
||||||
var (
|
var (
|
||||||
pend sync.WaitGroup
|
pend sync.WaitGroup
|
||||||
@ -111,14 +111,14 @@ func (ethash *Ethash) Seal(chain consensus.ChainReader, block *types.Block, resu
|
|||||||
select {
|
select {
|
||||||
case results <- result:
|
case results <- result:
|
||||||
default:
|
default:
|
||||||
log.Warn("Sealing result is not read by miner", "mode", "local", "sealhash", ethash.SealHash(block.Header()))
|
ethash.config.Log.Warn("Sealing result is not read by miner", "mode", "local", "sealhash", ethash.SealHash(block.Header()))
|
||||||
}
|
}
|
||||||
close(abort)
|
close(abort)
|
||||||
case <-ethash.update:
|
case <-ethash.update:
|
||||||
// Thread count was changed on user request, restart
|
// Thread count was changed on user request, restart
|
||||||
close(abort)
|
close(abort)
|
||||||
if err := ethash.Seal(chain, block, results, stop); err != nil {
|
if err := ethash.Seal(chain, block, results, stop); err != nil {
|
||||||
log.Error("Failed to restart sealing after update", "err", err)
|
ethash.config.Log.Error("Failed to restart sealing after update", "err", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Wait for all miners to terminate and return the block
|
// Wait for all miners to terminate and return the block
|
||||||
@ -143,7 +143,7 @@ func (ethash *Ethash) mine(block *types.Block, id int, seed uint64, abort chan s
|
|||||||
attempts = int64(0)
|
attempts = int64(0)
|
||||||
nonce = seed
|
nonce = seed
|
||||||
)
|
)
|
||||||
logger := log.New("miner", id)
|
logger := ethash.config.Log.New("miner", id)
|
||||||
logger.Trace("Started ethash search for new nonces", "seed", seed)
|
logger.Trace("Started ethash search for new nonces", "seed", seed)
|
||||||
search:
|
search:
|
||||||
for {
|
for {
|
||||||
@ -186,80 +186,219 @@ search:
|
|||||||
runtime.KeepAlive(dataset)
|
runtime.KeepAlive(dataset)
|
||||||
}
|
}
|
||||||
|
|
||||||
// remote is a standalone goroutine to handle remote mining related stuff.
|
// This is the timeout for HTTP requests to notify external miners.
|
||||||
func (ethash *Ethash) remote(notify []string, noverify bool) {
|
const remoteSealerTimeout = 1 * time.Second
|
||||||
var (
|
|
||||||
works = make(map[common.Hash]*types.Block)
|
|
||||||
rates = make(map[common.Hash]hashrate)
|
|
||||||
|
|
||||||
results chan<- *types.Block
|
type remoteSealer struct {
|
||||||
|
works map[common.Hash]*types.Block
|
||||||
|
rates map[common.Hash]hashrate
|
||||||
currentBlock *types.Block
|
currentBlock *types.Block
|
||||||
currentWork [4]string
|
currentWork [4]string
|
||||||
|
notifyCtx context.Context
|
||||||
|
cancelNotify context.CancelFunc // cancels all notification requests
|
||||||
|
reqWG sync.WaitGroup // tracks notification request goroutines
|
||||||
|
|
||||||
notifyTransport = &http.Transport{}
|
ethash *Ethash
|
||||||
notifyClient = &http.Client{
|
noverify bool
|
||||||
Transport: notifyTransport,
|
notifyURLs []string
|
||||||
Timeout: time.Second,
|
results chan<- *types.Block
|
||||||
|
workCh chan *sealTask // Notification channel to push new work and relative result channel to remote sealer
|
||||||
|
fetchWorkCh chan *sealWork // Channel used for remote sealer to fetch mining work
|
||||||
|
submitWorkCh chan *mineResult // Channel used for remote sealer to submit their mining result
|
||||||
|
fetchRateCh chan chan uint64 // Channel used to gather submitted hash rate for local or remote sealer.
|
||||||
|
submitRateCh chan *hashrate // Channel used for remote sealer to submit their mining hashrate
|
||||||
|
requestExit chan struct{}
|
||||||
|
exitCh chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// sealTask wraps a seal block with relative result channel for remote sealer thread.
|
||||||
|
type sealTask struct {
|
||||||
|
block *types.Block
|
||||||
|
results chan<- *types.Block
|
||||||
|
}
|
||||||
|
|
||||||
|
// mineResult wraps the pow solution parameters for the specified block.
|
||||||
|
type mineResult struct {
|
||||||
|
nonce types.BlockNonce
|
||||||
|
mixDigest common.Hash
|
||||||
|
hash common.Hash
|
||||||
|
|
||||||
|
errc chan error
|
||||||
|
}
|
||||||
|
|
||||||
|
// hashrate wraps the hash rate submitted by the remote sealer.
|
||||||
|
type hashrate struct {
|
||||||
|
id common.Hash
|
||||||
|
ping time.Time
|
||||||
|
rate uint64
|
||||||
|
|
||||||
|
done chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// sealWork wraps a seal work package for remote sealer.
|
||||||
|
type sealWork struct {
|
||||||
|
errc chan error
|
||||||
|
res chan [4]string
|
||||||
|
}
|
||||||
|
|
||||||
|
func startRemoteSealer(ethash *Ethash, urls []string, noverify bool) *remoteSealer {
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
s := &remoteSealer{
|
||||||
|
ethash: ethash,
|
||||||
|
noverify: noverify,
|
||||||
|
notifyURLs: urls,
|
||||||
|
notifyCtx: ctx,
|
||||||
|
cancelNotify: cancel,
|
||||||
|
works: make(map[common.Hash]*types.Block),
|
||||||
|
rates: make(map[common.Hash]hashrate),
|
||||||
|
workCh: make(chan *sealTask),
|
||||||
|
fetchWorkCh: make(chan *sealWork),
|
||||||
|
submitWorkCh: make(chan *mineResult),
|
||||||
|
fetchRateCh: make(chan chan uint64),
|
||||||
|
submitRateCh: make(chan *hashrate),
|
||||||
|
requestExit: make(chan struct{}),
|
||||||
|
exitCh: make(chan struct{}),
|
||||||
}
|
}
|
||||||
notifyReqs = make([]*http.Request, len(notify))
|
go s.loop()
|
||||||
)
|
return s
|
||||||
// notifyWork notifies all the specified mining endpoints of the availability of
|
}
|
||||||
// new work to be processed.
|
|
||||||
notifyWork := func() {
|
|
||||||
work := currentWork
|
|
||||||
blob, _ := json.Marshal(work)
|
|
||||||
|
|
||||||
for i, url := range notify {
|
func (s *remoteSealer) loop() {
|
||||||
// Terminate any previously pending request and create the new work
|
defer func() {
|
||||||
if notifyReqs[i] != nil {
|
s.ethash.config.Log.Trace("Ethash remote sealer is exiting")
|
||||||
notifyTransport.CancelRequest(notifyReqs[i])
|
s.cancelNotify()
|
||||||
}
|
s.reqWG.Wait()
|
||||||
notifyReqs[i], _ = http.NewRequest("POST", url, bytes.NewReader(blob))
|
close(s.exitCh)
|
||||||
notifyReqs[i].Header.Set("Content-Type", "application/json")
|
}()
|
||||||
|
|
||||||
// Push the new work concurrently to all the remote nodes
|
ticker := time.NewTicker(5 * time.Second)
|
||||||
go func(req *http.Request, url string) {
|
defer ticker.Stop()
|
||||||
res, err := notifyClient.Do(req)
|
|
||||||
if err != nil {
|
for {
|
||||||
log.Warn("Failed to notify remote miner", "err", err)
|
select {
|
||||||
|
case work := <-s.workCh:
|
||||||
|
// Update current work with new received block.
|
||||||
|
// Note same work can be past twice, happens when changing CPU threads.
|
||||||
|
s.results = work.results
|
||||||
|
s.makeWork(work.block)
|
||||||
|
s.notifyWork()
|
||||||
|
|
||||||
|
case work := <-s.fetchWorkCh:
|
||||||
|
// Return current mining work to remote miner.
|
||||||
|
if s.currentBlock == nil {
|
||||||
|
work.errc <- errNoMiningWork
|
||||||
} else {
|
} else {
|
||||||
log.Trace("Notified remote miner", "miner", url, "hash", log.Lazy{Fn: func() common.Hash { return common.HexToHash(work[0]) }}, "target", work[2])
|
work.res <- s.currentWork
|
||||||
res.Body.Close()
|
|
||||||
}
|
}
|
||||||
}(notifyReqs[i], url)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// makeWork creates a work package for external miner.
|
|
||||||
//
|
|
||||||
// The work package consists of 3 strings:
|
|
||||||
// result[0], 32 bytes hex encoded current block header pow-hash
|
|
||||||
// result[1], 32 bytes hex encoded seed hash used for DAG
|
|
||||||
// result[2], 32 bytes hex encoded boundary condition ("target"), 2^256/difficulty
|
|
||||||
// result[3], hex encoded block number
|
|
||||||
makeWork := func(block *types.Block) {
|
|
||||||
hash := ethash.SealHash(block.Header())
|
|
||||||
|
|
||||||
currentWork[0] = hash.Hex()
|
case result := <-s.submitWorkCh:
|
||||||
currentWork[1] = common.BytesToHash(SeedHash(block.NumberU64())).Hex()
|
// Verify submitted PoW solution based on maintained mining blocks.
|
||||||
currentWork[2] = common.BytesToHash(new(big.Int).Div(two256, block.Difficulty()).Bytes()).Hex()
|
if s.submitWork(result.nonce, result.mixDigest, result.hash) {
|
||||||
currentWork[3] = hexutil.EncodeBig(block.Number())
|
result.errc <- nil
|
||||||
|
} else {
|
||||||
|
result.errc <- errInvalidSealResult
|
||||||
|
}
|
||||||
|
|
||||||
|
case result := <-s.submitRateCh:
|
||||||
|
// Trace remote sealer's hash rate by submitted value.
|
||||||
|
s.rates[result.id] = hashrate{rate: result.rate, ping: time.Now()}
|
||||||
|
close(result.done)
|
||||||
|
|
||||||
|
case req := <-s.fetchRateCh:
|
||||||
|
// Gather all hash rate submitted by remote sealer.
|
||||||
|
var total uint64
|
||||||
|
for _, rate := range s.rates {
|
||||||
|
// this could overflow
|
||||||
|
total += rate.rate
|
||||||
|
}
|
||||||
|
req <- total
|
||||||
|
|
||||||
|
case <-ticker.C:
|
||||||
|
// Clear stale submitted hash rate.
|
||||||
|
for id, rate := range s.rates {
|
||||||
|
if time.Since(rate.ping) > 10*time.Second {
|
||||||
|
delete(s.rates, id)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Clear stale pending blocks
|
||||||
|
if s.currentBlock != nil {
|
||||||
|
for hash, block := range s.works {
|
||||||
|
if block.NumberU64()+staleThreshold <= s.currentBlock.NumberU64() {
|
||||||
|
delete(s.works, hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
case <-s.requestExit:
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// makeWork creates a work package for external miner.
|
||||||
|
//
|
||||||
|
// The work package consists of 3 strings:
|
||||||
|
// result[0], 32 bytes hex encoded current block header pow-hash
|
||||||
|
// result[1], 32 bytes hex encoded seed hash used for DAG
|
||||||
|
// result[2], 32 bytes hex encoded boundary condition ("target"), 2^256/difficulty
|
||||||
|
// result[3], hex encoded block number
|
||||||
|
func (s *remoteSealer) makeWork(block *types.Block) {
|
||||||
|
hash := s.ethash.SealHash(block.Header())
|
||||||
|
s.currentWork[0] = hash.Hex()
|
||||||
|
s.currentWork[1] = common.BytesToHash(SeedHash(block.NumberU64())).Hex()
|
||||||
|
s.currentWork[2] = common.BytesToHash(new(big.Int).Div(two256, block.Difficulty()).Bytes()).Hex()
|
||||||
|
s.currentWork[3] = hexutil.EncodeBig(block.Number())
|
||||||
|
|
||||||
// Trace the seal work fetched by remote sealer.
|
// Trace the seal work fetched by remote sealer.
|
||||||
currentBlock = block
|
s.currentBlock = block
|
||||||
works[hash] = block
|
s.works[hash] = block
|
||||||
|
}
|
||||||
|
|
||||||
|
// notifyWork notifies all the specified mining endpoints of the availability of
|
||||||
|
// new work to be processed.
|
||||||
|
func (s *remoteSealer) notifyWork() {
|
||||||
|
work := s.currentWork
|
||||||
|
blob, _ := json.Marshal(work)
|
||||||
|
s.reqWG.Add(len(s.notifyURLs))
|
||||||
|
for _, url := range s.notifyURLs {
|
||||||
|
go s.sendNotification(s.notifyCtx, url, blob, work)
|
||||||
}
|
}
|
||||||
// submitWork verifies the submitted pow solution, returning
|
}
|
||||||
// whether the solution was accepted or not (not can be both a bad pow as well as
|
|
||||||
// any other error, like no pending work or stale mining result).
|
func (s *remoteSealer) sendNotification(ctx context.Context, url string, json []byte, work [4]string) {
|
||||||
submitWork := func(nonce types.BlockNonce, mixDigest common.Hash, sealhash common.Hash) bool {
|
defer s.reqWG.Done()
|
||||||
if currentBlock == nil {
|
|
||||||
log.Error("Pending work without block", "sealhash", sealhash)
|
req, err := http.NewRequest("POST", url, bytes.NewReader(json))
|
||||||
|
if err != nil {
|
||||||
|
s.ethash.config.Log.Warn("Can't create remote miner notification", "err", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
ctx, cancel := context.WithTimeout(ctx, remoteSealerTimeout)
|
||||||
|
defer cancel()
|
||||||
|
req = req.WithContext(ctx)
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
|
||||||
|
resp, err := http.DefaultClient.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
s.ethash.config.Log.Warn("Failed to notify remote miner", "err", err)
|
||||||
|
} else {
|
||||||
|
s.ethash.config.Log.Trace("Notified remote miner", "miner", url, "hash", work[0], "target", work[2])
|
||||||
|
resp.Body.Close()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// submitWork verifies the submitted pow solution, returning
|
||||||
|
// whether the solution was accepted or not (not can be both a bad pow as well as
|
||||||
|
// any other error, like no pending work or stale mining result).
|
||||||
|
func (s *remoteSealer) submitWork(nonce types.BlockNonce, mixDigest common.Hash, sealhash common.Hash) bool {
|
||||||
|
if s.currentBlock == nil {
|
||||||
|
s.ethash.config.Log.Error("Pending work without block", "sealhash", sealhash)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
// Make sure the work submitted is present
|
// Make sure the work submitted is present
|
||||||
block := works[sealhash]
|
block := s.works[sealhash]
|
||||||
if block == nil {
|
if block == nil {
|
||||||
log.Warn("Work submitted but none pending", "sealhash", sealhash, "curnumber", currentBlock.NumberU64())
|
s.ethash.config.Log.Warn("Work submitted but none pending", "sealhash", sealhash, "curnumber", s.currentBlock.NumberU64())
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
// Verify the correctness of submitted result.
|
// Verify the correctness of submitted result.
|
||||||
@ -268,104 +407,34 @@ func (ethash *Ethash) remote(notify []string, noverify bool) {
|
|||||||
header.MixDigest = mixDigest
|
header.MixDigest = mixDigest
|
||||||
|
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
if !noverify {
|
if !s.noverify {
|
||||||
if err := ethash.verifySeal(nil, header, true); err != nil {
|
if err := s.ethash.verifySeal(nil, header, true); err != nil {
|
||||||
log.Warn("Invalid proof-of-work submitted", "sealhash", sealhash, "elapsed", common.PrettyDuration(time.Since(start)), "err", err)
|
s.ethash.config.Log.Warn("Invalid proof-of-work submitted", "sealhash", sealhash, "elapsed", common.PrettyDuration(time.Since(start)), "err", err)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Make sure the result channel is assigned.
|
// Make sure the result channel is assigned.
|
||||||
if results == nil {
|
if s.results == nil {
|
||||||
log.Warn("Ethash result channel is empty, submitted mining result is rejected")
|
s.ethash.config.Log.Warn("Ethash result channel is empty, submitted mining result is rejected")
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
log.Trace("Verified correct proof-of-work", "sealhash", sealhash, "elapsed", common.PrettyDuration(time.Since(start)))
|
s.ethash.config.Log.Trace("Verified correct proof-of-work", "sealhash", sealhash, "elapsed", common.PrettyDuration(time.Since(start)))
|
||||||
|
|
||||||
// Solutions seems to be valid, return to the miner and notify acceptance.
|
// Solutions seems to be valid, return to the miner and notify acceptance.
|
||||||
solution := block.WithSeal(header)
|
solution := block.WithSeal(header)
|
||||||
|
|
||||||
// The submitted solution is within the scope of acceptance.
|
// The submitted solution is within the scope of acceptance.
|
||||||
if solution.NumberU64()+staleThreshold > currentBlock.NumberU64() {
|
if solution.NumberU64()+staleThreshold > s.currentBlock.NumberU64() {
|
||||||
select {
|
select {
|
||||||
case results <- solution:
|
case s.results <- solution:
|
||||||
log.Debug("Work submitted is acceptable", "number", solution.NumberU64(), "sealhash", sealhash, "hash", solution.Hash())
|
s.ethash.config.Log.Debug("Work submitted is acceptable", "number", solution.NumberU64(), "sealhash", sealhash, "hash", solution.Hash())
|
||||||
return true
|
return true
|
||||||
default:
|
default:
|
||||||
log.Warn("Sealing result is not read by miner", "mode", "remote", "sealhash", sealhash)
|
s.ethash.config.Log.Warn("Sealing result is not read by miner", "mode", "remote", "sealhash", sealhash)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// The submitted block is too old to accept, drop it.
|
// The submitted block is too old to accept, drop it.
|
||||||
log.Warn("Work submitted is too old", "number", solution.NumberU64(), "sealhash", sealhash, "hash", solution.Hash())
|
s.ethash.config.Log.Warn("Work submitted is too old", "number", solution.NumberU64(), "sealhash", sealhash, "hash", solution.Hash())
|
||||||
return false
|
return false
|
||||||
}
|
|
||||||
|
|
||||||
ticker := time.NewTicker(5 * time.Second)
|
|
||||||
defer ticker.Stop()
|
|
||||||
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case work := <-ethash.workCh:
|
|
||||||
// Update current work with new received block.
|
|
||||||
// Note same work can be past twice, happens when changing CPU threads.
|
|
||||||
results = work.results
|
|
||||||
|
|
||||||
makeWork(work.block)
|
|
||||||
|
|
||||||
// Notify and requested URLs of the new work availability
|
|
||||||
notifyWork()
|
|
||||||
|
|
||||||
case work := <-ethash.fetchWorkCh:
|
|
||||||
// Return current mining work to remote miner.
|
|
||||||
if currentBlock == nil {
|
|
||||||
work.errc <- errNoMiningWork
|
|
||||||
} else {
|
|
||||||
work.res <- currentWork
|
|
||||||
}
|
|
||||||
|
|
||||||
case result := <-ethash.submitWorkCh:
|
|
||||||
// Verify submitted PoW solution based on maintained mining blocks.
|
|
||||||
if submitWork(result.nonce, result.mixDigest, result.hash) {
|
|
||||||
result.errc <- nil
|
|
||||||
} else {
|
|
||||||
result.errc <- errInvalidSealResult
|
|
||||||
}
|
|
||||||
|
|
||||||
case result := <-ethash.submitRateCh:
|
|
||||||
// Trace remote sealer's hash rate by submitted value.
|
|
||||||
rates[result.id] = hashrate{rate: result.rate, ping: time.Now()}
|
|
||||||
close(result.done)
|
|
||||||
|
|
||||||
case req := <-ethash.fetchRateCh:
|
|
||||||
// Gather all hash rate submitted by remote sealer.
|
|
||||||
var total uint64
|
|
||||||
for _, rate := range rates {
|
|
||||||
// this could overflow
|
|
||||||
total += rate.rate
|
|
||||||
}
|
|
||||||
req <- total
|
|
||||||
|
|
||||||
case <-ticker.C:
|
|
||||||
// Clear stale submitted hash rate.
|
|
||||||
for id, rate := range rates {
|
|
||||||
if time.Since(rate.ping) > 10*time.Second {
|
|
||||||
delete(rates, id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Clear stale pending blocks
|
|
||||||
if currentBlock != nil {
|
|
||||||
for hash, block := range works {
|
|
||||||
if block.NumberU64()+staleThreshold <= currentBlock.NumberU64() {
|
|
||||||
delete(works, hash)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
case errc := <-ethash.exitCh:
|
|
||||||
// Exit remote loop if ethash is closed and return relevant error.
|
|
||||||
errc <- nil
|
|
||||||
log.Trace("Ethash remote sealer is exiting")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -20,59 +20,39 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"math/big"
|
"math/big"
|
||||||
"net"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"net/http/httptest"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
|
"github.com/ethereum/go-ethereum/internal/testlog"
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Tests whether remote HTTP servers are correctly notified of new work.
|
// Tests whether remote HTTP servers are correctly notified of new work.
|
||||||
func TestRemoteNotify(t *testing.T) {
|
func TestRemoteNotify(t *testing.T) {
|
||||||
// Start a simple webserver to capture notifications
|
// Start a simple web server to capture notifications.
|
||||||
sink := make(chan [3]string)
|
sink := make(chan [3]string)
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
|
||||||
server := &http.Server{
|
|
||||||
Handler: http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
|
|
||||||
blob, err := ioutil.ReadAll(req.Body)
|
blob, err := ioutil.ReadAll(req.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to read miner notification: %v", err)
|
t.Errorf("failed to read miner notification: %v", err)
|
||||||
}
|
}
|
||||||
var work [3]string
|
var work [3]string
|
||||||
if err := json.Unmarshal(blob, &work); err != nil {
|
if err := json.Unmarshal(blob, &work); err != nil {
|
||||||
t.Fatalf("failed to unmarshal miner notification: %v", err)
|
t.Errorf("failed to unmarshal miner notification: %v", err)
|
||||||
}
|
}
|
||||||
sink <- work
|
sink <- work
|
||||||
}),
|
}))
|
||||||
}
|
defer server.Close()
|
||||||
// Open a custom listener to extract its local address
|
|
||||||
listener, err := net.Listen("tcp", "localhost:0")
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to open notification server: %v", err)
|
|
||||||
}
|
|
||||||
defer listener.Close()
|
|
||||||
|
|
||||||
go server.Serve(listener)
|
// Create the custom ethash engine.
|
||||||
|
ethash := NewTester([]string{server.URL}, false)
|
||||||
// Wait for server to start listening
|
|
||||||
var tries int
|
|
||||||
for tries = 0; tries < 10; tries++ {
|
|
||||||
conn, _ := net.DialTimeout("tcp", listener.Addr().String(), 1*time.Second)
|
|
||||||
if conn != nil {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if tries == 10 {
|
|
||||||
t.Fatal("tcp listener not ready for more than 10 seconds")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create the custom ethash engine
|
|
||||||
ethash := NewTester([]string{"http://" + listener.Addr().String()}, false)
|
|
||||||
defer ethash.Close()
|
defer ethash.Close()
|
||||||
|
|
||||||
// Stream a work task and ensure the notification bubbles out
|
// Stream a work task and ensure the notification bubbles out.
|
||||||
header := &types.Header{Number: big.NewInt(1), Difficulty: big.NewInt(100)}
|
header := &types.Header{Number: big.NewInt(1), Difficulty: big.NewInt(100)}
|
||||||
block := types.NewBlockWithHeader(header)
|
block := types.NewBlockWithHeader(header)
|
||||||
|
|
||||||
@ -97,46 +77,37 @@ func TestRemoteNotify(t *testing.T) {
|
|||||||
// Tests that pushing work packages fast to the miner doesn't cause any data race
|
// Tests that pushing work packages fast to the miner doesn't cause any data race
|
||||||
// issues in the notifications.
|
// issues in the notifications.
|
||||||
func TestRemoteMultiNotify(t *testing.T) {
|
func TestRemoteMultiNotify(t *testing.T) {
|
||||||
// Start a simple webserver to capture notifications
|
// Start a simple web server to capture notifications.
|
||||||
sink := make(chan [3]string, 64)
|
sink := make(chan [3]string, 64)
|
||||||
|
server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
|
||||||
server := &http.Server{
|
|
||||||
Handler: http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
|
|
||||||
blob, err := ioutil.ReadAll(req.Body)
|
blob, err := ioutil.ReadAll(req.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to read miner notification: %v", err)
|
t.Errorf("failed to read miner notification: %v", err)
|
||||||
}
|
}
|
||||||
var work [3]string
|
var work [3]string
|
||||||
if err := json.Unmarshal(blob, &work); err != nil {
|
if err := json.Unmarshal(blob, &work); err != nil {
|
||||||
t.Fatalf("failed to unmarshal miner notification: %v", err)
|
t.Errorf("failed to unmarshal miner notification: %v", err)
|
||||||
}
|
}
|
||||||
sink <- work
|
sink <- work
|
||||||
}),
|
}))
|
||||||
}
|
defer server.Close()
|
||||||
// Open a custom listener to extract its local address
|
|
||||||
listener, err := net.Listen("tcp", "localhost:0")
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to open notification server: %v", err)
|
|
||||||
}
|
|
||||||
defer listener.Close()
|
|
||||||
|
|
||||||
go server.Serve(listener)
|
// Create the custom ethash engine.
|
||||||
|
ethash := NewTester([]string{server.URL}, false)
|
||||||
// Create the custom ethash engine
|
ethash.config.Log = testlog.Logger(t, log.LvlWarn)
|
||||||
ethash := NewTester([]string{"http://" + listener.Addr().String()}, false)
|
|
||||||
defer ethash.Close()
|
defer ethash.Close()
|
||||||
|
|
||||||
// Stream a lot of work task and ensure all the notifications bubble out
|
// Stream a lot of work task and ensure all the notifications bubble out.
|
||||||
for i := 0; i < cap(sink); i++ {
|
for i := 0; i < cap(sink); i++ {
|
||||||
header := &types.Header{Number: big.NewInt(int64(i)), Difficulty: big.NewInt(100)}
|
header := &types.Header{Number: big.NewInt(int64(i)), Difficulty: big.NewInt(100)}
|
||||||
block := types.NewBlockWithHeader(header)
|
block := types.NewBlockWithHeader(header)
|
||||||
|
|
||||||
ethash.Seal(nil, block, nil, nil)
|
ethash.Seal(nil, block, nil, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
for i := 0; i < cap(sink); i++ {
|
for i := 0; i < cap(sink); i++ {
|
||||||
select {
|
select {
|
||||||
case <-sink:
|
case <-sink:
|
||||||
case <-time.After(3 * time.Second):
|
case <-time.After(10 * time.Second):
|
||||||
t.Fatalf("notification %d timed out", i)
|
t.Fatalf("notification %d timed out", i)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -206,10 +177,10 @@ func TestStaleSubmission(t *testing.T) {
|
|||||||
select {
|
select {
|
||||||
case res := <-results:
|
case res := <-results:
|
||||||
if res.Header().Nonce != fakeNonce {
|
if res.Header().Nonce != fakeNonce {
|
||||||
t.Errorf("case %d block nonce mismatch, want %s, get %s", id+1, fakeNonce, res.Header().Nonce)
|
t.Errorf("case %d block nonce mismatch, want %x, get %x", id+1, fakeNonce, res.Header().Nonce)
|
||||||
}
|
}
|
||||||
if res.Header().MixDigest != fakeDigest {
|
if res.Header().MixDigest != fakeDigest {
|
||||||
t.Errorf("case %d block digest mismatch, want %s, get %s", id+1, fakeDigest, res.Header().MixDigest)
|
t.Errorf("case %d block digest mismatch, want %x, get %x", id+1, fakeDigest, res.Header().MixDigest)
|
||||||
}
|
}
|
||||||
if res.Header().Difficulty.Uint64() != c.headers[c.submitIndex].Difficulty.Uint64() {
|
if res.Header().Difficulty.Uint64() != c.headers[c.submitIndex].Difficulty.Uint64() {
|
||||||
t.Errorf("case %d block difficulty mismatch, want %d, get %d", id+1, c.headers[c.submitIndex].Difficulty, res.Header().Difficulty)
|
t.Errorf("case %d block difficulty mismatch, want %d, get %d", id+1, c.headers[c.submitIndex].Difficulty, res.Header().Difficulty)
|
||||||
|
Loading…
Reference in New Issue
Block a user