forked from cerc-io/plugeth
cmd/swarm: hashes command (#19008)
This commit is contained in:
parent
85b3b1c8d6
commit
7f55b0cbd8
59
cmd/swarm/explore.go
Normal file
59
cmd/swarm/explore.go
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
// Copyright 2019 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
// Command bzzhash computes a swarm tree hash.
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
|
"gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
var hashesCommand = cli.Command{
|
||||||
|
Action: hashes,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "hashes",
|
||||||
|
Usage: "print all hashes of a file to STDOUT",
|
||||||
|
ArgsUsage: "<file>",
|
||||||
|
Description: "Prints all hashes of a file to STDOUT",
|
||||||
|
}
|
||||||
|
|
||||||
|
func hashes(ctx *cli.Context) {
|
||||||
|
args := ctx.Args()
|
||||||
|
if len(args) < 1 {
|
||||||
|
utils.Fatalf("Usage: swarm hashes <file name>")
|
||||||
|
}
|
||||||
|
f, err := os.Open(args[0])
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error opening file " + args[1])
|
||||||
|
}
|
||||||
|
defer f.Close()
|
||||||
|
|
||||||
|
fileStore := storage.NewFileStore(&storage.FakeChunkStore{}, storage.NewFileStoreParams())
|
||||||
|
refs, err := fileStore.GetAllReferences(context.TODO(), f, false)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("%v\n", err)
|
||||||
|
} else {
|
||||||
|
for _, r := range refs {
|
||||||
|
fmt.Println(r.String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@ -142,6 +142,8 @@ func init() {
|
|||||||
dbCommand,
|
dbCommand,
|
||||||
// See config.go
|
// See config.go
|
||||||
DumpConfigCommand,
|
DumpConfigCommand,
|
||||||
|
// hashesCommand
|
||||||
|
hashesCommand,
|
||||||
}
|
}
|
||||||
|
|
||||||
// append a hidden help subcommand to all commands that have subcommands
|
// append a hidden help subcommand to all commands that have subcommands
|
||||||
|
@ -75,7 +75,7 @@ func NewFileStore(store ChunkStore, params *FileStoreParams) *FileStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Public API. Main entry point for document retrieval directly. Used by the
|
// Retrieve is a public API. Main entry point for document retrieval directly. Used by the
|
||||||
// FS-aware API and httpaccess
|
// FS-aware API and httpaccess
|
||||||
// Chunk retrieval blocks on netStore requests with a timeout so reader will
|
// Chunk retrieval blocks on netStore requests with a timeout so reader will
|
||||||
// report error if retrieval of chunks within requested range time out.
|
// report error if retrieval of chunks within requested range time out.
|
||||||
@ -87,7 +87,7 @@ func (f *FileStore) Retrieve(ctx context.Context, addr Address) (reader *LazyChu
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Public API. Main entry point for document storage directly. Used by the
|
// Store is a public API. Main entry point for document storage directly. Used by the
|
||||||
// FS-aware API and httpaccess
|
// FS-aware API and httpaccess
|
||||||
func (f *FileStore) Store(ctx context.Context, data io.Reader, size int64, toEncrypt bool) (addr Address, wait func(context.Context) error, err error) {
|
func (f *FileStore) Store(ctx context.Context, data io.Reader, size int64, toEncrypt bool) (addr Address, wait func(context.Context) error, err error) {
|
||||||
putter := NewHasherStore(f.ChunkStore, f.hashFunc, toEncrypt)
|
putter := NewHasherStore(f.ChunkStore, f.hashFunc, toEncrypt)
|
||||||
@ -98,7 +98,7 @@ func (f *FileStore) HashSize() int {
|
|||||||
return f.hashFunc().Size()
|
return f.hashFunc().Size()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Public API. This endpoint returns all chunk hashes (only) for a given file
|
// GetAllReferences is a public API. This endpoint returns all chunk hashes (only) for a given file
|
||||||
func (f *FileStore) GetAllReferences(ctx context.Context, data io.Reader, toEncrypt bool) (addrs AddressCollection, err error) {
|
func (f *FileStore) GetAllReferences(ctx context.Context, data io.Reader, toEncrypt bool) (addrs AddressCollection, err error) {
|
||||||
// create a special kind of putter, which only will store the references
|
// create a special kind of putter, which only will store the references
|
||||||
putter := &HashExplorer{
|
putter := &HashExplorer{
|
||||||
|
@ -190,22 +190,18 @@ func TestGetAllReferences(t *testing.T) {
|
|||||||
}
|
}
|
||||||
fileStore := NewFileStore(localStore, NewFileStoreParams())
|
fileStore := NewFileStore(localStore, NewFileStoreParams())
|
||||||
|
|
||||||
checkRefs := func(dataSize int, expectedLen int) {
|
// testRuns[i] and expectedLen[i] are dataSize and expected length respectively
|
||||||
slice := testutil.RandomBytes(1, dataSize)
|
testRuns := []int{1024, 8192, 16000, 30000, 1000000}
|
||||||
|
expectedLens := []int{1, 3, 5, 9, 248}
|
||||||
|
for i, r := range testRuns {
|
||||||
|
slice := testutil.RandomBytes(1, r)
|
||||||
|
|
||||||
addrs, err := fileStore.GetAllReferences(context.Background(), bytes.NewReader(slice), false)
|
addrs, err := fileStore.GetAllReferences(context.Background(), bytes.NewReader(slice), false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
if len(addrs) != expectedLen {
|
if len(addrs) != expectedLens[i] {
|
||||||
t.Fatalf("Expected reference array length to be %d, but is %d", expectedLen, len(addrs))
|
t.Fatalf("Expected reference array length to be %d, but is %d", expectedLens[i], len(addrs))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// testRuns[i] and expectedLen[i] are dataSize and expected length respectively
|
|
||||||
testRuns := []int{1024, 8192, 16000, 30000, 1000000}
|
|
||||||
expectedLens := []int{1, 3, 5, 9, 248}
|
|
||||||
for i, r := range testRuns {
|
|
||||||
checkRefs(r, expectedLens[i])
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user