2018-06-20 07:06:27 -05:00
|
|
|
// Copyright 2018 The go-ethereum Authors
|
|
|
|
// This file is part of go-ethereum.
|
|
|
|
//
|
|
|
|
// go-ethereum is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// go-ethereum is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
|
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
2019-02-07 07:38:32 -06:00
|
|
|
"bytes"
|
2019-02-18 05:05:22 -06:00
|
|
|
"context"
|
2018-06-20 07:06:27 -05:00
|
|
|
"fmt"
|
2019-02-18 05:05:22 -06:00
|
|
|
"io/ioutil"
|
2018-12-11 02:21:58 -06:00
|
|
|
"math/rand"
|
2019-02-18 05:05:22 -06:00
|
|
|
"os"
|
2019-03-08 01:52:05 -06:00
|
|
|
"strings"
|
2018-06-20 07:06:27 -05:00
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/log"
|
2018-12-11 02:21:58 -06:00
|
|
|
"github.com/ethereum/go-ethereum/metrics"
|
2019-02-18 05:05:22 -06:00
|
|
|
"github.com/ethereum/go-ethereum/rpc"
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/api"
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
2019-02-07 07:38:32 -06:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/testutil"
|
2018-06-20 07:06:27 -05:00
|
|
|
"github.com/pborman/uuid"
|
|
|
|
|
|
|
|
cli "gopkg.in/urfave/cli.v1"
|
|
|
|
)
|
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
func uploadAndSyncCmd(ctx *cli.Context, tuid string) error {
|
|
|
|
randomBytes := testutil.RandomBytes(seed, filesize*1000)
|
2019-01-30 02:46:44 -06:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
errc := make(chan error)
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
go func() {
|
|
|
|
errc <- uplaodAndSync(ctx, randomBytes, tuid)
|
|
|
|
}()
|
|
|
|
|
|
|
|
select {
|
|
|
|
case err := <-errc:
|
|
|
|
if err != nil {
|
|
|
|
metrics.GetOrRegisterCounter(fmt.Sprintf("%s.fail", commandName), nil).Inc(1)
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
case <-time.After(time.Duration(timeout) * time.Second):
|
|
|
|
metrics.GetOrRegisterCounter(fmt.Sprintf("%s.timeout", commandName), nil).Inc(1)
|
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
e := fmt.Errorf("timeout after %v sec", timeout)
|
2019-02-07 07:38:32 -06:00
|
|
|
// trigger debug functionality on randomBytes
|
2019-02-18 05:05:22 -06:00
|
|
|
err := trackChunks(randomBytes[:])
|
|
|
|
if err != nil {
|
|
|
|
e = fmt.Errorf("%v; triggerChunkDebug failed: %v", e, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return e
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func trackChunks(testData []byte) error {
|
2019-03-11 05:45:34 -05:00
|
|
|
log.Warn("Test timed out, running chunk debug sequence")
|
2019-02-07 07:38:32 -06:00
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
addrs, err := getAllRefs(testData)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2019-02-07 07:38:32 -06:00
|
|
|
}
|
2019-02-18 05:05:22 -06:00
|
|
|
|
2019-03-08 01:52:05 -06:00
|
|
|
for i, ref := range addrs {
|
|
|
|
log.Trace(fmt.Sprintf("ref %d", i), "ref", ref)
|
|
|
|
}
|
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
for _, host := range hosts {
|
|
|
|
httpHost := fmt.Sprintf("ws://%s:%d", host, 8546)
|
2019-03-08 01:52:05 -06:00
|
|
|
|
|
|
|
hostChunks := []string{}
|
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
rpcClient, err := rpc.Dial(httpHost)
|
|
|
|
if err != nil {
|
2019-03-11 05:45:34 -05:00
|
|
|
log.Error("Error dialing host", "err", err)
|
2019-02-18 05:05:22 -06:00
|
|
|
return err
|
|
|
|
}
|
2019-03-11 05:45:34 -05:00
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
var hasInfo []api.HasInfo
|
|
|
|
err = rpcClient.Call(&hasInfo, "bzz_has", addrs)
|
|
|
|
if err != nil {
|
2019-03-11 05:45:34 -05:00
|
|
|
log.Error("Error calling host", "err", err)
|
2019-02-18 05:05:22 -06:00
|
|
|
return err
|
|
|
|
}
|
2019-03-08 01:52:05 -06:00
|
|
|
|
2019-03-11 05:45:34 -05:00
|
|
|
count := 0
|
|
|
|
for _, info := range hasInfo {
|
2019-03-08 01:52:05 -06:00
|
|
|
if info.Has {
|
|
|
|
hostChunks = append(hostChunks, "1")
|
|
|
|
} else {
|
|
|
|
hostChunks = append(hostChunks, "0")
|
2019-02-18 05:05:22 -06:00
|
|
|
count++
|
|
|
|
}
|
|
|
|
}
|
2019-03-11 05:45:34 -05:00
|
|
|
|
2019-02-18 05:05:22 -06:00
|
|
|
if count == 0 {
|
2019-03-08 01:52:05 -06:00
|
|
|
log.Info("host reported to have all chunks", "host", host)
|
2019-02-18 05:05:22 -06:00
|
|
|
}
|
2019-03-08 01:52:05 -06:00
|
|
|
|
|
|
|
log.Trace("chunks", "chunks", strings.Join(hostChunks, ""), "host", host)
|
2019-02-18 05:05:22 -06:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func getAllRefs(testData []byte) (storage.AddressCollection, error) {
|
|
|
|
log.Trace("Getting all references for given root hash")
|
|
|
|
datadir, err := ioutil.TempDir("", "chunk-debug")
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("unable to create temp dir: %v", err)
|
|
|
|
}
|
|
|
|
defer os.RemoveAll(datadir)
|
|
|
|
fileStore, err := storage.NewLocalFileStore(datadir, make([]byte, 32))
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Duration(trackTimeout)*time.Second)
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
reader := bytes.NewReader(testData)
|
|
|
|
return fileStore.GetAllReferences(ctx, reader, false)
|
2019-02-07 07:38:32 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
func uplaodAndSync(c *cli.Context, randomBytes []byte, tuid string) error {
|
|
|
|
log.Info("uploading to "+httpEndpoint(hosts[0])+" and syncing", "tuid", tuid, "seed", seed)
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2018-12-11 02:21:58 -06:00
|
|
|
t1 := time.Now()
|
2019-02-07 07:38:32 -06:00
|
|
|
hash, err := upload(randomBytes, httpEndpoint(hosts[0]))
|
2018-06-20 07:06:27 -05:00
|
|
|
if err != nil {
|
|
|
|
log.Error(err.Error())
|
|
|
|
return err
|
|
|
|
}
|
2019-02-07 07:38:32 -06:00
|
|
|
t2 := time.Since(t1)
|
|
|
|
metrics.GetOrRegisterResettingTimer("upload-and-sync.upload-time", nil).Update(t2)
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
fhash, err := digest(bytes.NewReader(randomBytes))
|
|
|
|
if err != nil {
|
|
|
|
log.Error(err.Error())
|
|
|
|
return err
|
|
|
|
}
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
log.Info("uploaded successfully", "tuid", tuid, "hash", hash, "took", t2, "digest", fmt.Sprintf("%x", fhash))
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2018-12-11 02:21:58 -06:00
|
|
|
time.Sleep(time.Duration(syncDelay) * time.Second)
|
2018-06-20 07:06:27 -05:00
|
|
|
|
|
|
|
wg := sync.WaitGroup{}
|
2018-12-11 02:21:58 -06:00
|
|
|
if single {
|
2019-02-07 07:38:32 -06:00
|
|
|
randIndex := 1 + rand.Intn(len(hosts)-1)
|
2018-06-20 07:06:27 -05:00
|
|
|
ruid := uuid.New()[:8]
|
|
|
|
wg.Add(1)
|
|
|
|
go func(endpoint string, ruid string) {
|
|
|
|
for {
|
2018-12-11 02:21:58 -06:00
|
|
|
start := time.Now()
|
2019-02-07 07:38:32 -06:00
|
|
|
err := fetch(hash, endpoint, fhash, ruid, tuid)
|
2018-06-20 07:06:27 -05:00
|
|
|
if err != nil {
|
|
|
|
continue
|
|
|
|
}
|
2019-02-07 07:38:32 -06:00
|
|
|
ended := time.Since(start)
|
2018-06-20 07:06:27 -05:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
metrics.GetOrRegisterResettingTimer("upload-and-sync.single.fetch-time", nil).Update(ended)
|
|
|
|
log.Info("fetch successful", "tuid", tuid, "ruid", ruid, "took", ended, "endpoint", endpoint)
|
2018-06-20 07:06:27 -05:00
|
|
|
wg.Done()
|
|
|
|
return
|
|
|
|
}
|
2019-02-07 07:38:32 -06:00
|
|
|
}(httpEndpoint(hosts[randIndex]), ruid)
|
2018-12-11 02:21:58 -06:00
|
|
|
} else {
|
2019-02-07 07:38:32 -06:00
|
|
|
for _, endpoint := range hosts[1:] {
|
2018-12-11 02:21:58 -06:00
|
|
|
ruid := uuid.New()[:8]
|
|
|
|
wg.Add(1)
|
|
|
|
go func(endpoint string, ruid string) {
|
|
|
|
for {
|
|
|
|
start := time.Now()
|
2019-02-07 07:38:32 -06:00
|
|
|
err := fetch(hash, endpoint, fhash, ruid, tuid)
|
2018-12-11 02:21:58 -06:00
|
|
|
if err != nil {
|
|
|
|
continue
|
|
|
|
}
|
2019-02-07 07:38:32 -06:00
|
|
|
ended := time.Since(start)
|
2018-12-11 02:21:58 -06:00
|
|
|
|
2019-02-07 07:38:32 -06:00
|
|
|
metrics.GetOrRegisterResettingTimer("upload-and-sync.each.fetch-time", nil).Update(ended)
|
|
|
|
log.Info("fetch successful", "tuid", tuid, "ruid", ruid, "took", ended, "endpoint", endpoint)
|
2018-12-11 02:21:58 -06:00
|
|
|
wg.Done()
|
|
|
|
return
|
|
|
|
}
|
2019-02-07 07:38:32 -06:00
|
|
|
}(httpEndpoint(endpoint), ruid)
|
2018-12-11 02:21:58 -06:00
|
|
|
}
|
2018-06-20 07:06:27 -05:00
|
|
|
}
|
|
|
|
wg.Wait()
|
2019-02-07 07:38:32 -06:00
|
|
|
log.Info("all hosts synced random file successfully")
|
2018-06-20 07:06:27 -05:00
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|