eth, les, swarm: fix go vet issues sufraced by log15

This commit is contained in:
Péter Szilágyi 2017-02-22 14:56:09 +02:00
parent d4fd06c3dc
commit 61e6bb1247
No known key found for this signature in database
GPG Key ID: E9AE538CEDF8293D
8 changed files with 10 additions and 10 deletions

@ -938,7 +938,7 @@ func (d *Downloader) fetchNodeData() error {
} }
if err != nil { if err != nil {
// If the node data processing failed, the root hash is very wrong, abort // If the node data processing failed, the root hash is very wrong, abort
log.Error(fmt.Sprintf("peer %d: state processing failed: %v", packet.PeerId(), err)) log.Error(fmt.Sprintf("peer %s: state processing failed: %v", packet.PeerId(), err))
d.cancel() d.cancel()
return return
} }

@ -441,7 +441,7 @@ func (pm *ProtocolManager) handleMsg(p *peer) error {
// Block header query, collect the requested headers and reply // Block header query, collect the requested headers and reply
case AnnounceMsg: case AnnounceMsg:
log.Debug(fmt.Sprint("<=== AnnounceMsg from peer %v:", p.id)) log.Debug(fmt.Sprintf("<=== AnnounceMsg from peer %v:", p.id))
var req announceData var req announceData
if err := msg.Decode(&req); err != nil { if err := msg.Decode(&req); err != nil {

@ -84,7 +84,7 @@ func (self *Api) Resolve(hostPort string, nameresolver bool) (storage.Key, error
err = ErrResolve(err) err = ErrResolve(err)
log.Warn(fmt.Sprintf("DNS error : %v", err)) log.Warn(fmt.Sprintf("DNS error : %v", err))
} }
log.Trace(fmt.Sprintf("host lookup: %v -> %v", err)) log.Trace(fmt.Sprintf("host lookup: %v -> %v", hostPort, contentHash))
return contentHash[:], err return contentHash[:], err
} }
func Parse(uri string) (hostPort, path string) { func Parse(uri string) (hostPort, path string) {

@ -211,7 +211,7 @@ only add if less than requesterCount peers forwarded the same request id so far
note this is done irrespective of status (searching or found) note this is done irrespective of status (searching or found)
*/ */
func (self *Depo) addRequester(rs *storage.RequestStatus, req *retrieveRequestMsgData) { func (self *Depo) addRequester(rs *storage.RequestStatus, req *retrieveRequestMsgData) {
log.Trace(fmt.Sprintf("Depo.addRequester: key %v - add peer to req.Id %v", req.Key.Log(), req.from, req.Id)) log.Trace(fmt.Sprintf("Depo.addRequester: key %v - add peer to req.Id %v", req.Key.Log(), req.Id))
list := rs.Requesters[req.Id] list := rs.Requesters[req.Id]
rs.Requesters[req.Id] = append(list, req) rs.Requesters[req.Id] = append(list, req)
} }

@ -120,7 +120,7 @@ func (self *KadDb) add(nrs []*NodeRecord, proximityBin func(Address) int) {
copy(newnodes[:], nodes[:dbcursor]) copy(newnodes[:], nodes[:dbcursor])
newnodes[dbcursor] = node newnodes[dbcursor] = node
copy(newnodes[dbcursor+1:], nodes[dbcursor:]) copy(newnodes[dbcursor+1:], nodes[dbcursor:])
log.Trace(fmt.Sprintf("new nodes: %v (keys: %v)\nnodes: %v", newnodes, nodes)) log.Trace(fmt.Sprintf("new nodes: %v, nodes: %v", newnodes, nodes))
self.Nodes[index] = newnodes self.Nodes[index] = newnodes
n++ n++
} }
@ -294,7 +294,7 @@ func (self *KadDb) save(path string, cb func(*NodeRecord, Node)) error {
} }
err = ioutil.WriteFile(path, data, os.ModePerm) err = ioutil.WriteFile(path, data, os.ModePerm)
if err != nil { if err != nil {
log.Warn(fmt.Sprintf("unable to save kaddb with %v nodes to %v: err", n, path, err)) log.Warn(fmt.Sprintf("unable to save kaddb with %v nodes to %v: %v", n, path, err))
} else { } else {
log.Info(fmt.Sprintf("saved kaddb with %v nodes to %v", n, path)) log.Info(fmt.Sprintf("saved kaddb with %v nodes to %v", n, path))
} }

@ -217,7 +217,7 @@ LOOP:
// if just switched to db mode and not quitting, then launch dbRead // if just switched to db mode and not quitting, then launch dbRead
// in a parallel go routine to send deliveries from db // in a parallel go routine to send deliveries from db
if inDb == 0 && quit != nil { if inDb == 0 && quit != nil {
log.Trace(fmt.Sprintf("syncDb[%v/%v] start dbRead")) log.Trace(fmt.Sprintf("syncDb[%v/%v] start dbRead", self.key.Log(), self.priority))
go self.dbRead(true, counter, deliver) go self.dbRead(true, counter, deliver)
} }
inDb++ inDb++

@ -480,7 +480,7 @@ LOOP:
stateCopy := *state stateCopy := *state
err := self.unsyncedKeys(unsynced, &stateCopy) err := self.unsyncedKeys(unsynced, &stateCopy)
if err != nil { if err != nil {
log.Warn(fmt.Sprintf("syncer[%v]: unable to send unsynced keys: %v", err)) log.Warn(fmt.Sprintf("syncer[%v]: unable to send unsynced keys: %v", self.key.Log(), err))
} }
self.state = state self.state = state
log.Debug(fmt.Sprintf("syncer[%v]: --> %v keys sent: (total: %v (%v), history: %v), sent sync state: %v", self.key.Log(), len(unsynced), keyCounts, keyCount, historyCnt, stateCopy)) log.Debug(fmt.Sprintf("syncer[%v]: --> %v keys sent: (total: %v (%v), history: %v), sent sync state: %v", self.key.Log(), len(unsynced), keyCounts, keyCount, historyCnt, stateCopy))
@ -553,7 +553,7 @@ LOOP:
log.Trace(fmt.Sprintf("syncer[%v]: (priority %v): request %v (synced = %v)", self.key.Log(), priority, req, state.Synced)) log.Trace(fmt.Sprintf("syncer[%v]: (priority %v): request %v (synced = %v)", self.key.Log(), priority, req, state.Synced))
unsynced = append(unsynced, sreq) unsynced = append(unsynced, sreq)
} else { } else {
log.Warn(fmt.Sprintf("syncer[%v]: (priority %v): error creating request for %v: %v)", self.key.Log(), priority, req, state.Synced, err)) log.Warn(fmt.Sprintf("syncer[%v]: (priority %v): error creating request for %v: %v)", self.key.Log(), priority, req, err))
} }
} }

@ -197,7 +197,7 @@ func (self *Swarm) Start(net *p2p.Server) error {
log.Debug(fmt.Sprintf("Swarm http proxy started on port: %v", self.config.Port)) log.Debug(fmt.Sprintf("Swarm http proxy started on port: %v", self.config.Port))
if self.corsString != "" { if self.corsString != "" {
log.Debug(fmt.Sprintf("Swarm http proxy started with corsdomain:", self.corsString)) log.Debug(fmt.Sprintf("Swarm http proxy started with corsdomain: %v", self.corsString))
} }
return nil return nil