2018-06-20 15:06:27 +03:00
|
|
|
// Copyright 2018 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package stream
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
2018-09-24 18:40:22 +03:00
|
|
|
"errors"
|
2018-06-20 15:06:27 +03:00
|
|
|
"fmt"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/metrics"
|
2019-04-30 10:28:46 +03:00
|
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/chunk"
|
2018-06-20 15:06:27 +03:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/log"
|
2019-04-30 10:28:46 +03:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/network"
|
2018-06-20 15:06:27 +03:00
|
|
|
pq "github.com/ethereum/go-ethereum/swarm/network/priorityqueue"
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/network/stream/intervals"
|
2018-07-13 18:40:28 +03:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/spancontext"
|
2018-06-20 15:06:27 +03:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/state"
|
|
|
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
2019-02-20 16:50:37 +03:00
|
|
|
"github.com/ethereum/go-ethereum/swarm/tracing"
|
2018-07-13 18:40:28 +03:00
|
|
|
opentracing "github.com/opentracing/opentracing-go"
|
2018-06-20 15:06:27 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
type notFoundError struct {
|
|
|
|
t string
|
|
|
|
s Stream
|
|
|
|
}
|
|
|
|
|
|
|
|
func newNotFoundError(t string, s Stream) *notFoundError {
|
|
|
|
return ¬FoundError{t: t, s: s}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (e *notFoundError) Error() string {
|
|
|
|
return fmt.Sprintf("%s not found for stream %q", e.t, e.s)
|
|
|
|
}
|
|
|
|
|
2018-09-24 18:40:22 +03:00
|
|
|
// ErrMaxPeerServers will be returned if peer server limit is reached.
|
|
|
|
// It will be sent in the SubscribeErrorMsg.
|
|
|
|
var ErrMaxPeerServers = errors.New("max peer servers")
|
|
|
|
|
2018-06-20 15:06:27 +03:00
|
|
|
// Peer is the Peer extension for the streaming protocol
|
|
|
|
type Peer struct {
|
2019-04-30 10:28:46 +03:00
|
|
|
*network.BzzPeer
|
2018-06-20 15:06:27 +03:00
|
|
|
streamer *Registry
|
|
|
|
pq *pq.PriorityQueue
|
|
|
|
serverMu sync.RWMutex
|
|
|
|
clientMu sync.RWMutex // protects both clients and clientParams
|
|
|
|
servers map[Stream]*server
|
|
|
|
clients map[Stream]*client
|
|
|
|
// clientParams map keeps required client arguments
|
|
|
|
// that are set on Registry.Subscribe and used
|
|
|
|
// on creating a new client in offered hashes handler.
|
|
|
|
clientParams map[Stream]*clientParams
|
|
|
|
quit chan struct{}
|
|
|
|
}
|
|
|
|
|
2018-07-13 18:40:28 +03:00
|
|
|
type WrappedPriorityMsg struct {
|
|
|
|
Context context.Context
|
|
|
|
Msg interface{}
|
|
|
|
}
|
|
|
|
|
2018-06-20 15:06:27 +03:00
|
|
|
// NewPeer is the constructor for Peer
|
2019-04-30 10:28:46 +03:00
|
|
|
func NewPeer(peer *network.BzzPeer, streamer *Registry) *Peer {
|
2018-06-20 15:06:27 +03:00
|
|
|
p := &Peer{
|
2019-04-30 10:28:46 +03:00
|
|
|
BzzPeer: peer,
|
2018-06-20 15:06:27 +03:00
|
|
|
pq: pq.New(int(PriorityQueue), PriorityQueueCap),
|
|
|
|
streamer: streamer,
|
|
|
|
servers: make(map[Stream]*server),
|
|
|
|
clients: make(map[Stream]*client),
|
|
|
|
clientParams: make(map[Stream]*clientParams),
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
}
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
2018-07-13 18:40:28 +03:00
|
|
|
go p.pq.Run(ctx, func(i interface{}) {
|
|
|
|
wmsg := i.(WrappedPriorityMsg)
|
2018-09-13 12:42:19 +03:00
|
|
|
err := p.Send(wmsg.Context, wmsg.Msg)
|
|
|
|
if err != nil {
|
|
|
|
log.Error("Message send error, dropping peer", "peer", p.ID(), "err", err)
|
2019-04-11 11:26:52 +03:00
|
|
|
p.Drop()
|
2018-09-13 12:42:19 +03:00
|
|
|
}
|
2018-07-13 18:40:28 +03:00
|
|
|
})
|
2018-09-13 12:42:19 +03:00
|
|
|
|
|
|
|
// basic monitoring for pq contention
|
|
|
|
go func(pq *pq.PriorityQueue) {
|
|
|
|
ticker := time.NewTicker(5 * time.Second)
|
|
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ticker.C:
|
2019-02-24 14:39:23 +03:00
|
|
|
var lenMaxi int
|
|
|
|
var capMaxi int
|
2018-09-13 12:42:19 +03:00
|
|
|
for k := range pq.Queues {
|
2019-02-24 14:39:23 +03:00
|
|
|
if lenMaxi < len(pq.Queues[k]) {
|
|
|
|
lenMaxi = len(pq.Queues[k])
|
2018-09-13 12:42:19 +03:00
|
|
|
}
|
|
|
|
|
2019-02-24 14:39:23 +03:00
|
|
|
if capMaxi < cap(pq.Queues[k]) {
|
|
|
|
capMaxi = cap(pq.Queues[k])
|
2018-09-13 12:42:19 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-24 14:39:23 +03:00
|
|
|
metrics.GetOrRegisterGauge(fmt.Sprintf("pq_len_%s", p.ID().TerminalString()), nil).Update(int64(lenMaxi))
|
|
|
|
metrics.GetOrRegisterGauge(fmt.Sprintf("pq_cap_%s", p.ID().TerminalString()), nil).Update(int64(capMaxi))
|
2018-09-13 12:42:19 +03:00
|
|
|
case <-p.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}(p.pq)
|
|
|
|
|
2018-06-20 15:06:27 +03:00
|
|
|
go func() {
|
|
|
|
<-p.quit
|
2019-02-20 16:50:37 +03:00
|
|
|
|
2018-06-20 15:06:27 +03:00
|
|
|
cancel()
|
|
|
|
}()
|
|
|
|
return p
|
|
|
|
}
|
|
|
|
|
|
|
|
// Deliver sends a storeRequestMsg protocol message to the peer
|
2018-10-21 10:30:41 +03:00
|
|
|
// Depending on the `syncing` parameter we send different message types
|
|
|
|
func (p *Peer) Deliver(ctx context.Context, chunk storage.Chunk, priority uint8, syncing bool) error {
|
|
|
|
var msg interface{}
|
|
|
|
|
2019-04-26 17:41:12 +03:00
|
|
|
metrics.GetOrRegisterCounter("peer.deliver", nil).Inc(1)
|
2018-10-21 10:30:41 +03:00
|
|
|
|
|
|
|
//we send different types of messages if delivery is for syncing or retrievals,
|
|
|
|
//even if handling and content of the message are the same,
|
|
|
|
//because swap accounting decides which messages need accounting based on the message type
|
|
|
|
if syncing {
|
|
|
|
msg = &ChunkDeliveryMsgSyncing{
|
|
|
|
Addr: chunk.Address(),
|
|
|
|
SData: chunk.Data(),
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
msg = &ChunkDeliveryMsgRetrieval{
|
|
|
|
Addr: chunk.Address(),
|
|
|
|
SData: chunk.Data(),
|
|
|
|
}
|
|
|
|
}
|
2018-07-13 18:40:28 +03:00
|
|
|
|
2019-02-20 16:50:37 +03:00
|
|
|
return p.SendPriority(ctx, msg, priority)
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// SendPriority sends message to the peer using the outgoing priority queue
|
2019-02-20 16:50:37 +03:00
|
|
|
func (p *Peer) SendPriority(ctx context.Context, msg interface{}, priority uint8) error {
|
2018-06-20 15:06:27 +03:00
|
|
|
defer metrics.GetOrRegisterResettingTimer(fmt.Sprintf("peer.sendpriority_t.%d", priority), nil).UpdateSince(time.Now())
|
2019-03-08 10:52:25 +03:00
|
|
|
ctx = tracing.StartSaveSpan(ctx)
|
2018-06-20 15:06:27 +03:00
|
|
|
metrics.GetOrRegisterCounter(fmt.Sprintf("peer.sendpriority.%d", priority), nil).Inc(1)
|
2018-07-13 18:40:28 +03:00
|
|
|
wmsg := WrappedPriorityMsg{
|
|
|
|
Context: ctx,
|
|
|
|
Msg: msg,
|
|
|
|
}
|
2018-09-13 12:42:19 +03:00
|
|
|
err := p.pq.Push(wmsg, int(priority))
|
2019-03-11 13:45:34 +03:00
|
|
|
if err != nil {
|
|
|
|
log.Error("err on p.pq.Push", "err", err, "peer", p.ID())
|
2018-09-13 12:42:19 +03:00
|
|
|
}
|
|
|
|
return err
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// SendOfferedHashes sends OfferedHashesMsg protocol msg
|
|
|
|
func (p *Peer) SendOfferedHashes(s *server, f, t uint64) error {
|
2018-07-13 18:40:28 +03:00
|
|
|
var sp opentracing.Span
|
|
|
|
ctx, sp := spancontext.StartSpan(
|
|
|
|
context.TODO(),
|
2019-02-20 16:50:37 +03:00
|
|
|
"send.offered.hashes",
|
|
|
|
)
|
2018-07-13 18:40:28 +03:00
|
|
|
defer sp.Finish()
|
|
|
|
|
2019-03-11 13:45:34 +03:00
|
|
|
defer metrics.GetOrRegisterResettingTimer("send.offered.hashes", nil).UpdateSince(time.Now())
|
|
|
|
|
2018-10-12 17:26:16 +03:00
|
|
|
hashes, from, to, proof, err := s.setNextBatch(f, t)
|
2018-06-20 15:06:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-09-19 19:10:40 +03:00
|
|
|
// true only when quitting
|
2018-06-20 15:06:27 +03:00
|
|
|
if len(hashes) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if proof == nil {
|
|
|
|
proof = &HandoverProof{
|
|
|
|
Handover: &Handover{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s.currentBatch = hashes
|
|
|
|
msg := &OfferedHashesMsg{
|
|
|
|
HandoverProof: proof,
|
|
|
|
Hashes: hashes,
|
|
|
|
From: from,
|
|
|
|
To: to,
|
|
|
|
Stream: s.stream,
|
|
|
|
}
|
|
|
|
log.Trace("Swarm syncer offer batch", "peer", p.ID(), "stream", s.stream, "len", len(hashes), "from", from, "to", to)
|
2019-02-20 16:50:37 +03:00
|
|
|
ctx = context.WithValue(ctx, "stream_send_tag", "send.offered.hashes")
|
|
|
|
return p.SendPriority(ctx, msg, s.priority)
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) getServer(s Stream) (*server, error) {
|
|
|
|
p.serverMu.RLock()
|
|
|
|
defer p.serverMu.RUnlock()
|
|
|
|
|
|
|
|
server := p.servers[s]
|
|
|
|
if server == nil {
|
|
|
|
return nil, newNotFoundError("server", s)
|
|
|
|
}
|
|
|
|
return server, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) setServer(s Stream, o Server, priority uint8) (*server, error) {
|
|
|
|
p.serverMu.Lock()
|
|
|
|
defer p.serverMu.Unlock()
|
|
|
|
|
|
|
|
if p.servers[s] != nil {
|
|
|
|
return nil, fmt.Errorf("server %s already registered", s)
|
|
|
|
}
|
2018-09-24 18:40:22 +03:00
|
|
|
|
|
|
|
if p.streamer.maxPeerServers > 0 && len(p.servers) >= p.streamer.maxPeerServers {
|
|
|
|
return nil, ErrMaxPeerServers
|
|
|
|
}
|
|
|
|
|
2018-10-12 17:26:16 +03:00
|
|
|
sessionIndex, err := o.SessionIndex()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2018-06-20 15:06:27 +03:00
|
|
|
os := &server{
|
2018-10-12 17:26:16 +03:00
|
|
|
Server: o,
|
|
|
|
stream: s,
|
|
|
|
priority: priority,
|
|
|
|
sessionIndex: sessionIndex,
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
|
|
|
p.servers[s] = os
|
|
|
|
return os, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) removeServer(s Stream) error {
|
|
|
|
p.serverMu.Lock()
|
|
|
|
defer p.serverMu.Unlock()
|
|
|
|
|
|
|
|
server, ok := p.servers[s]
|
|
|
|
if !ok {
|
|
|
|
return newNotFoundError("server", s)
|
|
|
|
}
|
|
|
|
server.Close()
|
|
|
|
delete(p.servers, s)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) getClient(ctx context.Context, s Stream) (c *client, err error) {
|
|
|
|
var params *clientParams
|
|
|
|
func() {
|
|
|
|
p.clientMu.RLock()
|
|
|
|
defer p.clientMu.RUnlock()
|
|
|
|
|
|
|
|
c = p.clients[s]
|
|
|
|
if c != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
params = p.clientParams[s]
|
|
|
|
}()
|
|
|
|
if c != nil {
|
|
|
|
return c, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
if params != nil {
|
|
|
|
//debug.PrintStack()
|
|
|
|
if err := params.waitClient(ctx); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
p.clientMu.RLock()
|
|
|
|
defer p.clientMu.RUnlock()
|
|
|
|
|
|
|
|
c = p.clients[s]
|
|
|
|
if c != nil {
|
|
|
|
return c, nil
|
|
|
|
}
|
|
|
|
return nil, newNotFoundError("client", s)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) getOrSetClient(s Stream, from, to uint64) (c *client, created bool, err error) {
|
|
|
|
p.clientMu.Lock()
|
|
|
|
defer p.clientMu.Unlock()
|
|
|
|
|
|
|
|
c = p.clients[s]
|
|
|
|
if c != nil {
|
|
|
|
return c, false, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
f, err := p.streamer.GetClientFunc(s.Name)
|
|
|
|
if err != nil {
|
|
|
|
return nil, false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
is, err := f(p, s.Key, s.Live)
|
|
|
|
if err != nil {
|
|
|
|
return nil, false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
cp, err := p.getClientParams(s)
|
|
|
|
if err != nil {
|
|
|
|
return nil, false, err
|
|
|
|
}
|
|
|
|
defer func() {
|
|
|
|
if err == nil {
|
|
|
|
if err := p.removeClientParams(s); err != nil {
|
|
|
|
log.Error("stream set client: remove client params", "stream", s, "peer", p, "err", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
intervalsKey := peerStreamIntervalsKey(p, s)
|
|
|
|
if s.Live {
|
|
|
|
// try to find previous history and live intervals and merge live into history
|
|
|
|
historyKey := peerStreamIntervalsKey(p, NewStream(s.Name, s.Key, false))
|
|
|
|
historyIntervals := &intervals.Intervals{}
|
|
|
|
err := p.streamer.intervalsStore.Get(historyKey, historyIntervals)
|
|
|
|
switch err {
|
|
|
|
case nil:
|
|
|
|
liveIntervals := &intervals.Intervals{}
|
|
|
|
err := p.streamer.intervalsStore.Get(intervalsKey, liveIntervals)
|
|
|
|
switch err {
|
|
|
|
case nil:
|
|
|
|
historyIntervals.Merge(liveIntervals)
|
|
|
|
if err := p.streamer.intervalsStore.Put(historyKey, historyIntervals); err != nil {
|
|
|
|
log.Error("stream set client: put history intervals", "stream", s, "peer", p, "err", err)
|
|
|
|
}
|
|
|
|
case state.ErrNotFound:
|
|
|
|
default:
|
|
|
|
log.Error("stream set client: get live intervals", "stream", s, "peer", p, "err", err)
|
|
|
|
}
|
|
|
|
case state.ErrNotFound:
|
|
|
|
default:
|
|
|
|
log.Error("stream set client: get history intervals", "stream", s, "peer", p, "err", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := p.streamer.intervalsStore.Put(intervalsKey, intervals.NewIntervals(from)); err != nil {
|
|
|
|
return nil, false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
next := make(chan error, 1)
|
|
|
|
c = &client{
|
|
|
|
Client: is,
|
|
|
|
stream: s,
|
|
|
|
priority: cp.priority,
|
|
|
|
to: cp.to,
|
|
|
|
next: next,
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
intervalsStore: p.streamer.intervalsStore,
|
|
|
|
intervalsKey: intervalsKey,
|
|
|
|
}
|
|
|
|
p.clients[s] = c
|
|
|
|
cp.clientCreated() // unblock all possible getClient calls that are waiting
|
|
|
|
next <- nil // this is to allow wantedKeysMsg before first batch arrives
|
|
|
|
return c, true, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) removeClient(s Stream) error {
|
|
|
|
p.clientMu.Lock()
|
|
|
|
defer p.clientMu.Unlock()
|
|
|
|
|
|
|
|
client, ok := p.clients[s]
|
|
|
|
if !ok {
|
|
|
|
return newNotFoundError("client", s)
|
|
|
|
}
|
|
|
|
client.close()
|
2018-09-24 18:40:22 +03:00
|
|
|
delete(p.clients, s)
|
2018-06-20 15:06:27 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) setClientParams(s Stream, params *clientParams) error {
|
|
|
|
p.clientMu.Lock()
|
|
|
|
defer p.clientMu.Unlock()
|
|
|
|
|
|
|
|
if p.clients[s] != nil {
|
|
|
|
return fmt.Errorf("client %s already exists", s)
|
|
|
|
}
|
|
|
|
if p.clientParams[s] != nil {
|
|
|
|
return fmt.Errorf("client params %s already set", s)
|
|
|
|
}
|
|
|
|
p.clientParams[s] = params
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) getClientParams(s Stream) (*clientParams, error) {
|
|
|
|
params := p.clientParams[s]
|
|
|
|
if params == nil {
|
|
|
|
return nil, fmt.Errorf("client params '%v' not provided to peer %v", s, p.ID())
|
|
|
|
}
|
|
|
|
return params, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) removeClientParams(s Stream) error {
|
|
|
|
_, ok := p.clientParams[s]
|
|
|
|
if !ok {
|
|
|
|
return newNotFoundError("client params", s)
|
|
|
|
}
|
|
|
|
delete(p.clientParams, s)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (p *Peer) close() {
|
2019-05-06 11:58:42 +03:00
|
|
|
p.serverMu.Lock()
|
|
|
|
defer p.serverMu.Unlock()
|
|
|
|
|
2018-06-20 15:06:27 +03:00
|
|
|
for _, s := range p.servers {
|
|
|
|
s.Close()
|
|
|
|
}
|
2019-05-06 11:58:42 +03:00
|
|
|
|
|
|
|
p.servers = nil
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
2019-04-30 10:28:46 +03:00
|
|
|
|
|
|
|
// runUpdateSyncing is a long running function that creates the initial
|
|
|
|
// syncing subscriptions to the peer and waits for neighbourhood depth change
|
|
|
|
// to create new ones or quit existing ones based on the new neighbourhood depth
|
|
|
|
// and if peer enters or leaves nearest neighbourhood by using
|
|
|
|
// syncSubscriptionsDiff and updateSyncSubscriptions functions.
|
|
|
|
func (p *Peer) runUpdateSyncing() {
|
|
|
|
timer := time.NewTimer(p.streamer.syncUpdateDelay)
|
|
|
|
defer timer.Stop()
|
|
|
|
|
|
|
|
select {
|
|
|
|
case <-timer.C:
|
|
|
|
case <-p.streamer.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
kad := p.streamer.delivery.kad
|
|
|
|
po := chunk.Proximity(p.BzzAddr.Over(), kad.BaseAddr())
|
|
|
|
|
|
|
|
depth := kad.NeighbourhoodDepth()
|
|
|
|
|
|
|
|
log.Debug("update syncing subscriptions: initial", "peer", p.ID(), "po", po, "depth", depth)
|
|
|
|
|
|
|
|
// initial subscriptions
|
|
|
|
p.updateSyncSubscriptions(syncSubscriptionsDiff(po, -1, depth, kad.MaxProxDisplay))
|
|
|
|
|
|
|
|
depthChangeSignal, unsubscribeDepthChangeSignal := kad.SubscribeToNeighbourhoodDepthChange()
|
|
|
|
defer unsubscribeDepthChangeSignal()
|
|
|
|
|
|
|
|
prevDepth := depth
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case _, ok := <-depthChangeSignal:
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// update subscriptions for this peer when depth changes
|
|
|
|
depth := kad.NeighbourhoodDepth()
|
|
|
|
log.Debug("update syncing subscriptions", "peer", p.ID(), "po", po, "depth", depth)
|
|
|
|
p.updateSyncSubscriptions(syncSubscriptionsDiff(po, prevDepth, depth, kad.MaxProxDisplay))
|
|
|
|
prevDepth = depth
|
|
|
|
case <-p.streamer.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
log.Debug("update syncing subscriptions: exiting", "peer", p.ID())
|
|
|
|
}
|
|
|
|
|
|
|
|
// updateSyncSubscriptions accepts two slices of integers, the first one
|
|
|
|
// representing proximity order bins for required syncing subscriptions
|
|
|
|
// and the second one representing bins for syncing subscriptions that
|
|
|
|
// need to be removed. This function sends request for subscription
|
|
|
|
// messages and quit messages for provided bins.
|
|
|
|
func (p *Peer) updateSyncSubscriptions(subBins, quitBins []int) {
|
|
|
|
if p.streamer.getPeer(p.ID()) == nil {
|
|
|
|
log.Debug("update syncing subscriptions", "peer not found", p.ID())
|
|
|
|
return
|
|
|
|
}
|
|
|
|
log.Debug("update syncing subscriptions", "peer", p.ID(), "subscribe", subBins, "quit", quitBins)
|
|
|
|
for _, po := range subBins {
|
|
|
|
p.subscribeSync(po)
|
|
|
|
}
|
|
|
|
for _, po := range quitBins {
|
|
|
|
p.quitSync(po)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// subscribeSync send the request for syncing subscriptions to the peer
|
|
|
|
// using subscriptionFunc. This function is used to request syncing subscriptions
|
|
|
|
// when new peer is added to the registry and on neighbourhood depth change.
|
|
|
|
func (p *Peer) subscribeSync(po int) {
|
|
|
|
err := subscriptionFunc(p.streamer, p.ID(), uint8(po))
|
|
|
|
if err != nil {
|
|
|
|
log.Error("subscription", "err", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// quitSync sends the quit message for live and history syncing streams to the peer.
|
|
|
|
// This function is used in runUpdateSyncing indirectly over updateSyncSubscriptions
|
|
|
|
// to remove unneeded syncing subscriptions on neighbourhood depth change.
|
|
|
|
func (p *Peer) quitSync(po int) {
|
|
|
|
live := NewStream("SYNC", FormatSyncBinKey(uint8(po)), true)
|
|
|
|
history := getHistoryStream(live)
|
|
|
|
err := p.streamer.Quit(p.ID(), live)
|
|
|
|
if err != nil && err != p2p.ErrShuttingDown {
|
|
|
|
log.Error("quit", "err", err, "peer", p.ID(), "stream", live)
|
|
|
|
}
|
|
|
|
err = p.streamer.Quit(p.ID(), history)
|
|
|
|
if err != nil && err != p2p.ErrShuttingDown {
|
|
|
|
log.Error("quit", "err", err, "peer", p.ID(), "stream", history)
|
|
|
|
}
|
|
|
|
|
|
|
|
err = p.removeServer(live)
|
|
|
|
if err != nil {
|
|
|
|
log.Error("remove server", "err", err, "peer", p.ID(), "stream", live)
|
|
|
|
}
|
|
|
|
err = p.removeServer(history)
|
|
|
|
if err != nil {
|
|
|
|
log.Error("remove server", "err", err, "peer", p.ID(), "stream", live)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// syncSubscriptionsDiff calculates to which proximity order bins a peer
|
|
|
|
// (with po peerPO) needs to be subscribed after kademlia neighbourhood depth
|
|
|
|
// change from prevDepth to newDepth. Max argument limits the number of
|
|
|
|
// proximity order bins. Returned values are slices of integers which represent
|
|
|
|
// proximity order bins, the first one to which additional subscriptions need to
|
|
|
|
// be requested and the second one which subscriptions need to be quit. Argument
|
|
|
|
// prevDepth with value less then 0 represents no previous depth, used for
|
|
|
|
// initial syncing subscriptions.
|
|
|
|
func syncSubscriptionsDiff(peerPO, prevDepth, newDepth, max int) (subBins, quitBins []int) {
|
|
|
|
newStart, newEnd := syncBins(peerPO, newDepth, max)
|
|
|
|
if prevDepth < 0 {
|
|
|
|
// no previous depth, return the complete range
|
|
|
|
// for subscriptions requests and nothing for quitting
|
|
|
|
return intRange(newStart, newEnd), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
prevStart, prevEnd := syncBins(peerPO, prevDepth, max)
|
|
|
|
|
|
|
|
if newStart < prevStart {
|
|
|
|
subBins = append(subBins, intRange(newStart, prevStart)...)
|
|
|
|
}
|
|
|
|
|
|
|
|
if prevStart < newStart {
|
|
|
|
quitBins = append(quitBins, intRange(prevStart, newStart)...)
|
|
|
|
}
|
|
|
|
|
|
|
|
if newEnd < prevEnd {
|
|
|
|
quitBins = append(quitBins, intRange(newEnd, prevEnd)...)
|
|
|
|
}
|
|
|
|
|
|
|
|
if prevEnd < newEnd {
|
|
|
|
subBins = append(subBins, intRange(prevEnd, newEnd)...)
|
|
|
|
}
|
|
|
|
|
|
|
|
return subBins, quitBins
|
|
|
|
}
|
|
|
|
|
|
|
|
// syncBins returns the range to which proximity order bins syncing
|
|
|
|
// subscriptions need to be requested, based on peer proximity and
|
|
|
|
// kademlia neighbourhood depth. Returned range is [start,end), inclusive for
|
|
|
|
// start and exclusive for end.
|
|
|
|
func syncBins(peerPO, depth, max int) (start, end int) {
|
|
|
|
if peerPO < depth {
|
|
|
|
// subscribe only to peerPO bin if it is not
|
|
|
|
// in the nearest neighbourhood
|
|
|
|
return peerPO, peerPO + 1
|
|
|
|
}
|
|
|
|
// subscribe from depth to max bin if the peer
|
|
|
|
// is in the nearest neighbourhood
|
|
|
|
return depth, max + 1
|
|
|
|
}
|
|
|
|
|
|
|
|
// intRange returns the slice of integers [start,end). The start
|
|
|
|
// is inclusive and the end is not.
|
|
|
|
func intRange(start, end int) (r []int) {
|
|
|
|
for i := start; i < end; i++ {
|
|
|
|
r = append(r, i)
|
|
|
|
}
|
|
|
|
return r
|
|
|
|
}
|