mirror of
https://github.com/go-micro/go-micro.git
synced 2024-11-24 08:02:32 +02:00
7b385bf163
* logs alloc Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * fix allocs Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * fix allocs Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * tunnel allocs Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * try to fix tunnel Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * cache cipher for send Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org> * more logger Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
1884 lines
50 KiB
Go
1884 lines
50 KiB
Go
package network
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"hash/fnv"
|
|
"io"
|
|
"math"
|
|
"math/rand"
|
|
"sort"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
"github.com/micro/go-micro/v2/client"
|
|
cmucp "github.com/micro/go-micro/v2/client/mucp"
|
|
rtr "github.com/micro/go-micro/v2/client/selector/router"
|
|
"github.com/micro/go-micro/v2/logger"
|
|
"github.com/micro/go-micro/v2/network/resolver/dns"
|
|
pbNet "github.com/micro/go-micro/v2/network/service/proto"
|
|
"github.com/micro/go-micro/v2/proxy"
|
|
"github.com/micro/go-micro/v2/router"
|
|
pbRtr "github.com/micro/go-micro/v2/router/service/proto"
|
|
"github.com/micro/go-micro/v2/server"
|
|
smucp "github.com/micro/go-micro/v2/server/mucp"
|
|
"github.com/micro/go-micro/v2/transport"
|
|
"github.com/micro/go-micro/v2/tunnel"
|
|
bun "github.com/micro/go-micro/v2/tunnel/broker"
|
|
tun "github.com/micro/go-micro/v2/tunnel/transport"
|
|
"github.com/micro/go-micro/v2/util/backoff"
|
|
pbUtil "github.com/micro/go-micro/v2/util/proto"
|
|
)
|
|
|
|
var (
|
|
// NetworkChannel is the name of the tunnel channel for passing network messages
|
|
NetworkChannel = "network"
|
|
// ControlChannel is the name of the tunnel channel for passing control message
|
|
ControlChannel = "control"
|
|
// DefaultLink is default network link
|
|
DefaultLink = "network"
|
|
// MaxConnections is the max number of network client connections
|
|
MaxConnections = 3
|
|
// MaxPeerErrors is the max number of peer errors before we remove it from network graph
|
|
MaxPeerErrors = 3
|
|
)
|
|
|
|
var (
|
|
// ErrClientNotFound is returned when client for tunnel channel could not be found
|
|
ErrClientNotFound = errors.New("client not found")
|
|
// ErrPeerLinkNotFound is returned when peer link could not be found in tunnel Links
|
|
ErrPeerLinkNotFound = errors.New("peer link not found")
|
|
// ErrPeerMaxExceeded is returned when peer has reached its max error count limit
|
|
ErrPeerMaxExceeded = errors.New("peer max errors exceeded")
|
|
)
|
|
|
|
// network implements Network interface
|
|
type network struct {
|
|
// node is network node
|
|
*node
|
|
// options configure the network
|
|
options Options
|
|
// rtr is network router
|
|
router router.Router
|
|
// proxy is network proxy
|
|
proxy proxy.Proxy
|
|
// tunnel is network tunnel
|
|
tunnel tunnel.Tunnel
|
|
// server is network server
|
|
server server.Server
|
|
// client is network client
|
|
client client.Client
|
|
|
|
// tunClient is a map of tunnel channel clients
|
|
tunClient map[string]tunnel.Session
|
|
// peerLinks is a map of links for each peer
|
|
peerLinks map[string]tunnel.Link
|
|
|
|
sync.RWMutex
|
|
// connected marks the network as connected
|
|
connected bool
|
|
// closed closes the network
|
|
closed chan bool
|
|
// whether we've discovered by the network
|
|
discovered chan bool
|
|
}
|
|
|
|
// message is network message
|
|
type message struct {
|
|
// msg is transport message
|
|
msg *transport.Message
|
|
// session is tunnel session
|
|
session tunnel.Session
|
|
}
|
|
|
|
// newNetwork returns a new network node
|
|
func newNetwork(opts ...Option) Network {
|
|
// create default options
|
|
options := DefaultOptions()
|
|
// initialize network options
|
|
for _, o := range opts {
|
|
o(&options)
|
|
}
|
|
|
|
// set the address to a hashed address
|
|
hasher := fnv.New64()
|
|
hasher.Write([]byte(options.Address + options.Id))
|
|
address := fmt.Sprintf("%d", hasher.Sum64())
|
|
|
|
// set the address to advertise
|
|
var advertise string
|
|
var peerAddress string
|
|
|
|
if len(options.Advertise) > 0 {
|
|
advertise = options.Advertise
|
|
peerAddress = options.Advertise
|
|
} else {
|
|
advertise = options.Address
|
|
peerAddress = address
|
|
}
|
|
|
|
// init tunnel address to the network bind address
|
|
options.Tunnel.Init(
|
|
tunnel.Address(options.Address),
|
|
)
|
|
|
|
// init router Id to the network id
|
|
options.Router.Init(
|
|
router.Id(options.Id),
|
|
router.Address(peerAddress),
|
|
)
|
|
|
|
// create tunnel client with tunnel transport
|
|
tunTransport := tun.NewTransport(
|
|
tun.WithTunnel(options.Tunnel),
|
|
)
|
|
|
|
// create the tunnel broker
|
|
tunBroker := bun.NewBroker(
|
|
bun.WithTunnel(options.Tunnel),
|
|
)
|
|
|
|
// server is network server
|
|
server := smucp.NewServer(
|
|
server.Id(options.Id),
|
|
server.Address(peerAddress),
|
|
server.Advertise(advertise),
|
|
server.Name(options.Name),
|
|
server.Transport(tunTransport),
|
|
server.Broker(tunBroker),
|
|
)
|
|
|
|
// client is network client
|
|
client := cmucp.NewClient(
|
|
client.Broker(tunBroker),
|
|
client.Transport(tunTransport),
|
|
client.Selector(
|
|
rtr.NewSelector(
|
|
rtr.WithRouter(options.Router),
|
|
),
|
|
),
|
|
)
|
|
|
|
network := &network{
|
|
node: &node{
|
|
id: options.Id,
|
|
address: peerAddress,
|
|
peers: make(map[string]*node),
|
|
status: newStatus(),
|
|
},
|
|
options: options,
|
|
router: options.Router,
|
|
proxy: options.Proxy,
|
|
tunnel: options.Tunnel,
|
|
server: server,
|
|
client: client,
|
|
tunClient: make(map[string]tunnel.Session),
|
|
peerLinks: make(map[string]tunnel.Link),
|
|
discovered: make(chan bool, 1),
|
|
}
|
|
|
|
network.node.network = network
|
|
|
|
return network
|
|
}
|
|
|
|
func (n *network) Init(opts ...Option) error {
|
|
n.Lock()
|
|
defer n.Unlock()
|
|
|
|
// TODO: maybe only allow reinit of certain opts
|
|
for _, o := range opts {
|
|
o(&n.options)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Options returns network options
|
|
func (n *network) Options() Options {
|
|
n.RLock()
|
|
defer n.RUnlock()
|
|
|
|
options := n.options
|
|
|
|
return options
|
|
}
|
|
|
|
// Name returns network name
|
|
func (n *network) Name() string {
|
|
n.RLock()
|
|
defer n.RUnlock()
|
|
|
|
name := n.options.Name
|
|
|
|
return name
|
|
}
|
|
|
|
// acceptNetConn accepts connections from NetworkChannel
|
|
func (n *network) acceptNetConn(l tunnel.Listener, recv chan *message) {
|
|
var i int
|
|
for {
|
|
// accept a connection
|
|
conn, err := l.Accept()
|
|
if err != nil {
|
|
sleep := backoff.Do(i)
|
|
logger.Debugf("Network tunnel [%s] accept error: %v, backing off for %v", ControlChannel, err, sleep)
|
|
time.Sleep(sleep)
|
|
i++
|
|
continue
|
|
}
|
|
|
|
select {
|
|
case <-n.closed:
|
|
if err := conn.Close(); err != nil {
|
|
logger.Debugf("Network tunnel [%s] failed to close connection: %v", NetworkChannel, err)
|
|
}
|
|
return
|
|
default:
|
|
// go handle NetworkChannel connection
|
|
go n.handleNetConn(conn, recv)
|
|
}
|
|
}
|
|
}
|
|
|
|
// acceptCtrlConn accepts connections from ControlChannel
|
|
func (n *network) acceptCtrlConn(l tunnel.Listener, recv chan *message) {
|
|
var i int
|
|
for {
|
|
// accept a connection
|
|
conn, err := l.Accept()
|
|
if err != nil {
|
|
sleep := backoff.Do(i)
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network tunnel [%s] accept error: %v, backing off for %v", ControlChannel, err, sleep)
|
|
}
|
|
time.Sleep(sleep)
|
|
i++
|
|
continue
|
|
}
|
|
|
|
select {
|
|
case <-n.closed:
|
|
if err := conn.Close(); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network tunnel [%s] failed to close connection: %v", ControlChannel, err)
|
|
}
|
|
}
|
|
return
|
|
default:
|
|
// go handle ControlChannel connection
|
|
go n.handleCtrlConn(conn, recv)
|
|
}
|
|
}
|
|
}
|
|
|
|
// maskRoute will mask the route so that we apply the right values
|
|
func (n *network) maskRoute(r *pbRtr.Route) {
|
|
hasher := fnv.New64()
|
|
// the routes service address
|
|
address := r.Address
|
|
|
|
// only hash the address if we're advertising our own local routes
|
|
// avoid hashing * based routes
|
|
if r.Router == n.Id() && r.Address != "*" {
|
|
// hash the service before advertising it
|
|
hasher.Reset()
|
|
// routes for multiple instances of a service will be collapsed here.
|
|
// TODO: once we store labels in the table this may need to change
|
|
// to include the labels in case they differ but highly unlikely
|
|
hasher.Write([]byte(r.Service + n.Address()))
|
|
address = fmt.Sprintf("%d", hasher.Sum64())
|
|
}
|
|
|
|
// calculate route metric to advertise
|
|
metric := n.getRouteMetric(r.Router, r.Gateway, r.Link)
|
|
|
|
// NOTE: we override Gateway, Link and Address here
|
|
r.Address = address
|
|
r.Gateway = n.Address()
|
|
r.Link = DefaultLink
|
|
r.Metric = metric
|
|
}
|
|
|
|
// advertise advertises routes to the network
|
|
func (n *network) advertise(advertChan <-chan *router.Advert) {
|
|
rnd := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
for {
|
|
select {
|
|
// process local adverts and randomly fire them at other nodes
|
|
case advert := <-advertChan:
|
|
// create a proto advert
|
|
var events []*pbRtr.Event
|
|
|
|
for _, event := range advert.Events {
|
|
// make a copy of the route
|
|
route := &pbRtr.Route{
|
|
Service: event.Route.Service,
|
|
Address: event.Route.Address,
|
|
Gateway: event.Route.Gateway,
|
|
Network: event.Route.Network,
|
|
Router: event.Route.Router,
|
|
Link: event.Route.Link,
|
|
Metric: event.Route.Metric,
|
|
}
|
|
|
|
// override the various values
|
|
n.maskRoute(route)
|
|
|
|
e := &pbRtr.Event{
|
|
Type: pbRtr.EventType(event.Type),
|
|
Timestamp: event.Timestamp.UnixNano(),
|
|
Route: route,
|
|
}
|
|
|
|
events = append(events, e)
|
|
}
|
|
|
|
msg := &pbRtr.Advert{
|
|
Id: advert.Id,
|
|
Type: pbRtr.AdvertType(advert.Type),
|
|
Timestamp: advert.Timestamp.UnixNano(),
|
|
Events: events,
|
|
}
|
|
|
|
// get a list of node peers
|
|
peers := n.Peers()
|
|
|
|
// continue if there is no one to send to
|
|
if len(peers) == 0 {
|
|
continue
|
|
}
|
|
|
|
// advertise to max 3 peers
|
|
max := len(peers)
|
|
if max > 3 {
|
|
max = 3
|
|
}
|
|
|
|
for i := 0; i < max; i++ {
|
|
if peer := n.node.GetPeerNode(peers[rnd.Intn(len(peers))].Id()); peer != nil {
|
|
if err := n.sendTo("advert", ControlChannel, peer, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to advertise routes to %s: %v", peer.Id(), err)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
case <-n.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// initNodes initializes tunnel with a list of resolved nodes
|
|
func (n *network) initNodes(startup bool) {
|
|
nodes, err := n.resolveNodes()
|
|
// NOTE: this condition never fires
|
|
// as resolveNodes() never returns error
|
|
if err != nil && !startup {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to init nodes: %v", err)
|
|
}
|
|
return
|
|
}
|
|
|
|
// strip self
|
|
var init []string
|
|
|
|
// our current address
|
|
advertised := n.server.Options().Advertise
|
|
|
|
for _, node := range nodes {
|
|
// skip self
|
|
if node == advertised {
|
|
continue
|
|
}
|
|
// add the node
|
|
init = append(init, node)
|
|
}
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
// initialize the tunnel
|
|
logger.Tracef("Network initialising nodes %+v\n", init)
|
|
}
|
|
|
|
n.tunnel.Init(
|
|
tunnel.Nodes(nodes...),
|
|
)
|
|
}
|
|
|
|
// resolveNodes resolves network nodes to addresses
|
|
func (n *network) resolveNodes() ([]string, error) {
|
|
// resolve the network address to network nodes
|
|
records, err := n.options.Resolver.Resolve(n.options.Name)
|
|
if err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to resolve nodes: %v", err)
|
|
}
|
|
}
|
|
|
|
// sort by lowest priority
|
|
if err == nil {
|
|
sort.Slice(records, func(i, j int) bool { return records[i].Priority < records[j].Priority })
|
|
}
|
|
|
|
// keep processing
|
|
|
|
nodeMap := make(map[string]bool)
|
|
|
|
// collect network node addresses
|
|
//nolint:prealloc
|
|
var nodes []string
|
|
var i int
|
|
|
|
for _, record := range records {
|
|
if _, ok := nodeMap[record.Address]; ok {
|
|
continue
|
|
}
|
|
|
|
nodeMap[record.Address] = true
|
|
nodes = append(nodes, record.Address)
|
|
|
|
i++
|
|
|
|
// break once MaxConnection nodes has been reached
|
|
if i == MaxConnections {
|
|
break
|
|
}
|
|
}
|
|
|
|
// use the DNS resolver to expand peers
|
|
dns := &dns.Resolver{}
|
|
|
|
// append seed nodes if we have them
|
|
for _, node := range n.options.Nodes {
|
|
// resolve anything that looks like a host name
|
|
records, err := dns.Resolve(node)
|
|
if err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Failed to resolve %v %v", node, err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// add to the node map
|
|
for _, record := range records {
|
|
if _, ok := nodeMap[record.Address]; !ok {
|
|
nodes = append(nodes, record.Address)
|
|
}
|
|
}
|
|
}
|
|
|
|
return nodes, nil
|
|
}
|
|
|
|
// handleNetConn handles network announcement messages
|
|
func (n *network) handleNetConn(s tunnel.Session, msg chan *message) {
|
|
for {
|
|
m := new(transport.Message)
|
|
if err := s.Recv(m); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network tunnel [%s] receive error: %v", NetworkChannel, err)
|
|
}
|
|
switch err {
|
|
case io.EOF, tunnel.ErrReadTimeout:
|
|
s.Close()
|
|
return
|
|
}
|
|
continue
|
|
}
|
|
|
|
// check if peer is set
|
|
peer := m.Header["Micro-Peer"]
|
|
|
|
// check who the message is intended for
|
|
if len(peer) > 0 && peer != n.options.Id {
|
|
continue
|
|
}
|
|
|
|
select {
|
|
case msg <- &message{
|
|
msg: m,
|
|
session: s,
|
|
}:
|
|
case <-n.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// handleCtrlConn handles ControlChannel connections
|
|
func (n *network) handleCtrlConn(s tunnel.Session, msg chan *message) {
|
|
for {
|
|
m := new(transport.Message)
|
|
if err := s.Recv(m); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network tunnel [%s] receive error: %v", ControlChannel, err)
|
|
}
|
|
switch err {
|
|
case io.EOF, tunnel.ErrReadTimeout:
|
|
s.Close()
|
|
return
|
|
}
|
|
continue
|
|
}
|
|
|
|
// check if peer is set
|
|
peer := m.Header["Micro-Peer"]
|
|
|
|
// check who the message is intended for
|
|
if len(peer) > 0 && peer != n.options.Id {
|
|
continue
|
|
}
|
|
|
|
select {
|
|
case msg <- &message{
|
|
msg: m,
|
|
session: s,
|
|
}:
|
|
case <-n.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// getHopCount queries network graph and returns hop count for given router
|
|
// NOTE: this should be called getHopeMetric
|
|
// - Routes for local services have hop count 1
|
|
// - Routes with ID of adjacent nodes have hop count 10
|
|
// - Routes by peers of the advertiser have hop count 100
|
|
// - Routes beyond node neighbourhood have hop count 1000
|
|
func (n *network) getHopCount(rtr string) int {
|
|
// make sure node.peers are not modified
|
|
n.node.RLock()
|
|
defer n.node.RUnlock()
|
|
|
|
// we are the origin of the route
|
|
if rtr == n.options.Id {
|
|
return 1
|
|
}
|
|
|
|
// the route origin is our peer
|
|
if _, ok := n.node.peers[rtr]; ok {
|
|
return 10
|
|
}
|
|
|
|
// the route origin is the peer of our peer
|
|
for _, peer := range n.node.peers {
|
|
for id := range peer.peers {
|
|
if rtr == id {
|
|
return 100
|
|
}
|
|
}
|
|
}
|
|
// otherwise we are three hops away
|
|
return 1000
|
|
}
|
|
|
|
// getRouteMetric calculates router metric and returns it
|
|
// Route metric is calculated based on link status and route hopd count
|
|
func (n *network) getRouteMetric(router string, gateway string, link string) int64 {
|
|
// set the route metric
|
|
n.RLock()
|
|
defer n.RUnlock()
|
|
|
|
// local links are marked as 1
|
|
if link == "local" && gateway == "" {
|
|
return 1
|
|
}
|
|
|
|
// local links from other gateways as 2
|
|
if link == "local" && gateway != "" {
|
|
return 2
|
|
}
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network looking up %s link to gateway: %s", link, gateway)
|
|
}
|
|
// attempt to find link based on gateway address
|
|
lnk, ok := n.peerLinks[gateway]
|
|
if !ok {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to find a link to gateway: %s", gateway)
|
|
}
|
|
// no link found so infinite metric returned
|
|
return math.MaxInt64
|
|
}
|
|
|
|
// calculating metric
|
|
|
|
delay := lnk.Delay()
|
|
hops := n.getHopCount(router)
|
|
length := lnk.Length()
|
|
|
|
// make sure delay is non-zero
|
|
if delay == 0 {
|
|
delay = 1
|
|
}
|
|
|
|
// make sure length is non-zero
|
|
if length == 0 {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Link length is 0 %v %v", link, lnk.Length())
|
|
}
|
|
length = 10e9
|
|
}
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network calculated metric %v delay %v length %v distance %v", (delay*length*int64(hops))/10e6, delay, length, hops)
|
|
}
|
|
|
|
return (delay * length * int64(hops)) / 10e6
|
|
}
|
|
|
|
// processCtrlChan processes messages received on ControlChannel
|
|
func (n *network) processCtrlChan(listener tunnel.Listener) {
|
|
defer listener.Close()
|
|
|
|
// receive control message queue
|
|
recv := make(chan *message, 128)
|
|
|
|
// accept ControlChannel cconnections
|
|
go n.acceptCtrlConn(listener, recv)
|
|
|
|
for {
|
|
select {
|
|
case m := <-recv:
|
|
// switch on type of message and take action
|
|
switch m.msg.Header["Micro-Method"] {
|
|
case "advert":
|
|
pbRtrAdvert := &pbRtr.Advert{}
|
|
|
|
if err := proto.Unmarshal(m.msg.Body, pbRtrAdvert); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network fail to unmarshal advert message: %v", err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// don't process your own messages
|
|
if pbRtrAdvert.Id == n.options.Id {
|
|
continue
|
|
}
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network received advert message from: %s", pbRtrAdvert.Id)
|
|
}
|
|
|
|
// loookup advertising node in our peer topology
|
|
advertNode := n.node.GetPeerNode(pbRtrAdvert.Id)
|
|
if advertNode == nil {
|
|
// if we can't find the node in our topology (MaxDepth) we skipp prcessing adverts
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network skipping advert message from unknown peer: %s", pbRtrAdvert.Id)
|
|
}
|
|
continue
|
|
}
|
|
|
|
var events []*router.Event
|
|
|
|
for _, event := range pbRtrAdvert.Events {
|
|
// for backwards compatibility reasons
|
|
if event == nil || event.Route == nil {
|
|
continue
|
|
}
|
|
|
|
// we know the advertising node is not the origin of the route
|
|
if pbRtrAdvert.Id != event.Route.Router {
|
|
// if the origin router is not the advertising node peer
|
|
// we can't rule out potential routing loops so we bail here
|
|
if peer := advertNode.GetPeerNode(event.Route.Router); peer == nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network skipping advert message from peer: %s", pbRtrAdvert.Id)
|
|
}
|
|
continue
|
|
}
|
|
}
|
|
|
|
route := router.Route{
|
|
Service: event.Route.Service,
|
|
Address: event.Route.Address,
|
|
Gateway: event.Route.Gateway,
|
|
Network: event.Route.Network,
|
|
Router: event.Route.Router,
|
|
Link: event.Route.Link,
|
|
Metric: event.Route.Metric,
|
|
}
|
|
|
|
// calculate route metric and add to the advertised metric
|
|
// we need to make sure we do not overflow math.MaxInt64
|
|
metric := n.getRouteMetric(event.Route.Router, event.Route.Gateway, event.Route.Link)
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network metric for router %s and gateway %s: %v", event.Route.Router, event.Route.Gateway, metric)
|
|
}
|
|
|
|
// check we don't overflow max int 64
|
|
if d := route.Metric + metric; d <= 0 {
|
|
// set to max int64 if we overflow
|
|
route.Metric = math.MaxInt64
|
|
} else {
|
|
// set the combined value of metrics otherwise
|
|
route.Metric = d
|
|
}
|
|
|
|
// create router event
|
|
e := &router.Event{
|
|
Type: router.EventType(event.Type),
|
|
Timestamp: time.Unix(0, pbRtrAdvert.Timestamp),
|
|
Route: route,
|
|
}
|
|
events = append(events, e)
|
|
}
|
|
|
|
// if no events are eligible for processing continue
|
|
if len(events) == 0 {
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network no events to be processed by router: %s", n.options.Id)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// create an advert and process it
|
|
advert := &router.Advert{
|
|
Id: pbRtrAdvert.Id,
|
|
Type: router.AdvertType(pbRtrAdvert.Type),
|
|
Timestamp: time.Unix(0, pbRtrAdvert.Timestamp),
|
|
TTL: time.Duration(pbRtrAdvert.Ttl),
|
|
Events: events,
|
|
}
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network router %s processing advert: %s", n.Id(), advert.Id)
|
|
}
|
|
if err := n.router.Process(advert); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to process advert %s: %v", advert.Id, err)
|
|
}
|
|
}
|
|
}
|
|
case <-n.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// processNetChan processes messages received on NetworkChannel
|
|
func (n *network) processNetChan(listener tunnel.Listener) {
|
|
defer listener.Close()
|
|
|
|
// receive network message queue
|
|
recv := make(chan *message, 128)
|
|
|
|
// accept NetworkChannel connections
|
|
go n.acceptNetConn(listener, recv)
|
|
|
|
for {
|
|
select {
|
|
case m := <-recv:
|
|
// switch on type of message and take action
|
|
switch m.msg.Header["Micro-Method"] {
|
|
case "connect":
|
|
// mark the time the message has been received
|
|
now := time.Now()
|
|
|
|
pbNetConnect := &pbNet.Connect{}
|
|
if err := proto.Unmarshal(m.msg.Body, pbNetConnect); err != nil {
|
|
logger.Debugf("Network tunnel [%s] connect unmarshal error: %v", NetworkChannel, err)
|
|
continue
|
|
}
|
|
|
|
// don't process your own messages
|
|
if pbNetConnect.Node.Id == n.options.Id {
|
|
continue
|
|
}
|
|
|
|
logger.Debugf("Network received connect message from: %s", pbNetConnect.Node.Id)
|
|
|
|
peer := &node{
|
|
id: pbNetConnect.Node.Id,
|
|
address: pbNetConnect.Node.Address,
|
|
link: m.msg.Header["Micro-Link"],
|
|
peers: make(map[string]*node),
|
|
status: newStatus(),
|
|
lastSeen: now,
|
|
}
|
|
|
|
// update peer links
|
|
|
|
// TODO: should we do this only if we manage to add a peer
|
|
// What should we do if the peer links failed to be updated?
|
|
if err := n.updatePeerLinks(peer); err != nil {
|
|
logger.Debugf("Network failed updating peer links: %s", err)
|
|
}
|
|
|
|
// add peer to the list of node peers
|
|
if err := n.AddPeer(peer); err == ErrPeerExists {
|
|
logger.Tracef("Network peer exists, refreshing: %s", peer.id)
|
|
// update lastSeen time for the peer
|
|
if err := n.RefreshPeer(peer.id, peer.link, now); err != nil {
|
|
logger.Debugf("Network failed refreshing peer %s: %v", peer.id, err)
|
|
}
|
|
}
|
|
|
|
// we send the sync message because someone has sent connect
|
|
// and wants to either connect or reconnect to the network
|
|
// The faster it gets the network config (routes and peer graph)
|
|
// the faster the network converges to a stable state
|
|
|
|
go func() {
|
|
// get node peer graph to send back to the connecting node
|
|
node := PeersToProto(n.node, MaxDepth)
|
|
|
|
msg := &pbNet.Sync{
|
|
Peer: node,
|
|
}
|
|
|
|
// get a list of the best routes for each service in our routing table
|
|
routes, err := n.getProtoRoutes()
|
|
if err != nil {
|
|
logger.Debugf("Network node %s failed listing routes: %v", n.id, err)
|
|
}
|
|
// attached the routes to the message
|
|
msg.Routes = routes
|
|
|
|
// send sync message to the newly connected peer
|
|
if err := n.sendTo("sync", NetworkChannel, peer, msg); err != nil {
|
|
logger.Debugf("Network failed to send sync message: %v", err)
|
|
}
|
|
}()
|
|
case "peer":
|
|
// mark the time the message has been received
|
|
now := time.Now()
|
|
pbNetPeer := &pbNet.Peer{}
|
|
|
|
if err := proto.Unmarshal(m.msg.Body, pbNetPeer); err != nil {
|
|
logger.Debugf("Network tunnel [%s] peer unmarshal error: %v", NetworkChannel, err)
|
|
continue
|
|
}
|
|
|
|
// don't process your own messages
|
|
if pbNetPeer.Node.Id == n.options.Id {
|
|
continue
|
|
}
|
|
|
|
logger.Debugf("Network received peer message from: %s %s", pbNetPeer.Node.Id, pbNetPeer.Node.Address)
|
|
|
|
peer := &node{
|
|
id: pbNetPeer.Node.Id,
|
|
address: pbNetPeer.Node.Address,
|
|
link: m.msg.Header["Micro-Link"],
|
|
peers: make(map[string]*node),
|
|
status: newPeerStatus(pbNetPeer),
|
|
lastSeen: now,
|
|
}
|
|
|
|
// update peer links
|
|
|
|
// TODO: should we do this only if we manage to add a peer
|
|
// What should we do if the peer links failed to be updated?
|
|
if err := n.updatePeerLinks(peer); err != nil {
|
|
logger.Debugf("Network failed updating peer links: %s", err)
|
|
}
|
|
|
|
// if it's a new peer i.e. we do not have it in our graph, we request full sync
|
|
if err := n.node.AddPeer(peer); err == nil {
|
|
go func() {
|
|
// marshal node graph into protobuf
|
|
node := PeersToProto(n.node, MaxDepth)
|
|
|
|
msg := &pbNet.Sync{
|
|
Peer: node,
|
|
}
|
|
|
|
// get a list of the best routes for each service in our routing table
|
|
routes, err := n.getProtoRoutes()
|
|
if err != nil {
|
|
logger.Debugf("Network node %s failed listing routes: %v", n.id, err)
|
|
}
|
|
// attached the routes to the message
|
|
msg.Routes = routes
|
|
|
|
// send sync message to the newly connected peer
|
|
if err := n.sendTo("sync", NetworkChannel, peer, msg); err != nil {
|
|
logger.Debugf("Network failed to send sync message: %v", err)
|
|
}
|
|
}()
|
|
|
|
continue
|
|
// if we already have the peer in our graph, skip further steps
|
|
} else if err != ErrPeerExists {
|
|
logger.Debugf("Network got error adding peer %v", err)
|
|
continue
|
|
}
|
|
|
|
logger.Tracef("Network peer exists, refreshing: %s", pbNetPeer.Node.Id)
|
|
|
|
// update lastSeen time for the peer
|
|
if err := n.RefreshPeer(peer.id, peer.link, now); err != nil {
|
|
logger.Debugf("Network failed refreshing peer %s: %v", pbNetPeer.Node.Id, err)
|
|
}
|
|
|
|
// NOTE: we don't unpack MaxDepth toplogy
|
|
peer = UnpackPeerTopology(pbNetPeer, now, MaxDepth-1)
|
|
// update the link
|
|
peer.link = m.msg.Header["Micro-Link"]
|
|
|
|
logger.Tracef("Network updating topology of node: %s", n.node.id)
|
|
if err := n.node.UpdatePeer(peer); err != nil {
|
|
logger.Debugf("Network failed to update peers: %v", err)
|
|
}
|
|
|
|
// tell the connect loop that we've been discovered
|
|
// so it stops sending connect messages out
|
|
select {
|
|
case n.discovered <- true:
|
|
default:
|
|
// don't block here
|
|
}
|
|
case "sync":
|
|
// record the timestamp of the message receipt
|
|
now := time.Now()
|
|
|
|
pbNetSync := &pbNet.Sync{}
|
|
if err := proto.Unmarshal(m.msg.Body, pbNetSync); err != nil {
|
|
logger.Debugf("Network tunnel [%s] sync unmarshal error: %v", NetworkChannel, err)
|
|
continue
|
|
}
|
|
|
|
// don't process your own messages
|
|
if pbNetSync.Peer.Node.Id == n.options.Id {
|
|
continue
|
|
}
|
|
|
|
logger.Debugf("Network received sync message from: %s", pbNetSync.Peer.Node.Id)
|
|
|
|
peer := &node{
|
|
id: pbNetSync.Peer.Node.Id,
|
|
address: pbNetSync.Peer.Node.Address,
|
|
link: m.msg.Header["Micro-Link"],
|
|
peers: make(map[string]*node),
|
|
status: newPeerStatus(pbNetSync.Peer),
|
|
lastSeen: now,
|
|
}
|
|
|
|
// update peer links
|
|
|
|
// TODO: should we do this only if we manage to add a peer
|
|
// What should we do if the peer links failed to be updated?
|
|
if err := n.updatePeerLinks(peer); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed updating peer links: %s", err)
|
|
}
|
|
}
|
|
|
|
// add peer to the list of node peers
|
|
if err := n.node.AddPeer(peer); err == ErrPeerExists {
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network peer exists, refreshing: %s", peer.id)
|
|
}
|
|
// update lastSeen time for the existing node
|
|
if err := n.RefreshPeer(peer.id, peer.link, now); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed refreshing peer %s: %v", peer.id, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// when we receive a sync message we update our routing table
|
|
// and send a peer message back to the network to announce our presence
|
|
|
|
// add all the routes we have received in the sync message
|
|
for _, pbRoute := range pbNetSync.Routes {
|
|
// unmarshal the routes received from remote peer
|
|
route := pbUtil.ProtoToRoute(pbRoute)
|
|
// continue if we are the originator of the route
|
|
if route.Router == n.router.Options().Id {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s skipping route addition: route already present", n.id)
|
|
}
|
|
continue
|
|
}
|
|
|
|
metric := n.getRouteMetric(route.Router, route.Gateway, route.Link)
|
|
// check we don't overflow max int 64
|
|
if d := route.Metric + metric; d <= 0 {
|
|
// set to max int64 if we overflow
|
|
route.Metric = math.MaxInt64
|
|
} else {
|
|
// set the combined value of metrics otherwise
|
|
route.Metric = d
|
|
}
|
|
|
|
/////////////////////////////////////////////////////////////////////
|
|
// maybe we should not be this clever ¯\_(ツ)_/¯ //
|
|
/////////////////////////////////////////////////////////////////////
|
|
// lookup best routes for the services in the just received route
|
|
q := []router.QueryOption{
|
|
router.QueryService(route.Service),
|
|
router.QueryStrategy(n.router.Options().Advertise),
|
|
}
|
|
|
|
routes, err := n.router.Table().Query(q...)
|
|
if err != nil && err != router.ErrRouteNotFound {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s failed listing best routes for %s: %v", n.id, route.Service, err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// we found no routes for the given service
|
|
// create the new route we have just received
|
|
if len(routes) == 0 {
|
|
if err := n.router.Table().Create(route); err != nil && err != router.ErrDuplicateRoute {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s failed to add route: %v", n.id, err)
|
|
}
|
|
}
|
|
continue
|
|
}
|
|
|
|
// find the best route for the given service
|
|
// from the routes that we would advertise
|
|
bestRoute := routes[0]
|
|
for _, r := range routes[0:] {
|
|
if bestRoute.Metric > r.Metric {
|
|
bestRoute = r
|
|
}
|
|
}
|
|
|
|
// Take the best route to given service and:
|
|
// only add new routes if the metric is better
|
|
// than the metric of our best route
|
|
|
|
if bestRoute.Metric <= route.Metric {
|
|
continue
|
|
}
|
|
///////////////////////////////////////////////////////////////////////
|
|
///////////////////////////////////////////////////////////////////////
|
|
|
|
// add route to the routing table
|
|
if err := n.router.Table().Create(route); err != nil && err != router.ErrDuplicateRoute {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s failed to add route: %v", n.id, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// update your sync timestamp
|
|
// NOTE: this might go away as we will be doing full table advert to random peer
|
|
if err := n.RefreshSync(now); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed refreshing sync time: %v", err)
|
|
}
|
|
}
|
|
|
|
go func() {
|
|
// get node peer graph to send back to the syncing node
|
|
msg := PeersToProto(n.node, MaxDepth)
|
|
|
|
// advertise yourself to the new node
|
|
if err := n.sendTo("peer", NetworkChannel, peer, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to advertise peers: %v", err)
|
|
}
|
|
}
|
|
}()
|
|
case "close":
|
|
pbNetClose := &pbNet.Close{}
|
|
if err := proto.Unmarshal(m.msg.Body, pbNetClose); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network tunnel [%s] close unmarshal error: %v", NetworkChannel, err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// don't process your own messages
|
|
if pbNetClose.Node.Id == n.options.Id {
|
|
continue
|
|
}
|
|
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network received close message from: %s", pbNetClose.Node.Id)
|
|
}
|
|
|
|
peer := &node{
|
|
id: pbNetClose.Node.Id,
|
|
address: pbNetClose.Node.Address,
|
|
}
|
|
|
|
if err := n.DeletePeerNode(peer.id); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to delete node %s routes: %v", peer.id, err)
|
|
}
|
|
}
|
|
|
|
if err := n.prunePeerRoutes(peer); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed pruning peer %s routes: %v", peer.id, err)
|
|
}
|
|
}
|
|
|
|
// NOTE: we should maybe advertise this to the network so we converge faster on closed nodes
|
|
// as opposed to our waiting until the node eventually gets pruned; something to think about
|
|
|
|
// delete peer from the peerLinks
|
|
n.Lock()
|
|
delete(n.peerLinks, pbNetClose.Node.Address)
|
|
n.Unlock()
|
|
}
|
|
case <-n.closed:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// pruneRoutes prunes routes return by given query
|
|
func (n *network) pruneRoutes(q ...router.QueryOption) error {
|
|
routes, err := n.router.Table().Query(q...)
|
|
if err != nil && err != router.ErrRouteNotFound {
|
|
return err
|
|
}
|
|
|
|
for _, route := range routes {
|
|
if err := n.router.Table().Delete(route); err != nil && err != router.ErrRouteNotFound {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// pruneNodeRoutes prunes routes that were either originated by or routable via given node
|
|
func (n *network) prunePeerRoutes(peer *node) error {
|
|
// lookup all routes originated by router
|
|
q := []router.QueryOption{
|
|
router.QueryRouter(peer.id),
|
|
}
|
|
if err := n.pruneRoutes(q...); err != nil {
|
|
return err
|
|
}
|
|
|
|
// lookup all routes routable via gw
|
|
q = []router.QueryOption{
|
|
router.QueryGateway(peer.address),
|
|
}
|
|
if err := n.pruneRoutes(q...); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// manage the process of announcing to peers and prune any peer nodes that have not been
|
|
// seen for a period of time. Also removes all the routes either originated by or routable
|
|
// by the stale nodes. it also resolves nodes periodically and adds them to the tunnel
|
|
func (n *network) manage() {
|
|
rnd := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
announce := time.NewTicker(AnnounceTime)
|
|
defer announce.Stop()
|
|
prune := time.NewTicker(PruneTime)
|
|
defer prune.Stop()
|
|
resolve := time.NewTicker(ResolveTime)
|
|
defer resolve.Stop()
|
|
netsync := time.NewTicker(SyncTime)
|
|
defer netsync.Stop()
|
|
|
|
// list of links we've sent to
|
|
links := make(map[string]time.Time)
|
|
|
|
for {
|
|
select {
|
|
case <-n.closed:
|
|
return
|
|
case <-announce.C:
|
|
current := make(map[string]time.Time)
|
|
|
|
// build link map of current links
|
|
for _, link := range n.tunnel.Links() {
|
|
if n.isLoopback(link) {
|
|
continue
|
|
}
|
|
// get an existing timestamp if it exists
|
|
current[link.Id()] = links[link.Id()]
|
|
}
|
|
|
|
// replace link map
|
|
// we do this because a growing map is not
|
|
// garbage collected
|
|
links = current
|
|
|
|
n.RLock()
|
|
var i int
|
|
// create a list of peers to send to
|
|
var peers []*node
|
|
|
|
// check peers to see if they need to be sent to
|
|
for _, peer := range n.peers {
|
|
if i >= 3 {
|
|
break
|
|
}
|
|
|
|
// get last sent
|
|
lastSent := links[peer.link]
|
|
|
|
// check when we last sent to the peer
|
|
// and send a peer message if we havent
|
|
if lastSent.IsZero() || time.Since(lastSent) > KeepAliveTime {
|
|
link := peer.link
|
|
id := peer.id
|
|
|
|
// might not exist for some weird reason
|
|
if len(link) == 0 {
|
|
// set the link via peer links
|
|
l, ok := n.peerLinks[peer.address]
|
|
if ok {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network link not found for peer %s cannot announce", peer.id)
|
|
}
|
|
continue
|
|
}
|
|
link = l.Id()
|
|
}
|
|
|
|
// add to the list of peers we're going to send to
|
|
peers = append(peers, &node{
|
|
id: id,
|
|
link: link,
|
|
})
|
|
|
|
// increment our count
|
|
i++
|
|
}
|
|
}
|
|
|
|
n.RUnlock()
|
|
|
|
// peers to proto
|
|
msg := PeersToProto(n.node, MaxDepth)
|
|
|
|
// we're only going to send to max 3 peers at any given tick
|
|
for _, peer := range peers {
|
|
// advertise yourself to the network
|
|
if err := n.sendTo("peer", NetworkChannel, peer, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to advertise peer %s: %v", peer.id, err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// update last sent time
|
|
links[peer.link] = time.Now()
|
|
}
|
|
|
|
// now look at links we may not have sent to. this may occur
|
|
// where a connect message was lost
|
|
for link, lastSent := range links {
|
|
if !lastSent.IsZero() || time.Since(lastSent) < KeepAliveTime {
|
|
continue
|
|
}
|
|
|
|
peer := &node{
|
|
// unknown id of the peer
|
|
link: link,
|
|
}
|
|
|
|
// unknown link and peer so lets do the connect flow
|
|
if err := n.sendTo("connect", NetworkChannel, peer, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to connect %s: %v", peer.id, err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
links[peer.link] = time.Now()
|
|
}
|
|
case <-prune.C:
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s pruning stale peers", n.id)
|
|
}
|
|
pruned := n.PruneStalePeers(PruneTime)
|
|
|
|
for id, peer := range pruned {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network peer exceeded prune time: %s", id)
|
|
}
|
|
n.Lock()
|
|
delete(n.peerLinks, peer.address)
|
|
n.Unlock()
|
|
|
|
if err := n.prunePeerRoutes(peer); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed pruning peer %s routes: %v", id, err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// get a list of all routes
|
|
routes, err := n.options.Router.Table().List()
|
|
if err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed listing routes when pruning peers: %v", err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// collect all the router IDs in the routing table
|
|
routers := make(map[string]bool)
|
|
|
|
for _, route := range routes {
|
|
// check if its been processed
|
|
if _, ok := routers[route.Router]; ok {
|
|
continue
|
|
}
|
|
|
|
// mark as processed
|
|
routers[route.Router] = true
|
|
|
|
// if the router is in our peer graph do NOT delete routes originated by it
|
|
if peer := n.node.GetPeerNode(route.Router); peer != nil {
|
|
continue
|
|
}
|
|
// otherwise delete all the routes originated by it
|
|
if err := n.pruneRoutes(router.QueryRouter(route.Router)); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed deleting routes by %s: %v", route.Router, err)
|
|
}
|
|
}
|
|
}
|
|
case <-netsync.C:
|
|
// get a list of node peers
|
|
peers := n.Peers()
|
|
|
|
// skip when there are no peers
|
|
if len(peers) == 0 {
|
|
continue
|
|
}
|
|
|
|
// pick a random peer from the list of peers and request full sync
|
|
peer := n.node.GetPeerNode(peers[rnd.Intn(len(peers))].Id())
|
|
// skip if we can't find randmly selected peer
|
|
if peer == nil {
|
|
continue
|
|
}
|
|
|
|
go func() {
|
|
// get node peer graph to send back to the connecting node
|
|
node := PeersToProto(n.node, MaxDepth)
|
|
|
|
msg := &pbNet.Sync{
|
|
Peer: node,
|
|
}
|
|
|
|
// get a list of the best routes for each service in our routing table
|
|
routes, err := n.getProtoRoutes()
|
|
if err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node %s failed listing routes: %v", n.id, err)
|
|
}
|
|
}
|
|
// attached the routes to the message
|
|
msg.Routes = routes
|
|
|
|
// send sync message to the newly connected peer
|
|
if err := n.sendTo("sync", NetworkChannel, peer, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to send sync message: %v", err)
|
|
}
|
|
}
|
|
}()
|
|
case <-resolve.C:
|
|
n.initNodes(false)
|
|
}
|
|
}
|
|
}
|
|
|
|
// getAdvertProtoRoutes returns a list of routes to advertise to remote peer
|
|
// based on the advertisement strategy encoded in protobuf
|
|
// It returns error if the routes failed to be retrieved from the routing table
|
|
func (n *network) getProtoRoutes() ([]*pbRtr.Route, error) {
|
|
// get a list of the best routes for each service in our routing table
|
|
q := []router.QueryOption{
|
|
router.QueryStrategy(n.router.Options().Advertise),
|
|
}
|
|
|
|
routes, err := n.router.Table().Query(q...)
|
|
if err != nil && err != router.ErrRouteNotFound {
|
|
return nil, err
|
|
}
|
|
|
|
// encode the routes to protobuf
|
|
pbRoutes := make([]*pbRtr.Route, 0, len(routes))
|
|
for _, route := range routes {
|
|
// generate new route proto
|
|
pbRoute := pbUtil.RouteToProto(route)
|
|
// mask the route before outbounding
|
|
n.maskRoute(pbRoute)
|
|
// add to list of routes
|
|
pbRoutes = append(pbRoutes, pbRoute)
|
|
}
|
|
|
|
return pbRoutes, nil
|
|
}
|
|
|
|
func (n *network) sendConnect() {
|
|
// send connect message to NetworkChannel
|
|
// NOTE: in theory we could do this as soon as
|
|
// Dial to NetworkChannel succeeds, but instead
|
|
// we initialize all other node resources first
|
|
msg := &pbNet.Connect{
|
|
Node: &pbNet.Node{
|
|
Id: n.node.id,
|
|
Address: n.node.address,
|
|
},
|
|
}
|
|
|
|
if err := n.sendMsg("connect", NetworkChannel, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to send connect message: %s", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// sendTo sends a message to a specific node as a one off.
|
|
// we need this because when links die, we have no discovery info,
|
|
// and sending to an existing multicast link doesn't immediately work
|
|
func (n *network) sendTo(method, channel string, peer *node, msg proto.Message) error {
|
|
body, err := proto.Marshal(msg)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Create a unicast connection to the peer but don't do the open/accept flow
|
|
c, err := n.tunnel.Dial(channel, tunnel.DialWait(false), tunnel.DialLink(peer.link))
|
|
if err != nil {
|
|
if peerNode := n.GetPeerNode(peer.id); peerNode != nil {
|
|
// update node status when error happens
|
|
peerNode.status.err.Update(err)
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network increment peer %v error count to: %d", peerNode, peerNode, peerNode.status.Error().Count())
|
|
}
|
|
if count := peerNode.status.Error().Count(); count == MaxPeerErrors {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network peer %v error count exceeded %d. Prunning.", peerNode, MaxPeerErrors)
|
|
}
|
|
n.PrunePeer(peerNode.id)
|
|
}
|
|
}
|
|
return err
|
|
}
|
|
defer c.Close()
|
|
|
|
id := peer.id
|
|
|
|
if len(id) == 0 {
|
|
id = peer.link
|
|
}
|
|
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network sending %s message from: %s to %s", method, n.options.Id, id)
|
|
}
|
|
tmsg := &transport.Message{
|
|
Header: map[string]string{
|
|
"Micro-Method": method,
|
|
},
|
|
Body: body,
|
|
}
|
|
|
|
// setting the peer header
|
|
if len(peer.id) > 0 {
|
|
tmsg.Header["Micro-Peer"] = peer.id
|
|
}
|
|
|
|
if err := c.Send(tmsg); err != nil {
|
|
// TODO: Lookup peer in our graph
|
|
if peerNode := n.GetPeerNode(peer.id); peerNode != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network found peer %s: %v", peer.id, peerNode)
|
|
}
|
|
// update node status when error happens
|
|
peerNode.status.err.Update(err)
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network increment node peer %p %v count to: %d", peerNode, peerNode, peerNode.status.Error().Count())
|
|
}
|
|
if count := peerNode.status.Error().Count(); count == MaxPeerErrors {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network node peer %v count exceeded %d: %d", peerNode, MaxPeerErrors, peerNode.status.Error().Count())
|
|
}
|
|
n.PrunePeer(peerNode.id)
|
|
}
|
|
}
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// sendMsg sends a message to the tunnel channel
|
|
func (n *network) sendMsg(method, channel string, msg proto.Message) error {
|
|
body, err := proto.Marshal(msg)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// check if the channel client is initialized
|
|
n.RLock()
|
|
client, ok := n.tunClient[channel]
|
|
if !ok || client == nil {
|
|
n.RUnlock()
|
|
return ErrClientNotFound
|
|
}
|
|
n.RUnlock()
|
|
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network sending %s message from: %s", method, n.options.Id)
|
|
}
|
|
|
|
return client.Send(&transport.Message{
|
|
Header: map[string]string{
|
|
"Micro-Method": method,
|
|
},
|
|
Body: body,
|
|
})
|
|
}
|
|
|
|
// updatePeerLinks updates link for a given peer
|
|
func (n *network) updatePeerLinks(peer *node) error {
|
|
n.Lock()
|
|
defer n.Unlock()
|
|
|
|
linkId := peer.link
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
logger.Tracef("Network looking up link %s in the peer links", linkId)
|
|
}
|
|
|
|
// lookup the peer link
|
|
var peerLink tunnel.Link
|
|
|
|
for _, link := range n.tunnel.Links() {
|
|
if link.Id() == linkId {
|
|
peerLink = link
|
|
break
|
|
}
|
|
}
|
|
|
|
if peerLink == nil {
|
|
return ErrPeerLinkNotFound
|
|
}
|
|
|
|
if logger.V(logger.TraceLevel, logger.DefaultLogger) {
|
|
// if the peerLink is found in the returned links update peerLinks
|
|
logger.Tracef("Network updating peer links for peer %s", peer.address)
|
|
}
|
|
|
|
// lookup a link and update it if better link is available
|
|
if link, ok := n.peerLinks[peer.address]; ok {
|
|
// if the existing has better Length then the new, replace it
|
|
if link.Length() < peerLink.Length() {
|
|
n.peerLinks[peer.address] = peerLink
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// add peerLink to the peerLinks map
|
|
n.peerLinks[peer.address] = peerLink
|
|
|
|
return nil
|
|
}
|
|
|
|
// isLoopback checks if a link is a loopback to ourselves
|
|
func (n *network) isLoopback(link tunnel.Link) bool {
|
|
// skip loopback
|
|
if link.Loopback() {
|
|
return true
|
|
}
|
|
|
|
// our advertise address
|
|
loopback := n.server.Options().Advertise
|
|
// actual address
|
|
address := n.tunnel.Address()
|
|
|
|
// if remote is ourselves
|
|
switch link.Remote() {
|
|
case loopback, address:
|
|
return true
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
// connect will wait for a link to be established and send the connect
|
|
// message. We're trying to ensure convergence pretty quickly. So we want
|
|
// to hear back. In the case we become completely disconnected we'll
|
|
// connect again once a new link is established
|
|
func (n *network) connect() {
|
|
// discovered lets us know what we received a peer message back
|
|
var discovered bool
|
|
var attempts int
|
|
|
|
for {
|
|
// connected is used to define if the link is connected
|
|
var connected bool
|
|
|
|
// check the links state
|
|
for _, link := range n.tunnel.Links() {
|
|
// skip loopback
|
|
if n.isLoopback(link) {
|
|
continue
|
|
}
|
|
|
|
if link.State() == "connected" {
|
|
connected = true
|
|
break
|
|
}
|
|
}
|
|
|
|
// if we're not connected wait
|
|
if !connected {
|
|
// reset discovered
|
|
discovered = false
|
|
// sleep for a second
|
|
time.Sleep(time.Second)
|
|
// now try again
|
|
continue
|
|
}
|
|
|
|
// we're connected but are we discovered?
|
|
if !discovered {
|
|
// recreate the clients because all the tunnel links are gone
|
|
// so we haven't send discovery beneath
|
|
// NOTE: when starting the tunnel for the first time we might be recreating potentially
|
|
// well functioning tunnel clients as "discovered" will be false until the
|
|
// n.discovered channel is read at some point later on.
|
|
if err := n.createClients(); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Failed to recreate network/control clients: %v", err)
|
|
}
|
|
continue
|
|
}
|
|
|
|
// send the connect message
|
|
n.sendConnect()
|
|
}
|
|
|
|
// check if we've been discovered
|
|
select {
|
|
case <-n.discovered:
|
|
discovered = true
|
|
attempts = 0
|
|
case <-n.closed:
|
|
return
|
|
case <-time.After(time.Second + backoff.Do(attempts)):
|
|
// we have to try again
|
|
attempts++
|
|
}
|
|
}
|
|
}
|
|
|
|
// Connect connects the network
|
|
func (n *network) Connect() error {
|
|
n.Lock()
|
|
defer n.Unlock()
|
|
|
|
// connect network tunnel
|
|
if err := n.tunnel.Connect(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// return if already connected
|
|
if n.connected {
|
|
// initialise the nodes
|
|
n.initNodes(false)
|
|
// send the connect message
|
|
go n.sendConnect()
|
|
return nil
|
|
}
|
|
|
|
// initialise the nodes
|
|
n.initNodes(true)
|
|
|
|
// set our internal node address
|
|
// if advertise address is not set
|
|
if len(n.options.Advertise) == 0 {
|
|
n.server.Init(server.Advertise(n.tunnel.Address()))
|
|
}
|
|
|
|
// listen on NetworkChannel
|
|
netListener, err := n.tunnel.Listen(
|
|
NetworkChannel,
|
|
tunnel.ListenMode(tunnel.Multicast),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// listen on ControlChannel
|
|
ctrlListener, err := n.tunnel.Listen(
|
|
ControlChannel,
|
|
tunnel.ListenMode(tunnel.Multicast),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// dial into ControlChannel to send route adverts
|
|
ctrlClient, err := n.tunnel.Dial(
|
|
ControlChannel,
|
|
tunnel.DialMode(tunnel.Multicast),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
n.tunClient[ControlChannel] = ctrlClient
|
|
|
|
// dial into NetworkChannel to send network messages
|
|
netClient, err := n.tunnel.Dial(
|
|
NetworkChannel,
|
|
tunnel.DialMode(tunnel.Multicast),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
n.tunClient[NetworkChannel] = netClient
|
|
|
|
// create closed channel
|
|
n.closed = make(chan bool)
|
|
|
|
// start the router
|
|
if err := n.options.Router.Start(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// start advertising routes
|
|
advertChan, err := n.options.Router.Advertise()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// start the server
|
|
if err := n.server.Start(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// advertise service routes
|
|
go n.advertise(advertChan)
|
|
// listen to network messages
|
|
go n.processNetChan(netListener)
|
|
// accept and process routes
|
|
go n.processCtrlChan(ctrlListener)
|
|
// manage connection once links are established
|
|
go n.connect()
|
|
// resolve nodes, broadcast announcements and prune stale nodes
|
|
go n.manage()
|
|
|
|
// we're now connected
|
|
n.connected = true
|
|
|
|
return nil
|
|
}
|
|
|
|
func (n *network) close() error {
|
|
// stop the server
|
|
if err := n.server.Stop(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// stop the router
|
|
if err := n.router.Stop(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// close the tunnel
|
|
if err := n.tunnel.Close(); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// createClients is used to create new clients in the event we lose all the tunnels
|
|
func (n *network) createClients() error {
|
|
// dial into ControlChannel to send route adverts
|
|
ctrlClient, err := n.tunnel.Dial(ControlChannel, tunnel.DialMode(tunnel.Multicast))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// dial into NetworkChannel to send network messages
|
|
netClient, err := n.tunnel.Dial(NetworkChannel, tunnel.DialMode(tunnel.Multicast))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
n.Lock()
|
|
defer n.Unlock()
|
|
|
|
// set the control client
|
|
c, ok := n.tunClient[ControlChannel]
|
|
if ok {
|
|
c.Close()
|
|
}
|
|
n.tunClient[ControlChannel] = ctrlClient
|
|
|
|
// set the network client
|
|
c, ok = n.tunClient[NetworkChannel]
|
|
if ok {
|
|
c.Close()
|
|
}
|
|
n.tunClient[NetworkChannel] = netClient
|
|
|
|
return nil
|
|
}
|
|
|
|
// Close closes network connection
|
|
func (n *network) Close() error {
|
|
n.Lock()
|
|
|
|
if !n.connected {
|
|
n.Unlock()
|
|
return nil
|
|
}
|
|
|
|
select {
|
|
case <-n.closed:
|
|
n.Unlock()
|
|
return nil
|
|
default:
|
|
close(n.closed)
|
|
|
|
// set connected to false
|
|
n.connected = false
|
|
|
|
// unlock the lock otherwise we'll deadlock sending the close
|
|
n.Unlock()
|
|
|
|
msg := &pbNet.Close{
|
|
Node: &pbNet.Node{
|
|
Id: n.node.id,
|
|
Address: n.node.address,
|
|
},
|
|
}
|
|
|
|
if err := n.sendMsg("close", NetworkChannel, msg); err != nil {
|
|
if logger.V(logger.DebugLevel, logger.DefaultLogger) {
|
|
logger.Debugf("Network failed to send close message: %s", err)
|
|
}
|
|
}
|
|
<-time.After(time.Millisecond * 100)
|
|
}
|
|
|
|
return n.close()
|
|
}
|
|
|
|
// Client returns network client
|
|
func (n *network) Client() client.Client {
|
|
return n.client
|
|
}
|
|
|
|
// Server returns network server
|
|
func (n *network) Server() server.Server {
|
|
return n.server
|
|
}
|