mirror of
https://github.com/zerotier/ZeroTierOne.git
synced 2024-12-23 14:52:24 +00:00
689 lines
18 KiB
C++
689 lines
18 KiB
C++
/*
|
|
* Copyright (c)2019 ZeroTier, Inc.
|
|
*
|
|
* Use of this software is governed by the Business Source License included
|
|
* in the LICENSE.TXT file in the project's root directory.
|
|
*
|
|
* Change Date: 2023-01-01
|
|
*
|
|
* On the date above, in accordance with the Business Source License, use
|
|
* of this software will be governed by version 2.0 of the Apache License.
|
|
*/
|
|
/****/
|
|
|
|
#ifndef ZT_TOPOLOGY_HPP
|
|
#define ZT_TOPOLOGY_HPP
|
|
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
|
|
#include <vector>
|
|
#include <stdexcept>
|
|
#include <algorithm>
|
|
#include <utility>
|
|
#include <set>
|
|
|
|
#include "Constants.hpp"
|
|
#include "../include/ZeroTierOne.h"
|
|
|
|
#include "Address.hpp"
|
|
#include "Identity.hpp"
|
|
#include "Peer.hpp"
|
|
#include "Path.hpp"
|
|
#include "Mutex.hpp"
|
|
#include "InetAddress.hpp"
|
|
#include "Hashtable.hpp"
|
|
#include "Locator.hpp"
|
|
#include "SharedPtr.hpp"
|
|
#include "ScopedPtr.hpp"
|
|
|
|
namespace ZeroTier {
|
|
|
|
class RuntimeEnvironment;
|
|
|
|
/**
|
|
* Database of network topology
|
|
*/
|
|
class Topology
|
|
{
|
|
private:
|
|
struct _RootRankingFunction
|
|
{
|
|
ZT_ALWAYS_INLINE _RootRankingFunction() : bestRoot(),bestRootLatency(0xffff) {}
|
|
ZT_ALWAYS_INLINE bool operator()(const SharedPtr<Peer> &peer,const std::vector<InetAddress> &phy)
|
|
{
|
|
const unsigned int lat = peer->latency(now);
|
|
if ((!bestRoot)||((lat <= bestRootLatency)&&(peer->getAppropriatePath(now,false)))) {
|
|
bestRoot = peer;
|
|
bestRootLatency = lat;
|
|
}
|
|
return true;
|
|
}
|
|
int64_t now;
|
|
SharedPtr<Peer> bestRoot;
|
|
unsigned int bestRootLatency;
|
|
};
|
|
|
|
ZT_ALWAYS_INLINE void _updateDynamicRootIdentities()
|
|
{
|
|
// assumes _dynamicRoots_l is locked
|
|
_dynamicRootIdentities.clear();
|
|
Hashtable< Str,Locator >::Iterator i(_dynamicRoots);
|
|
Str *k = (Str *)0;
|
|
Locator *v = (Locator *)0;
|
|
while (i.next(k,v)) {
|
|
if (*v)
|
|
_dynamicRootIdentities.set(v->id(),true);
|
|
}
|
|
}
|
|
|
|
public:
|
|
ZT_ALWAYS_INLINE Topology(const RuntimeEnvironment *renv,const Identity &myId) :
|
|
RR(renv),
|
|
_myIdentity(myId),
|
|
_numConfiguredPhysicalPaths(0),
|
|
_lastUpdatedBestRoot(0) {}
|
|
ZT_ALWAYS_INLINE ~Topology() {}
|
|
|
|
/**
|
|
* Add a peer to database
|
|
*
|
|
* This will not replace existing peers. In that case the existing peer
|
|
* record is returned.
|
|
*
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
|
* @param peer Peer to add
|
|
* @return New or existing peer (should replace 'peer')
|
|
*/
|
|
ZT_ALWAYS_INLINE SharedPtr<Peer> add(const SharedPtr<Peer> &peer)
|
|
{
|
|
SharedPtr<Peer> np;
|
|
{
|
|
Mutex::Lock _l(_peers_l);
|
|
SharedPtr<Peer> &hp = _peers[peer->address()];
|
|
if (!hp)
|
|
hp = peer;
|
|
np = hp;
|
|
}
|
|
return np;
|
|
}
|
|
|
|
/**
|
|
* Get a peer from its address
|
|
*
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
|
* @param zta ZeroTier address of peer
|
|
* @return Peer or NULL if not found
|
|
*/
|
|
ZT_ALWAYS_INLINE SharedPtr<Peer> get(const Address &zta)
|
|
{
|
|
if (zta == _myIdentity.address())
|
|
return SharedPtr<Peer>();
|
|
Mutex::Lock l1(_peers_l);
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
|
if (ap)
|
|
return *ap;
|
|
return SharedPtr<Peer>();
|
|
}
|
|
|
|
/**
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
|
* @param zta ZeroTier address of peer
|
|
* @return Identity or NULL identity if not found
|
|
*/
|
|
ZT_ALWAYS_INLINE Identity getIdentity(void *tPtr,const Address &zta)
|
|
{
|
|
if (zta == _myIdentity.address()) {
|
|
return _myIdentity;
|
|
} else {
|
|
Mutex::Lock _l(_peers_l);
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
|
if (ap)
|
|
return (*ap)->identity();
|
|
}
|
|
return Identity();
|
|
}
|
|
|
|
/**
|
|
* Get a Path object for a given local and remote physical address, creating if needed
|
|
*
|
|
* @param l Local socket
|
|
* @param r Remote address
|
|
* @return Pointer to canonicalized Path object
|
|
*/
|
|
ZT_ALWAYS_INLINE SharedPtr<Path> getPath(const int64_t l,const InetAddress &r)
|
|
{
|
|
Mutex::Lock _l(_paths_l);
|
|
SharedPtr<Path> &p = _paths[Path::HashKey(l,r)];
|
|
if (!p)
|
|
p.set(new Path(l,r));
|
|
return p;
|
|
}
|
|
|
|
/**
|
|
* @param id Identity to check
|
|
* @return True if this identity corresponds to a root
|
|
*/
|
|
ZT_ALWAYS_INLINE bool isRoot(const Identity &id) const
|
|
{
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
if (_dynamicRootIdentities.contains(id))
|
|
return true;
|
|
}
|
|
{
|
|
Mutex::Lock l(_staticRoots_l);
|
|
if (_staticRoots.contains(id))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* Do periodic tasks such as database cleanup
|
|
*/
|
|
ZT_ALWAYS_INLINE void doPeriodicTasks(int64_t now)
|
|
{
|
|
{
|
|
Mutex::Lock _l1(_peers_l);
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(_peers);
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
if (!(*p)->alive(now)) {
|
|
_peers.erase(*a);
|
|
}
|
|
}
|
|
}
|
|
{
|
|
Mutex::Lock _l(_paths_l);
|
|
Hashtable< Path::HashKey,SharedPtr<Path> >::Iterator i(_paths);
|
|
Path::HashKey *k = (Path::HashKey *)0;
|
|
SharedPtr<Path> *p = (SharedPtr<Path> *)0;
|
|
while (i.next(k,p)) {
|
|
if (p->references() <= 1)
|
|
_paths.erase(*k);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param now Current time
|
|
* @return Number of peers with active direct paths
|
|
*/
|
|
inline unsigned long countActive(int64_t now) const
|
|
{
|
|
unsigned long cnt = 0;
|
|
Mutex::Lock _l(_peers_l);
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(const_cast<Topology *>(this)->_peers);
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
const SharedPtr<Path> pp((*p)->getAppropriatePath(now,false));
|
|
if (pp)
|
|
++cnt;
|
|
}
|
|
return cnt;
|
|
}
|
|
|
|
/**
|
|
* Apply a function or function object to all peers
|
|
*
|
|
* This locks the peer map during execution, so calls to get() etc. during
|
|
* eachPeer() will deadlock.
|
|
*
|
|
* @param f Function to apply
|
|
* @tparam F Function or function object type
|
|
*/
|
|
template<typename F>
|
|
ZT_ALWAYS_INLINE void eachPeer(F f)
|
|
{
|
|
Mutex::Lock l(_peers_l);
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(_peers);
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
if (!f(*((const SharedPtr<Peer> *)p)))
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Apply a function or function object to all roots
|
|
*
|
|
* This locks the root list during execution but other operations
|
|
* are fine.
|
|
*
|
|
* @param f Function to apply f(peer,IPs)
|
|
* @tparam F function or function object type
|
|
*/
|
|
template<typename F>
|
|
ZT_ALWAYS_INLINE void eachRoot(F f)
|
|
{
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
Hashtable< Str,Locator >::Iterator i(_dynamicRoots);
|
|
Str *k = (Str *)0;
|
|
Locator *v = (Locator *)0;
|
|
while (i.next(k,v)) {
|
|
if (*v) {
|
|
for(std::vector<Identity>::const_iterator id(v->virt().begin());id!=v->virt().end();++id) {
|
|
const SharedPtr<Peer> *ap;
|
|
{
|
|
Mutex::Lock l2(_peers_l);
|
|
ap = _peers.get(id->address());
|
|
}
|
|
if (ap) {
|
|
if (!f(*ap,v->phy()))
|
|
return;
|
|
} else {
|
|
SharedPtr<Peer> p(new Peer(RR,_myIdentity,*id));
|
|
{
|
|
Mutex::Lock l2(_peers_l);
|
|
_peers.set(id->address(),p);
|
|
}
|
|
if (!f(p,v->phy()))
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
{
|
|
Mutex::Lock l(_staticRoots_l);
|
|
Hashtable< Identity,std::vector<InetAddress> >::Iterator i(_staticRoots);
|
|
Identity *k = (Identity *)0;
|
|
std::vector<InetAddress> *v = (std::vector<InetAddress> *)0;
|
|
while (i.next(k,v)) {
|
|
if (!v->empty()) {
|
|
const SharedPtr<Peer> *ap;
|
|
{
|
|
Mutex::Lock l2(_peers_l);
|
|
ap = _peers.get(k->address());
|
|
}
|
|
if (ap) {
|
|
if (!f(*ap,*v))
|
|
return;
|
|
} else {
|
|
SharedPtr<Peer> p(new Peer(RR,_myIdentity,*k));
|
|
{
|
|
Mutex::Lock l2(_peers_l);
|
|
_peers.set(k->address(),p);
|
|
}
|
|
if (!f(p,*v))
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @return Current best root (updated automatically each second)
|
|
*/
|
|
inline SharedPtr<Peer> root(const int64_t now)
|
|
{
|
|
Mutex::Lock l(_bestRoot_l);
|
|
if ((!_bestRoot)||((now - _lastUpdatedBestRoot) > 1000)) {
|
|
_lastUpdatedBestRoot = now;
|
|
_RootRankingFunction rrf;
|
|
rrf.now = now;
|
|
eachRoot(rrf);
|
|
_bestRoot = rrf.bestRoot;
|
|
}
|
|
return _bestRoot;
|
|
}
|
|
|
|
/**
|
|
* Set or update a static root entry
|
|
*
|
|
* @param id Static root's identity
|
|
* @param addrs Static root's IP address(es)
|
|
*/
|
|
inline void setStaticRoot(const Identity &id,const std::vector<InetAddress> &addrs)
|
|
{
|
|
Mutex::Lock l(_staticRoots_l);
|
|
_staticRoots[id] = addrs;
|
|
}
|
|
|
|
/**
|
|
* Remove a static root
|
|
*
|
|
* @param id Identity to remove
|
|
*/
|
|
inline void removeStaticRoot(const Identity &id)
|
|
{
|
|
Mutex::Lock l(_staticRoots_l);
|
|
_staticRoots.erase(id);
|
|
}
|
|
|
|
/**
|
|
* Clear all static roots
|
|
*/
|
|
inline void removeStaticRoot()
|
|
{
|
|
Mutex::Lock l(_staticRoots_l);
|
|
_staticRoots.clear();
|
|
}
|
|
|
|
/**
|
|
* Iterate through all dynamic roots
|
|
*
|
|
* @param f Function of (Str,Locator)
|
|
*/
|
|
template<typename F>
|
|
ZT_ALWAYS_INLINE void eachDynamicRoot(F f) const
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
Str *k = (Str *)0;
|
|
Locator *v = (Locator *)0;
|
|
Hashtable< Str,Locator >::Iterator i(const_cast<Topology *>(this)->_dynamicRoots);
|
|
while (i.next(k,v)) {
|
|
if (!f(*k,*v))
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Set or update dynamic root if new locator is newer
|
|
*
|
|
* This does not check signatures or internal validity of the locator.
|
|
*
|
|
* @param dnsName DNS name used to retrive root
|
|
* @param latestLocator Latest locator
|
|
* @return True if locator is newer or if a new entry was created
|
|
*/
|
|
inline bool setDynamicRoot(const Str &dnsName,const Locator &latestLocator)
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
if (latestLocator) {
|
|
Locator &ll = _dynamicRoots[dnsName];
|
|
if (ll.timestamp() < latestLocator.timestamp()) {
|
|
ll = latestLocator;
|
|
_updateDynamicRootIdentities();
|
|
return true;
|
|
}
|
|
} else if (!_dynamicRoots.contains(dnsName)) {
|
|
_dynamicRoots[dnsName];
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* Remove a dynamic root entry
|
|
*
|
|
* @param dnsName DNS name to remove
|
|
*/
|
|
inline void removeDynamicRoot(const Str &dnsName)
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
_dynamicRoots.erase(dnsName);
|
|
_updateDynamicRootIdentities();
|
|
}
|
|
|
|
/**
|
|
* Remove all dynamic roots
|
|
*/
|
|
inline void clearDynamicRoots()
|
|
{
|
|
Mutex::Lock l(_dynamicRoots_l);
|
|
_dynamicRoots.clear();
|
|
_dynamicRootIdentities.clear();
|
|
}
|
|
|
|
/**
|
|
* @param Current time
|
|
* @return ZT_RootList as returned by the external CAPI
|
|
*/
|
|
inline ZT_RootList *apiRoots(const int64_t now) const
|
|
{
|
|
Mutex::Lock l1(_staticRoots_l);
|
|
Mutex::Lock l2(_dynamicRoots_l);
|
|
|
|
// The memory allocated here has room for all roots plus the maximum size
|
|
// of their DNS names, identities, and up to 16 physical addresses. Most
|
|
// roots will have two: one V4 and one V6.
|
|
const unsigned int totalRoots = _staticRoots.size() + _dynamicRoots.size();
|
|
ZT_RootList *rl = reinterpret_cast<ZT_RootList *>(malloc(sizeof(ZT_RootList) + (sizeof(ZT_Root) * totalRoots) + ((sizeof(struct sockaddr_storage) * ZT_MAX_PEER_NETWORK_PATHS) * totalRoots) + ((ZT_IDENTITY_STRING_BUFFER_LENGTH + 1024) * totalRoots)));
|
|
if (!rl) {
|
|
return nullptr;
|
|
}
|
|
|
|
unsigned int c = 0;
|
|
char *nameBufPtr = reinterpret_cast<char *>(rl) + sizeof(ZT_RootList) + (sizeof(ZT_Root) * totalRoots);
|
|
struct sockaddr_storage *addrBuf = reinterpret_cast<struct sockaddr_storage *>(nameBufPtr);
|
|
nameBufPtr += (sizeof(struct sockaddr_storage) * ZT_MAX_PEER_NETWORK_PATHS) * totalRoots;
|
|
|
|
_bestRoot_l.lock();
|
|
const Peer *const bestRootPtr = _bestRoot.ptr();
|
|
_bestRoot_l.unlock();
|
|
|
|
{
|
|
Str *k = (Str *)0;
|
|
Locator *v = (Locator *)0;
|
|
Hashtable< Str,Locator >::Iterator i(const_cast<Topology *>(this)->_dynamicRoots);
|
|
while (i.next(k,v)) {
|
|
rl->roots[c].dnsName = nameBufPtr;
|
|
const char *p = k->c_str();
|
|
while (*p)
|
|
*(nameBufPtr++) = *(p++);
|
|
*(nameBufPtr++) = (char)0;
|
|
|
|
if (v->id()) {
|
|
rl->roots[c].identity = nameBufPtr;
|
|
v->id().toString(false,nameBufPtr);
|
|
nameBufPtr += strlen(nameBufPtr) + 1;
|
|
}
|
|
|
|
rl->roots[c].addresses = addrBuf;
|
|
unsigned int ac = 0;
|
|
for(unsigned int j=(unsigned int)v->phy().size();(ac<j)&&(ac<16);++ac)
|
|
*(addrBuf++) = v->phy()[ac];
|
|
rl->roots[c].addressCount = ac;
|
|
|
|
_peers_l.lock();
|
|
const SharedPtr<Peer> *psptr = _peers.get(v->id().address());
|
|
if (psptr) {
|
|
rl->roots[c].preferred = (psptr->ptr() == bestRootPtr) ? 1 : 0;
|
|
rl->roots[c].online = (*psptr)->alive(now) ? 1 : 0;
|
|
}
|
|
_peers_l.unlock();
|
|
|
|
++c;
|
|
}
|
|
}
|
|
|
|
{
|
|
Hashtable< Identity,std::vector<InetAddress> >::Iterator i(const_cast<Topology *>(this)->_staticRoots);
|
|
Identity *k = (Identity *)0;
|
|
std::vector<InetAddress> *v = (std::vector<InetAddress> *)0;
|
|
while (i.next(k,v)) {
|
|
rl->roots[c].dnsName = nullptr;
|
|
|
|
rl->roots[c].identity = nameBufPtr;
|
|
k->toString(false,nameBufPtr);
|
|
nameBufPtr += strlen(nameBufPtr) + 1;
|
|
|
|
rl->roots[c].addresses = addrBuf;
|
|
unsigned int ac = 0;
|
|
for(unsigned int j=(unsigned int)v->size();(ac<j)&&(ac<16);++ac)
|
|
*(addrBuf++) = (*v)[ac];
|
|
rl->roots[c].addressCount = ac;
|
|
|
|
_peers_l.lock();
|
|
const SharedPtr<Peer> *psptr = _peers.get(k->address());
|
|
if (psptr) {
|
|
rl->roots[c].preferred = (psptr->ptr() == bestRootPtr) ? 1 : 0;
|
|
rl->roots[c].online = (*psptr)->alive(now) ? 1 : 0;
|
|
}
|
|
_peers_l.unlock();
|
|
|
|
++c;
|
|
}
|
|
}
|
|
|
|
rl->count = c;
|
|
|
|
return rl;
|
|
}
|
|
|
|
/**
|
|
* Get the best relay to a given address, which may or may not be a root
|
|
*
|
|
* @param now Current time
|
|
* @param toAddr Destination address
|
|
* @return Best current relay or NULL if none
|
|
*/
|
|
ZT_ALWAYS_INLINE SharedPtr<Peer> findRelayTo(const int64_t now,const Address &toAddr)
|
|
{
|
|
// TODO: in the future this will check 'mesh-like' relays and if enabled consult LF for other roots (for if this is a root)
|
|
return root(now);
|
|
}
|
|
|
|
/**
|
|
* @param allPeers vector to fill with all current peers
|
|
*/
|
|
ZT_ALWAYS_INLINE void getAllPeers(std::vector< SharedPtr<Peer> > &allPeers) const
|
|
{
|
|
Mutex::Lock l(_peers_l);
|
|
allPeers.clear();
|
|
allPeers.reserve(_peers.size());
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(*(const_cast<Hashtable< Address,SharedPtr<Peer> > *>(&_peers)));
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
allPeers.push_back(*p);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Get info about a path
|
|
*
|
|
* The supplied result variables are not modified if no special config info is found.
|
|
*
|
|
* @param physicalAddress Physical endpoint address
|
|
* @param mtu Variable set to MTU
|
|
* @param trustedPathId Variable set to trusted path ID
|
|
*/
|
|
ZT_ALWAYS_INLINE void getOutboundPathInfo(const InetAddress &physicalAddress,unsigned int &mtu,uint64_t &trustedPathId)
|
|
{
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress)) {
|
|
trustedPathId = _physicalPathConfig[i].second.trustedPathId;
|
|
mtu = _physicalPathConfig[i].second.mtu;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Get the payload MTU for an outbound physical path (returns default if not configured)
|
|
*
|
|
* @param physicalAddress Physical endpoint address
|
|
* @return MTU
|
|
*/
|
|
ZT_ALWAYS_INLINE unsigned int getOutboundPathMtu(const InetAddress &physicalAddress)
|
|
{
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress))
|
|
return _physicalPathConfig[i].second.mtu;
|
|
}
|
|
return ZT_DEFAULT_PHYSMTU;
|
|
}
|
|
|
|
/**
|
|
* Get the outbound trusted path ID for a physical address, or 0 if none
|
|
*
|
|
* @param physicalAddress Physical address to which we are sending the packet
|
|
* @return Trusted path ID or 0 if none (0 is not a valid trusted path ID)
|
|
*/
|
|
ZT_ALWAYS_INLINE uint64_t getOutboundPathTrust(const InetAddress &physicalAddress)
|
|
{
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress))
|
|
return _physicalPathConfig[i].second.trustedPathId;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Check whether in incoming trusted path marked packet is valid
|
|
*
|
|
* @param physicalAddress Originating physical address
|
|
* @param trustedPathId Trusted path ID from packet (from MAC field)
|
|
*/
|
|
ZT_ALWAYS_INLINE bool shouldInboundPathBeTrusted(const InetAddress &physicalAddress,const uint64_t trustedPathId)
|
|
{
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
if ((_physicalPathConfig[i].second.trustedPathId == trustedPathId)&&(_physicalPathConfig[i].first.containsAddress(physicalAddress)))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* Set or clear physical path configuration (called via Node::setPhysicalPathConfiguration)
|
|
*/
|
|
inline void setPhysicalPathConfiguration(const struct sockaddr_storage *pathNetwork,const ZT_PhysicalPathConfiguration *pathConfig)
|
|
{
|
|
if (!pathNetwork) {
|
|
_numConfiguredPhysicalPaths = 0;
|
|
} else {
|
|
std::map<InetAddress,ZT_PhysicalPathConfiguration> cpaths;
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i)
|
|
cpaths[_physicalPathConfig[i].first] = _physicalPathConfig[i].second;
|
|
|
|
if (pathConfig) {
|
|
ZT_PhysicalPathConfiguration pc(*pathConfig);
|
|
|
|
if (pc.mtu <= 0)
|
|
pc.mtu = ZT_DEFAULT_PHYSMTU;
|
|
else if (pc.mtu < ZT_MIN_PHYSMTU)
|
|
pc.mtu = ZT_MIN_PHYSMTU;
|
|
else if (pc.mtu > ZT_MAX_PHYSMTU)
|
|
pc.mtu = ZT_MAX_PHYSMTU;
|
|
|
|
cpaths[*(reinterpret_cast<const InetAddress *>(pathNetwork))] = pc;
|
|
} else {
|
|
cpaths.erase(*(reinterpret_cast<const InetAddress *>(pathNetwork)));
|
|
}
|
|
|
|
unsigned int cnt = 0;
|
|
for(std::map<InetAddress,ZT_PhysicalPathConfiguration>::const_iterator i(cpaths.begin());((i!=cpaths.end())&&(cnt<ZT_MAX_CONFIGURABLE_PATHS));++i) {
|
|
_physicalPathConfig[cnt].first = i->first;
|
|
_physicalPathConfig[cnt].second = i->second;
|
|
++cnt;
|
|
}
|
|
_numConfiguredPhysicalPaths = cnt;
|
|
}
|
|
}
|
|
|
|
private:
|
|
const RuntimeEnvironment *const RR;
|
|
const Identity _myIdentity;
|
|
|
|
std::pair<InetAddress,ZT_PhysicalPathConfiguration> _physicalPathConfig[ZT_MAX_CONFIGURABLE_PATHS];
|
|
unsigned int _numConfiguredPhysicalPaths;
|
|
|
|
Hashtable< Address,SharedPtr<Peer> > _peers;
|
|
Hashtable< Path::HashKey,SharedPtr<Path> > _paths;
|
|
|
|
Hashtable< Str,Locator > _dynamicRoots;
|
|
Hashtable< Identity,bool > _dynamicRootIdentities;
|
|
Hashtable< Identity,std::vector<InetAddress> > _staticRoots;
|
|
|
|
int64_t _lastUpdatedBestRoot;
|
|
SharedPtr<Peer> _bestRoot;
|
|
|
|
Mutex _peers_l;
|
|
Mutex _paths_l;
|
|
Mutex _dynamicRoots_l;
|
|
Mutex _staticRoots_l;
|
|
Mutex _bestRoot_l;
|
|
};
|
|
|
|
} // namespace ZeroTier
|
|
|
|
#endif
|