2013-07-04 20:56:19 +00:00
|
|
|
/*
|
2015-02-17 21:11:34 +00:00
|
|
|
* ZeroTier One - Network Virtualization Everywhere
|
2019-01-14 18:25:53 +00:00
|
|
|
* Copyright (C) 2011-2019 ZeroTier, Inc. https://www.zerotier.com/
|
2013-07-04 20:56:19 +00:00
|
|
|
*
|
|
|
|
* This program is free software: you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License as published by
|
|
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
|
|
* (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
2019-01-14 18:25:53 +00:00
|
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2017-04-28 03:47:25 +00:00
|
|
|
*
|
|
|
|
* --
|
|
|
|
*
|
|
|
|
* You can be released from the requirements of the license by purchasing
|
|
|
|
* a commercial license. Buying such a license is mandatory as soon as you
|
|
|
|
* develop commercial closed-source software that incorporates or links
|
|
|
|
* directly against ZeroTier software without disclosing the source code
|
|
|
|
* of your own application.
|
2013-07-04 20:56:19 +00:00
|
|
|
*/
|
|
|
|
|
2013-12-07 00:49:20 +00:00
|
|
|
#ifndef ZT_TOPOLOGY_HPP
|
|
|
|
#define ZT_TOPOLOGY_HPP
|
2013-07-04 20:56:19 +00:00
|
|
|
|
2013-07-06 18:58:34 +00:00
|
|
|
#include <stdio.h>
|
|
|
|
#include <string.h>
|
2013-09-17 19:53:59 +00:00
|
|
|
|
2013-07-04 20:56:19 +00:00
|
|
|
#include <vector>
|
|
|
|
#include <stdexcept>
|
2014-10-14 23:38:27 +00:00
|
|
|
#include <algorithm>
|
2015-10-13 15:49:36 +00:00
|
|
|
#include <utility>
|
2013-07-04 20:56:19 +00:00
|
|
|
|
2013-10-05 14:19:12 +00:00
|
|
|
#include "Constants.hpp"
|
2016-07-12 15:29:50 +00:00
|
|
|
#include "../include/ZeroTierOne.h"
|
2014-08-14 23:52:22 +00:00
|
|
|
|
2013-07-04 20:56:19 +00:00
|
|
|
#include "Address.hpp"
|
2014-08-14 23:52:22 +00:00
|
|
|
#include "Identity.hpp"
|
2013-07-04 20:56:19 +00:00
|
|
|
#include "Peer.hpp"
|
2016-09-01 22:43:07 +00:00
|
|
|
#include "Path.hpp"
|
2013-07-04 20:56:19 +00:00
|
|
|
#include "Mutex.hpp"
|
|
|
|
#include "InetAddress.hpp"
|
2015-09-04 19:14:21 +00:00
|
|
|
#include "Hashtable.hpp"
|
2013-07-04 20:56:19 +00:00
|
|
|
|
|
|
|
namespace ZeroTier {
|
|
|
|
|
|
|
|
class RuntimeEnvironment;
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Database of network topology
|
|
|
|
*/
|
2013-08-05 20:06:16 +00:00
|
|
|
class Topology
|
2013-07-04 20:56:19 +00:00
|
|
|
{
|
|
|
|
public:
|
2019-08-14 17:35:57 +00:00
|
|
|
inline Topology(const RuntimeEnvironment *renv,void *tPtr) :
|
|
|
|
RR(renv),
|
|
|
|
_numConfiguredPhysicalPaths(0) {}
|
|
|
|
inline ~Topology() {}
|
2013-07-04 20:56:19 +00:00
|
|
|
|
|
|
|
/**
|
2013-10-05 14:19:12 +00:00
|
|
|
* Add a peer to database
|
2013-07-04 20:56:19 +00:00
|
|
|
*
|
2013-10-05 14:19:12 +00:00
|
|
|
* This will not replace existing peers. In that case the existing peer
|
|
|
|
* record is returned.
|
|
|
|
*
|
2017-03-28 00:03:17 +00:00
|
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
2013-10-05 14:19:12 +00:00
|
|
|
* @param peer Peer to add
|
|
|
|
* @return New or existing peer (should replace 'peer')
|
2013-07-04 20:56:19 +00:00
|
|
|
*/
|
2019-08-14 17:35:57 +00:00
|
|
|
inline SharedPtr<Peer> addPeer(void *tPtr,const SharedPtr<Peer> &peer)
|
|
|
|
{
|
|
|
|
SharedPtr<Peer> np;
|
|
|
|
{
|
|
|
|
Mutex::Lock _l(_peers_m);
|
|
|
|
SharedPtr<Peer> &hp = _peers[peer->address()];
|
|
|
|
if (!hp)
|
|
|
|
hp = peer;
|
|
|
|
np = hp;
|
|
|
|
}
|
|
|
|
return np;
|
|
|
|
}
|
2013-07-04 20:56:19 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Get a peer from its address
|
2015-07-28 00:02:43 +00:00
|
|
|
*
|
2017-03-28 00:03:17 +00:00
|
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
2013-07-04 20:56:19 +00:00
|
|
|
* @param zta ZeroTier address of peer
|
|
|
|
* @return Peer or NULL if not found
|
|
|
|
*/
|
2019-08-14 17:35:57 +00:00
|
|
|
inline SharedPtr<Peer> getPeer(void *tPtr,const Address &zta) const
|
|
|
|
{
|
|
|
|
if (zta == RR->identity.address())
|
|
|
|
return SharedPtr<Peer>();
|
2019-08-14 19:48:45 +00:00
|
|
|
Mutex::Lock _l(_peers_m);
|
|
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
|
|
|
return ((ap) ? *ap : SharedPtr<Peer>());
|
2019-08-14 17:35:57 +00:00
|
|
|
}
|
2013-10-21 18:12:00 +00:00
|
|
|
|
2017-07-06 17:25:36 +00:00
|
|
|
/**
|
|
|
|
* @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call
|
|
|
|
* @param zta ZeroTier address of peer
|
|
|
|
* @return Identity or NULL identity if not found
|
|
|
|
*/
|
2019-08-14 17:35:57 +00:00
|
|
|
inline Identity getIdentity(void *tPtr,const Address &zta)
|
|
|
|
{
|
|
|
|
if (zta == RR->identity.address()) {
|
|
|
|
return RR->identity;
|
|
|
|
} else {
|
|
|
|
Mutex::Lock _l(_peers_m);
|
|
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
|
|
|
if (ap)
|
|
|
|
return (*ap)->identity();
|
|
|
|
}
|
|
|
|
return Identity();
|
|
|
|
}
|
2019-08-14 19:48:45 +00:00
|
|
|
|
2015-10-27 23:52:44 +00:00
|
|
|
/**
|
|
|
|
* Get a peer only if it is presently in memory (no disk cache)
|
|
|
|
*
|
2015-11-02 23:38:53 +00:00
|
|
|
* This also does not update the lastUsed() time for peers, which means
|
|
|
|
* that it won't prevent them from falling out of RAM. This is currently
|
|
|
|
* used in the Cluster code to update peer info without forcing all peers
|
|
|
|
* across the entire cluster to remain in memory cache.
|
|
|
|
*
|
2015-10-27 23:52:44 +00:00
|
|
|
* @param zta ZeroTier address
|
|
|
|
*/
|
2015-11-06 01:22:22 +00:00
|
|
|
inline SharedPtr<Peer> getPeerNoCache(const Address &zta)
|
2015-10-27 23:52:44 +00:00
|
|
|
{
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex::Lock _l(_peers_m);
|
2015-10-30 20:39:28 +00:00
|
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
2015-11-02 23:38:53 +00:00
|
|
|
if (ap)
|
2015-10-27 23:52:44 +00:00
|
|
|
return *ap;
|
|
|
|
return SharedPtr<Peer>();
|
|
|
|
}
|
|
|
|
|
2016-09-01 22:43:07 +00:00
|
|
|
/**
|
|
|
|
* Get a Path object for a given local and remote physical address, creating if needed
|
|
|
|
*
|
2017-07-06 18:45:22 +00:00
|
|
|
* @param l Local socket
|
2016-09-01 22:43:07 +00:00
|
|
|
* @param r Remote address
|
|
|
|
* @return Pointer to canonicalized Path object
|
|
|
|
*/
|
2017-07-06 18:45:22 +00:00
|
|
|
inline SharedPtr<Path> getPath(const int64_t l,const InetAddress &r)
|
2016-09-01 22:43:07 +00:00
|
|
|
{
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex::Lock _l(_paths_m);
|
2016-09-01 22:43:07 +00:00
|
|
|
SharedPtr<Path> &p = _paths[Path::HashKey(l,r)];
|
|
|
|
if (!p)
|
2018-04-06 15:10:34 +00:00
|
|
|
p.set(new Path(l,r));
|
2016-09-01 22:43:07 +00:00
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
2019-08-14 17:35:57 +00:00
|
|
|
inline SharedPtr<Peer> getUpstreamPeer() const
|
|
|
|
{
|
2019-08-14 19:48:45 +00:00
|
|
|
// TODO
|
2019-08-14 17:35:57 +00:00
|
|
|
return SharedPtr<Peer>();
|
|
|
|
}
|
2016-11-18 00:20:41 +00:00
|
|
|
|
2019-08-14 17:35:57 +00:00
|
|
|
inline bool isUpstream(const Identity &id) const
|
|
|
|
{
|
2019-08-14 19:48:45 +00:00
|
|
|
// TODO
|
2019-08-14 17:35:57 +00:00
|
|
|
return false;
|
|
|
|
}
|
2019-08-14 19:48:45 +00:00
|
|
|
|
2019-08-14 17:35:57 +00:00
|
|
|
inline ZT_PeerRole role(const Address &ztaddr) const
|
|
|
|
{
|
2019-08-14 19:48:45 +00:00
|
|
|
// TODO
|
2019-08-14 17:35:57 +00:00
|
|
|
return ZT_PEER_ROLE_LEAF;
|
|
|
|
}
|
2016-12-06 00:09:42 +00:00
|
|
|
|
2019-08-14 19:48:45 +00:00
|
|
|
inline void getAlwaysContact(Hashtable< Address,std::vector<InetAddress> > &eps) const
|
2017-01-27 21:27:52 +00:00
|
|
|
{
|
2019-08-14 19:48:45 +00:00
|
|
|
// TODO
|
2017-01-27 21:27:52 +00:00
|
|
|
}
|
|
|
|
|
2016-08-04 17:18:33 +00:00
|
|
|
inline std::vector<Address> upstreamAddresses() const
|
|
|
|
{
|
2019-08-06 21:46:13 +00:00
|
|
|
// TODO
|
2019-07-17 15:52:08 +00:00
|
|
|
return std::vector<Address>();
|
2015-10-13 15:49:36 +00:00
|
|
|
}
|
|
|
|
|
2019-08-14 17:35:57 +00:00
|
|
|
inline void doPeriodicTasks(void *tPtr,int64_t now)
|
|
|
|
{
|
|
|
|
{
|
|
|
|
Mutex::Lock _l1(_peers_m);
|
|
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(_peers);
|
|
|
|
Address *a = (Address *)0;
|
|
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
|
|
while (i.next(a,p)) {
|
|
|
|
if (!(*p)->isAlive(now)) {
|
|
|
|
_peers.erase(*a);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
{
|
|
|
|
Mutex::Lock _l(_paths_m);
|
|
|
|
Hashtable< Path::HashKey,SharedPtr<Path> >::Iterator i(_paths);
|
|
|
|
Path::HashKey *k = (Path::HashKey *)0;
|
|
|
|
SharedPtr<Path> *p = (SharedPtr<Path> *)0;
|
|
|
|
while (i.next(k,p)) {
|
|
|
|
if (p->references() <= 1)
|
|
|
|
_paths.erase(*k);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2013-07-04 20:56:19 +00:00
|
|
|
|
2015-10-26 19:41:08 +00:00
|
|
|
/**
|
2015-11-13 20:14:28 +00:00
|
|
|
* @param now Current time
|
2015-10-28 21:29:08 +00:00
|
|
|
* @return Number of peers with active direct paths
|
2015-10-26 19:41:08 +00:00
|
|
|
*/
|
2017-10-02 22:52:57 +00:00
|
|
|
inline unsigned long countActive(int64_t now) const
|
2015-11-13 20:14:28 +00:00
|
|
|
{
|
|
|
|
unsigned long cnt = 0;
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex::Lock _l(_peers_m);
|
2015-11-13 20:14:28 +00:00
|
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(const_cast<Topology *>(this)->_peers);
|
|
|
|
Address *a = (Address *)0;
|
|
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
|
|
while (i.next(a,p)) {
|
2018-05-01 23:32:15 +00:00
|
|
|
const SharedPtr<Path> pp((*p)->getAppropriatePath(now,false));
|
2017-10-25 19:42:14 +00:00
|
|
|
if (pp)
|
2017-04-15 00:53:32 +00:00
|
|
|
++cnt;
|
2015-11-13 20:14:28 +00:00
|
|
|
}
|
|
|
|
return cnt;
|
|
|
|
}
|
2015-10-26 19:41:08 +00:00
|
|
|
|
2013-07-04 20:56:19 +00:00
|
|
|
/**
|
|
|
|
* Apply a function or function object to all peers
|
|
|
|
*
|
|
|
|
* @param f Function to apply
|
|
|
|
* @tparam F Function or function object type
|
|
|
|
*/
|
|
|
|
template<typename F>
|
|
|
|
inline void eachPeer(F f)
|
|
|
|
{
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex::Lock _l(_peers_m);
|
2015-10-02 00:09:01 +00:00
|
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(_peers);
|
2015-09-04 19:14:21 +00:00
|
|
|
Address *a = (Address *)0;
|
|
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
2015-10-30 21:11:10 +00:00
|
|
|
while (i.next(a,p)) {
|
|
|
|
f(*this,*((const SharedPtr<Peer> *)p));
|
|
|
|
}
|
2013-07-04 20:56:19 +00:00
|
|
|
}
|
|
|
|
|
2015-04-09 01:45:21 +00:00
|
|
|
/**
|
2019-08-14 17:35:57 +00:00
|
|
|
* @return All peers by address (unsorted)
|
2015-04-09 01:45:21 +00:00
|
|
|
*/
|
2015-09-04 19:14:21 +00:00
|
|
|
inline std::vector< std::pair< Address,SharedPtr<Peer> > > allPeers() const
|
2015-04-09 01:45:21 +00:00
|
|
|
{
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex::Lock _l(_peers_m);
|
2015-10-02 00:09:01 +00:00
|
|
|
return _peers.entries();
|
2015-04-09 01:45:21 +00:00
|
|
|
}
|
|
|
|
|
2017-09-01 23:25:34 +00:00
|
|
|
/**
|
|
|
|
* Get info about a path
|
|
|
|
*
|
|
|
|
* The supplied result variables are not modified if no special config info is found.
|
|
|
|
*
|
|
|
|
* @param physicalAddress Physical endpoint address
|
|
|
|
* @param mtu Variable set to MTU
|
|
|
|
* @param trustedPathId Variable set to trusted path ID
|
|
|
|
*/
|
|
|
|
inline void getOutboundPathInfo(const InetAddress &physicalAddress,unsigned int &mtu,uint64_t &trustedPathId)
|
|
|
|
{
|
|
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress)) {
|
|
|
|
trustedPathId = _physicalPathConfig[i].second.trustedPathId;
|
|
|
|
mtu = _physicalPathConfig[i].second.mtu;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Get the payload MTU for an outbound physical path (returns default if not configured)
|
|
|
|
*
|
|
|
|
* @param physicalAddress Physical endpoint address
|
|
|
|
* @return MTU
|
|
|
|
*/
|
|
|
|
inline unsigned int getOutboundPathMtu(const InetAddress &physicalAddress)
|
|
|
|
{
|
|
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress))
|
|
|
|
return _physicalPathConfig[i].second.mtu;
|
|
|
|
}
|
|
|
|
return ZT_DEFAULT_PHYSMTU;
|
|
|
|
}
|
|
|
|
|
2016-07-12 15:29:50 +00:00
|
|
|
/**
|
|
|
|
* Get the outbound trusted path ID for a physical address, or 0 if none
|
|
|
|
*
|
|
|
|
* @param physicalAddress Physical address to which we are sending the packet
|
|
|
|
* @return Trusted path ID or 0 if none (0 is not a valid trusted path ID)
|
|
|
|
*/
|
|
|
|
inline uint64_t getOutboundPathTrust(const InetAddress &physicalAddress)
|
|
|
|
{
|
2017-09-01 23:25:34 +00:00
|
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
|
|
if (_physicalPathConfig[i].first.containsAddress(physicalAddress))
|
|
|
|
return _physicalPathConfig[i].second.trustedPathId;
|
2016-07-12 15:29:50 +00:00
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Check whether in incoming trusted path marked packet is valid
|
|
|
|
*
|
|
|
|
* @param physicalAddress Originating physical address
|
|
|
|
* @param trustedPathId Trusted path ID from packet (from MAC field)
|
|
|
|
*/
|
|
|
|
inline bool shouldInboundPathBeTrusted(const InetAddress &physicalAddress,const uint64_t trustedPathId)
|
|
|
|
{
|
2017-09-01 23:25:34 +00:00
|
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i) {
|
|
|
|
if ((_physicalPathConfig[i].second.trustedPathId == trustedPathId)&&(_physicalPathConfig[i].first.containsAddress(physicalAddress)))
|
2016-07-12 15:29:50 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2017-09-01 23:25:34 +00:00
|
|
|
* Set or clear physical path configuration (called via Node::setPhysicalPathConfiguration)
|
2016-07-12 15:29:50 +00:00
|
|
|
*/
|
2019-08-14 17:35:57 +00:00
|
|
|
inline void setPhysicalPathConfiguration(const struct sockaddr_storage *pathNetwork,const ZT_PhysicalPathConfiguration *pathConfig)
|
|
|
|
{
|
|
|
|
if (!pathNetwork) {
|
|
|
|
_numConfiguredPhysicalPaths = 0;
|
|
|
|
} else {
|
|
|
|
std::map<InetAddress,ZT_PhysicalPathConfiguration> cpaths;
|
|
|
|
for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i<j;++i)
|
|
|
|
cpaths[_physicalPathConfig[i].first] = _physicalPathConfig[i].second;
|
|
|
|
|
|
|
|
if (pathConfig) {
|
|
|
|
ZT_PhysicalPathConfiguration pc(*pathConfig);
|
|
|
|
|
|
|
|
if (pc.mtu <= 0)
|
|
|
|
pc.mtu = ZT_DEFAULT_PHYSMTU;
|
|
|
|
else if (pc.mtu < ZT_MIN_PHYSMTU)
|
|
|
|
pc.mtu = ZT_MIN_PHYSMTU;
|
|
|
|
else if (pc.mtu > ZT_MAX_PHYSMTU)
|
|
|
|
pc.mtu = ZT_MAX_PHYSMTU;
|
|
|
|
|
|
|
|
cpaths[*(reinterpret_cast<const InetAddress *>(pathNetwork))] = pc;
|
|
|
|
} else {
|
|
|
|
cpaths.erase(*(reinterpret_cast<const InetAddress *>(pathNetwork)));
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned int cnt = 0;
|
|
|
|
for(std::map<InetAddress,ZT_PhysicalPathConfiguration>::const_iterator i(cpaths.begin());((i!=cpaths.end())&&(cnt<ZT_MAX_CONFIGURABLE_PATHS));++i) {
|
|
|
|
_physicalPathConfig[cnt].first = i->first;
|
|
|
|
_physicalPathConfig[cnt].second = i->second;
|
|
|
|
++cnt;
|
|
|
|
}
|
|
|
|
_numConfiguredPhysicalPaths = cnt;
|
|
|
|
}
|
|
|
|
}
|
2016-07-12 15:29:50 +00:00
|
|
|
|
2013-07-04 20:56:19 +00:00
|
|
|
private:
|
2015-11-13 20:14:28 +00:00
|
|
|
const RuntimeEnvironment *const RR;
|
2017-09-01 23:25:34 +00:00
|
|
|
std::pair<InetAddress,ZT_PhysicalPathConfiguration> _physicalPathConfig[ZT_MAX_CONFIGURABLE_PATHS];
|
2019-08-06 21:46:13 +00:00
|
|
|
unsigned int _numConfiguredPhysicalPaths;
|
2015-10-02 00:09:01 +00:00
|
|
|
Hashtable< Address,SharedPtr<Peer> > _peers;
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex _peers_m;
|
2016-09-01 22:43:07 +00:00
|
|
|
Hashtable< Path::HashKey,SharedPtr<Path> > _paths;
|
2017-01-28 00:16:06 +00:00
|
|
|
Mutex _paths_m;
|
2013-07-04 20:56:19 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace ZeroTier
|
|
|
|
|
|
|
|
#endif
|