mirror of
https://github.com/zerotier/ZeroTierOne.git
synced 2024-12-22 14:22:23 +00:00
334 lines
8.8 KiB
C++
334 lines
8.8 KiB
C++
/*
|
|
* ZeroTier One - Network Virtualization Everywhere
|
|
* Copyright (C) 2011-2016 ZeroTier, Inc. https://www.zerotier.com/
|
|
*
|
|
* This program is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#ifndef ZT_TOPOLOGY_HPP
|
|
#define ZT_TOPOLOGY_HPP
|
|
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
|
|
#include <vector>
|
|
#include <stdexcept>
|
|
#include <algorithm>
|
|
#include <utility>
|
|
|
|
#include "Constants.hpp"
|
|
#include "../include/ZeroTierOne.h"
|
|
|
|
#include "Address.hpp"
|
|
#include "Identity.hpp"
|
|
#include "Peer.hpp"
|
|
#include "Mutex.hpp"
|
|
#include "InetAddress.hpp"
|
|
#include "Hashtable.hpp"
|
|
#include "World.hpp"
|
|
|
|
namespace ZeroTier {
|
|
|
|
class RuntimeEnvironment;
|
|
|
|
/**
|
|
* Database of network topology
|
|
*/
|
|
class Topology
|
|
{
|
|
public:
|
|
Topology(const RuntimeEnvironment *renv);
|
|
~Topology();
|
|
|
|
/**
|
|
* Add a peer to database
|
|
*
|
|
* This will not replace existing peers. In that case the existing peer
|
|
* record is returned.
|
|
*
|
|
* @param peer Peer to add
|
|
* @return New or existing peer (should replace 'peer')
|
|
*/
|
|
SharedPtr<Peer> addPeer(const SharedPtr<Peer> &peer);
|
|
|
|
/**
|
|
* Get a peer from its address
|
|
*
|
|
* @param zta ZeroTier address of peer
|
|
* @return Peer or NULL if not found
|
|
*/
|
|
SharedPtr<Peer> getPeer(const Address &zta);
|
|
|
|
/**
|
|
* Get a peer only if it is presently in memory (no disk cache)
|
|
*
|
|
* This also does not update the lastUsed() time for peers, which means
|
|
* that it won't prevent them from falling out of RAM. This is currently
|
|
* used in the Cluster code to update peer info without forcing all peers
|
|
* across the entire cluster to remain in memory cache.
|
|
*
|
|
* @param zta ZeroTier address
|
|
*/
|
|
inline SharedPtr<Peer> getPeerNoCache(const Address &zta)
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
const SharedPtr<Peer> *const ap = _peers.get(zta);
|
|
if (ap)
|
|
return *ap;
|
|
return SharedPtr<Peer>();
|
|
}
|
|
|
|
/**
|
|
* Get the identity of a peer
|
|
*
|
|
* @param zta ZeroTier address of peer
|
|
* @return Identity or NULL Identity if not found
|
|
*/
|
|
Identity getIdentity(const Address &zta);
|
|
|
|
/**
|
|
* Cache an identity
|
|
*
|
|
* This is done automatically on addPeer(), and so is only useful for
|
|
* cluster identity replication.
|
|
*
|
|
* @param id Identity to cache
|
|
*/
|
|
void saveIdentity(const Identity &id);
|
|
|
|
/**
|
|
* Get the current favorite root server
|
|
*
|
|
* @return Root server with lowest latency or NULL if none
|
|
*/
|
|
inline SharedPtr<Peer> getBestRoot() { return getBestRoot((const Address *)0,0,false); }
|
|
|
|
/**
|
|
* Get the best root server, avoiding root servers listed in an array
|
|
*
|
|
* This will get the best root server (lowest latency, etc.) but will
|
|
* try to avoid the listed root servers, only using them if no others
|
|
* are available.
|
|
*
|
|
* @param avoid Nodes to avoid
|
|
* @param avoidCount Number of nodes to avoid
|
|
* @param strictAvoid If false, consider avoided root servers anyway if no non-avoid root servers are available
|
|
* @return Root server or NULL if none available
|
|
*/
|
|
SharedPtr<Peer> getBestRoot(const Address *avoid,unsigned int avoidCount,bool strictAvoid);
|
|
|
|
/**
|
|
* @param id Identity to check
|
|
* @return True if this is a designated root server in this world
|
|
*/
|
|
inline bool isRoot(const Identity &id) const
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
return (std::find(_rootAddresses.begin(),_rootAddresses.end(),id.address()) != _rootAddresses.end());
|
|
}
|
|
|
|
/**
|
|
* @param id Identity to check
|
|
* @return True if this is a root server or a network preferred relay from one of our networks
|
|
*/
|
|
bool isUpstream(const Identity &id) const;
|
|
|
|
/**
|
|
* @return Vector of root server addresses
|
|
*/
|
|
inline std::vector<Address> rootAddresses() const
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
return _rootAddresses;
|
|
}
|
|
|
|
/**
|
|
* @return Vector of active upstream addresses (including roots)
|
|
*/
|
|
inline std::vector<Address> upstreamAddresses() const
|
|
{
|
|
return rootAddresses();
|
|
}
|
|
|
|
/**
|
|
* @return Current World (copy)
|
|
*/
|
|
inline World world() const
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
return _world;
|
|
}
|
|
|
|
/**
|
|
* @return Current world ID
|
|
*/
|
|
inline uint64_t worldId() const
|
|
{
|
|
return _world.id(); // safe to read without lock, and used from within eachPeer() so don't lock
|
|
}
|
|
|
|
/**
|
|
* @return Current world timestamp
|
|
*/
|
|
inline uint64_t worldTimestamp() const
|
|
{
|
|
return _world.timestamp(); // safe to read without lock, and used from within eachPeer() so don't lock
|
|
}
|
|
|
|
/**
|
|
* Validate new world and update if newer and signature is okay
|
|
*
|
|
* @param newWorld Potential new world definition revision
|
|
* @return True if an update actually occurred
|
|
*/
|
|
bool worldUpdateIfValid(const World &newWorld);
|
|
|
|
/**
|
|
* Clean and flush database
|
|
*/
|
|
void clean(uint64_t now);
|
|
|
|
/**
|
|
* @param now Current time
|
|
* @return Number of peers with active direct paths
|
|
*/
|
|
inline unsigned long countActive(uint64_t now) const
|
|
{
|
|
unsigned long cnt = 0;
|
|
Mutex::Lock _l(_lock);
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(const_cast<Topology *>(this)->_peers);
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
cnt += (unsigned long)((*p)->hasActiveDirectPath(now));
|
|
}
|
|
return cnt;
|
|
}
|
|
|
|
/**
|
|
* Apply a function or function object to all peers
|
|
*
|
|
* Note: explicitly template this by reference if you want the object
|
|
* passed by reference instead of copied.
|
|
*
|
|
* Warning: be careful not to use features in these that call any other
|
|
* methods of Topology that may lock _lock, otherwise a recursive lock
|
|
* and deadlock or lock corruption may occur.
|
|
*
|
|
* @param f Function to apply
|
|
* @tparam F Function or function object type
|
|
*/
|
|
template<typename F>
|
|
inline void eachPeer(F f)
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
Hashtable< Address,SharedPtr<Peer> >::Iterator i(_peers);
|
|
Address *a = (Address *)0;
|
|
SharedPtr<Peer> *p = (SharedPtr<Peer> *)0;
|
|
while (i.next(a,p)) {
|
|
#ifdef ZT_TRACE
|
|
if (!(*p)) {
|
|
fprintf(stderr,"FATAL BUG: eachPeer() caught NULL peer for %s -- peer pointers in Topology should NEVER be NULL" ZT_EOL_S,a->toString().c_str());
|
|
abort();
|
|
}
|
|
#endif
|
|
f(*this,*((const SharedPtr<Peer> *)p));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @return All currently active peers by address (unsorted)
|
|
*/
|
|
inline std::vector< std::pair< Address,SharedPtr<Peer> > > allPeers() const
|
|
{
|
|
Mutex::Lock _l(_lock);
|
|
return _peers.entries();
|
|
}
|
|
|
|
/**
|
|
* @return True if I am a root server in the current World
|
|
*/
|
|
inline bool amRoot() const throw() { return _amRoot; }
|
|
|
|
/**
|
|
* Get the outbound trusted path ID for a physical address, or 0 if none
|
|
*
|
|
* @param physicalAddress Physical address to which we are sending the packet
|
|
* @return Trusted path ID or 0 if none (0 is not a valid trusted path ID)
|
|
*/
|
|
inline uint64_t getOutboundPathTrust(const InetAddress &physicalAddress)
|
|
{
|
|
for(unsigned int i=0;i<_trustedPathCount;++i) {
|
|
if (_trustedPathNetworks[i].containsAddress(physicalAddress))
|
|
return _trustedPathIds[i];
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Check whether in incoming trusted path marked packet is valid
|
|
*
|
|
* @param physicalAddress Originating physical address
|
|
* @param trustedPathId Trusted path ID from packet (from MAC field)
|
|
*/
|
|
inline bool shouldInboundPathBeTrusted(const InetAddress &physicalAddress,const uint64_t trustedPathId)
|
|
{
|
|
for(unsigned int i=0;i<_trustedPathCount;++i) {
|
|
if ((_trustedPathIds[i] == trustedPathId)&&(_trustedPathNetworks[i].containsAddress(physicalAddress)))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* Set trusted paths in this topology
|
|
*
|
|
* @param networks Array of networks (prefix/netmask bits)
|
|
* @param ids Array of trusted path IDs
|
|
* @param count Number of trusted paths (if larger than ZT_MAX_TRUSTED_PATHS overflow is ignored)
|
|
*/
|
|
inline void setTrustedPaths(const InetAddress *networks,const uint64_t *ids,unsigned int count)
|
|
{
|
|
if (count > ZT_MAX_TRUSTED_PATHS)
|
|
count = ZT_MAX_TRUSTED_PATHS;
|
|
Mutex::Lock _l(_lock);
|
|
for(unsigned int i=0;i<count;++i) {
|
|
_trustedPathIds[i] = ids[i];
|
|
_trustedPathNetworks[i] = networks[i];
|
|
}
|
|
_trustedPathCount = count;
|
|
}
|
|
|
|
private:
|
|
Identity _getIdentity(const Address &zta);
|
|
void _setWorld(const World &newWorld);
|
|
|
|
const RuntimeEnvironment *const RR;
|
|
|
|
uint64_t _trustedPathIds[ZT_MAX_TRUSTED_PATHS];
|
|
InetAddress _trustedPathNetworks[ZT_MAX_TRUSTED_PATHS];
|
|
unsigned int _trustedPathCount;
|
|
World _world;
|
|
Hashtable< Address,SharedPtr<Peer> > _peers;
|
|
std::vector< Address > _rootAddresses;
|
|
std::vector< SharedPtr<Peer> > _rootPeers;
|
|
bool _amRoot;
|
|
|
|
Mutex _lock;
|
|
};
|
|
|
|
} // namespace ZeroTier
|
|
|
|
#endif
|