/* * ZeroTier One - Network Virtualization Everywhere * Copyright (C) 2011-2015 ZeroTier, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see . * * -- * * ZeroTier may be used and distributed under the terms of the GPLv3, which * are available at: http://www.gnu.org/licenses/gpl-3.0.html * * If you would like to embed ZeroTier into a commercial application or * redistribute it in a modified binary form, please contact ZeroTier Networks * LLC. Start here: http://www.zerotier.com/ */ #ifndef ZT_TOPOLOGY_HPP #define ZT_TOPOLOGY_HPP #include #include #include #include #include #include #include "Constants.hpp" #include "Address.hpp" #include "Identity.hpp" #include "Peer.hpp" #include "Mutex.hpp" #include "InetAddress.hpp" #include "Hashtable.hpp" #include "World.hpp" namespace ZeroTier { class RuntimeEnvironment; /** * Database of network topology */ class Topology { public: Topology(const RuntimeEnvironment *renv); ~Topology(); /** * Add a peer to database * * This will not replace existing peers. In that case the existing peer * record is returned. * * @param peer Peer to add * @return New or existing peer (should replace 'peer') */ SharedPtr addPeer(const SharedPtr &peer); /** * Get a peer from its address * * @param zta ZeroTier address of peer * @return Peer or NULL if not found */ SharedPtr getPeer(const Address &zta); /** * Get a peer only if it is presently in memory (no disk cache) * * This also does not update the lastUsed() time for peers, which means * that it won't prevent them from falling out of RAM. This is currently * used in the Cluster code to update peer info without forcing all peers * across the entire cluster to remain in memory cache. * * @param zta ZeroTier address */ inline SharedPtr getPeerNoCache(const Address &zta) { Mutex::Lock _l(_lock); const SharedPtr *const ap = _peers.get(zta); if (ap) return *ap; return SharedPtr(); } /** * Get the identity of a peer * * @param zta ZeroTier address of peer * @return Identity or NULL Identity if not found */ Identity getIdentity(const Address &zta); /** * Cache an identity * * This is done automatically on addPeer(), and so is only useful for * cluster identity replication. * * @param id Identity to cache */ void saveIdentity(const Identity &id); /** * Get the current favorite root server * * @return Root server with lowest latency or NULL if none */ inline SharedPtr getBestRoot() { return getBestRoot((const Address *)0,0,false); } /** * Get the best root server, avoiding root servers listed in an array * * This will get the best root server (lowest latency, etc.) but will * try to avoid the listed root servers, only using them if no others * are available. * * @param avoid Nodes to avoid * @param avoidCount Number of nodes to avoid * @param strictAvoid If false, consider avoided root servers anyway if no non-avoid root servers are available * @return Root server or NULL if none available */ SharedPtr getBestRoot(const Address *avoid,unsigned int avoidCount,bool strictAvoid); /** * @param id Identity to check * @return True if this is a designated root server in this world */ inline bool isRoot(const Identity &id) const { Mutex::Lock _l(_lock); return (std::find(_rootAddresses.begin(),_rootAddresses.end(),id.address()) != _rootAddresses.end()); } /** * @param id Identity to check * @return True if this is a root server or a network preferred relay from one of our networks */ bool isUpstream(const Identity &id) const; /** * @return Vector of root server addresses */ inline std::vector
rootAddresses() const { Mutex::Lock _l(_lock); return _rootAddresses; } /** * @return Current World (copy) */ inline World world() const { Mutex::Lock _l(_lock); return _world; } /** * @return Current world ID */ inline uint64_t worldId() const { return _world.id(); // safe to read without lock, and used from within eachPeer() so don't lock } /** * @return Current world timestamp */ inline uint64_t worldTimestamp() const { return _world.timestamp(); // safe to read without lock, and used from within eachPeer() so don't lock } /** * Validate new world and update if newer and signature is okay * * @param newWorld Potential new world definition revision * @return True if an update actually occurred */ bool worldUpdateIfValid(const World &newWorld); /** * Clean and flush database */ void clean(uint64_t now); /** * @param now Current time * @return Number of peers with active direct paths */ inline unsigned long countActive(uint64_t now) const { unsigned long cnt = 0; Mutex::Lock _l(_lock); Hashtable< Address,SharedPtr >::Iterator i(const_cast(this)->_peers); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { cnt += (unsigned long)((*p)->hasActiveDirectPath(now)); } return cnt; } /** * Apply a function or function object to all peers * * Note: explicitly template this by reference if you want the object * passed by reference instead of copied. * * Warning: be careful not to use features in these that call any other * methods of Topology that may lock _lock, otherwise a recursive lock * and deadlock or lock corruption may occur. * * @param f Function to apply * @tparam F Function or function object type */ template inline void eachPeer(F f) { Mutex::Lock _l(_lock); Hashtable< Address,SharedPtr >::Iterator i(_peers); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { #ifdef ZT_TRACE if (!(*p)) { fprintf(stderr,"FATAL BUG: eachPeer() caught NULL peer for %s -- peer pointers in Topology should NEVER be NULL" ZT_EOL_S,a->toString().c_str()); abort(); } #endif f(*this,*((const SharedPtr *)p)); } } /** * @return All currently active peers by address (unsorted) */ inline std::vector< std::pair< Address,SharedPtr > > allPeers() const { Mutex::Lock _l(_lock); return _peers.entries(); } /** * @return True if I am a root server in the current World */ inline bool amRoot() const throw() { return _amRoot; } private: Identity _getIdentity(const Address &zta); void _setWorld(const World &newWorld); const RuntimeEnvironment *const RR; World _world; Hashtable< Address,SharedPtr > _peers; std::vector< Address > _rootAddresses; std::vector< SharedPtr > _rootPeers; bool _amRoot; Mutex _lock; }; } // namespace ZeroTier #endif