/* * Copyright (c)2019 ZeroTier, Inc. * * Use of this software is governed by the Business Source License included * in the LICENSE.TXT file in the project's root directory. * * Change Date: 2023-01-01 * * On the date above, in accordance with the Business Source License, use * of this software will be governed by version 2.0 of the Apache License. */ /****/ #ifndef ZT_TOPOLOGY_HPP #define ZT_TOPOLOGY_HPP #include #include #include #include #include #include #include #include "Constants.hpp" #include "../include/ZeroTierOne.h" #include "Address.hpp" #include "Identity.hpp" #include "Peer.hpp" #include "Path.hpp" #include "Mutex.hpp" #include "InetAddress.hpp" #include "Hashtable.hpp" #include "Locator.hpp" #include "SharedPtr.hpp" #include "ScopedPtr.hpp" namespace ZeroTier { class RuntimeEnvironment; /** * Database of network topology */ class Topology { private: struct _RootRankingFunction { ZT_ALWAYS_INLINE _RootRankingFunction() : bestRoot(),bestRootLatency(0xffff) {} ZT_ALWAYS_INLINE bool operator()(const SharedPtr &peer,const std::vector &phy) { } SharedPtr bestRoot; unsigned int bestRootLatency; }; public: ZT_ALWAYS_INLINE Topology(const RuntimeEnvironment *renv,const Identity &myId) : RR(renv), _myIdentity(myId), _numConfiguredPhysicalPaths(0) {} ZT_ALWAYS_INLINE ~Topology() {} /** * Add a peer to database * * This will not replace existing peers. In that case the existing peer * record is returned. * * @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call * @param peer Peer to add * @return New or existing peer (should replace 'peer') */ ZT_ALWAYS_INLINE SharedPtr add(const SharedPtr &peer) { SharedPtr np; { Mutex::Lock _l(_peers_l); SharedPtr &hp = _peers[peer->address()]; if (!hp) hp = peer; np = hp; } return np; } /** * Get a peer from its address * * @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call * @param zta ZeroTier address of peer * @return Peer or NULL if not found */ ZT_ALWAYS_INLINE SharedPtr get(const Address &zta) { if (zta == _myIdentity.address()) return SharedPtr(); Mutex::Lock l1(_peers_l); const SharedPtr *const ap = _peers.get(zta); if (ap) return *ap; #if 0 Mutex::Lock l2(_roots_m); for(std::vector::const_iterator r(_roots.begin());r!=_roots.end();++r) { if (r->address() == zta) { try { SharedPtr rp(new Peer(RR,_myIdentity,r->id())); _peers[zta] = rp; return rp; } catch ( ... ) {} } } #endif return SharedPtr(); } /** * @param tPtr Thread pointer to be handed through to any callbacks called as a result of this call * @param zta ZeroTier address of peer * @return Identity or NULL identity if not found */ ZT_ALWAYS_INLINE Identity getIdentity(void *tPtr,const Address &zta) { if (zta == _myIdentity.address()) { return _myIdentity; } else { Mutex::Lock _l(_peers_l); const SharedPtr *const ap = _peers.get(zta); if (ap) return (*ap)->identity(); } return Identity(); } /** * Get a Path object for a given local and remote physical address, creating if needed * * @param l Local socket * @param r Remote address * @return Pointer to canonicalized Path object */ ZT_ALWAYS_INLINE SharedPtr getPath(const int64_t l,const InetAddress &r) { Mutex::Lock _l(_paths_l); SharedPtr &p = _paths[Path::HashKey(l,r)]; if (!p) p.set(new Path(l,r)); return p; } /** * @param id Identity to check * @return True if this identity corresponds to a root */ ZT_ALWAYS_INLINE bool isRoot(const Identity &id) const { #if 0 Mutex::Lock l(_roots_m); for(std::vector::const_iterator r(_roots.begin());r!=_roots.end();++r) { if (r->is(id)) return true; } #endif return false; } /** * Do periodic tasks such as database cleanup */ ZT_ALWAYS_INLINE void doPeriodicTasks(int64_t now) { { Mutex::Lock _l1(_peers_l); Hashtable< Address,SharedPtr >::Iterator i(_peers); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { if (!(*p)->alive(now)) { _peers.erase(*a); } } } { Mutex::Lock _l(_paths_l); Hashtable< Path::HashKey,SharedPtr >::Iterator i(_paths); Path::HashKey *k = (Path::HashKey *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(k,p)) { if (p->references() <= 1) _paths.erase(*k); } } } /** * @param now Current time * @return Number of peers with active direct paths */ ZT_ALWAYS_INLINE unsigned long countActive(int64_t now) const { unsigned long cnt = 0; Mutex::Lock _l(_peers_l); Hashtable< Address,SharedPtr >::Iterator i(const_cast(this)->_peers); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { const SharedPtr pp((*p)->getAppropriatePath(now,false)); if (pp) ++cnt; } return cnt; } /** * Apply a function or function object to all peers * * This locks the peer map during execution, so calls to get() etc. during * eachPeer() will deadlock. * * @param f Function to apply * @tparam F Function or function object type */ template ZT_ALWAYS_INLINE void eachPeer(F f) { Mutex::Lock l(_peers_l); Hashtable< Address,SharedPtr >::Iterator i(_peers); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { f(*((const SharedPtr *)p)); } } #if 0 /** * Apply a function or function object to all roots * * This locks the root list during execution but other operations * are fine. * * @param f Function to apply * @tparam F function or function object type */ template ZT_ALWAYS_INLINE void eachRoot(F f) { Mutex::Lock l(_roots_m); SharedPtr rp; for(std::vector::const_iterator i(_roots.begin());i!=_roots.end();++i) { { Mutex::Lock l2(_peers_l); const SharedPtr *const ap = _peers.get(i->address()); if (ap) { rp = *ap; } else { rp.set(new Peer(RR,_myIdentity,i->id())); _peers.set(rp->address(),rp); } } f(*i,rp); } } /** * Get the best root, rescanning and re-ranking roots periodically * * @param now Current time * @return Best/fastest currently connected root or NULL if none */ inline SharedPtr root(const int64_t now) { Mutex::Lock l(_bestRoot_m); if ((!_bestRoot)||((now - _lastRankedBestRoot) >= ZT_FIND_BEST_ROOT_PERIOD)) { _bestRoot.zero(); Mutex::Lock l2(_roots_m); SharedPtr rp; long bestQuality = 2147483647; for(std::vector::const_iterator i(_roots.begin());i!=_roots.end();++i) { { Mutex::Lock l2(_peers_l); const SharedPtr *const ap = _peers.get(i->address()); if (ap) { rp = *ap; } else { rp.set(new Peer(RR,_myIdentity,i->id())); _peers.set(rp->address(),rp); } } SharedPtr path(rp->getAppropriatePath(now,false)); if (path) { const long pq = path->quality(now); if (pq < bestQuality) { bestQuality = pq; _bestRoot = rp; } } } } return _bestRoot; } #endif /** * Apply a function or function object to all roots * * This locks the root list during execution but other operations * are fine. * * @param f Function to apply f(peer,IPs) * @tparam F function or function object type */ template inline void eachRoot(F f) { { Mutex::Lock l(_staticRoots_l); Hashtable< Identity,std::vector >::Iterator i(_staticRoots); Identity *k = (Identity *)0; std::vector *v = (std::vector *)0; while (i.next(k,v)) { if (!v->empty()) { const SharedPtr *ap; { Mutex::Lock l2(_peers_l); ap = _peers.get(k->address()); } if (ap) { if (!f(*ap,*v)) return; } else { SharedPtr p(new Peer(RR,_myIdentity,*k)); { Mutex::Lock l2(_peers_l); _peers.set(k->address(),p); } if (!f(p,*v)) return; } } } } { Mutex::Lock l(_dynamicRoots_l); Hashtable< Str,Locator >::Iterator i(_dynamicRoots); Str *k = (Str *)0; Locator *v = (Locator *)0; while (i.next(k,v)) { if (*v) { for(std::vector::const_iterator id(v->virt().begin());id!=v->virt().end();++id) { const SharedPtr *ap; { Mutex::Lock l2(_peers_l); ap = _peers.get(id->address()); } if (ap) { if (!f(*ap,v->phy())) return; } else { SharedPtr p(new Peer(RR,_myIdentity,*id)); { Mutex::Lock l2(_peers_l); _peers.set(id->address(),p); } if (!f(p,v->phy())) return; } } } } } } inline SharedPtr root(const int64_t now) { _RootRankingFunction rrf; eachRoot(rrf); } /** * @return Names of dynamic roots currently known by the system */ ZT_ALWAYS_INLINE std::vector dynamicRootNames() const { Mutex::Lock l(_dynamicRoots_l); return _dynamicRoots.keys(); } /** * Set or update a static root entry * * @param id Static root's identity * @param addrs Static root's IP address(es) */ ZT_ALWAYS_INLINE void setStaticRoot(const Identity &id,const std::vector &addrs) { Mutex::Lock l(_staticRoots_l); _staticRoots[id] = addrs; } /** * Set or update dynamic root if new locator is newer and valid * * This checks internal validity of the new locator including its internal self-signature. * It does not check any DNS signatures. * * @param dnsName DNS name used to retrive root * @param latestLocator Latest locator * @return True if latest locator is internally valid and newer */ ZT_ALWAYS_INLINE bool setDynamicRoot(const Str &dnsName,const Locator &latestLocator) { Mutex::Lock l(_dynamicRoots_l); Locator &ll = _dynamicRoots[dnsName]; if (ll.timestamp() < latestLocator.timestamp()) { ll = latestLocator; return true; } return false; } /** * Get the best relay to a given address, which may or may not be a root * * @param now Current time * @param toAddr Destination address * @return Best current relay or NULL if none */ ZT_ALWAYS_INLINE SharedPtr findRelayTo(const int64_t now,const Address &toAddr) { // TODO: in the future this will check 'mesh-like' relays and if enabled consult LF for other roots (for if this is a root) //return root(now); } /** * @param allPeers vector to fill with all current peers */ inline void getAllPeers(std::vector< SharedPtr > &allPeers) const { Mutex::Lock l(_peers_l); allPeers.clear(); allPeers.reserve(_peers.size()); Hashtable< Address,SharedPtr >::Iterator i(*(const_cast > *>(&_peers))); Address *a = (Address *)0; SharedPtr *p = (SharedPtr *)0; while (i.next(a,p)) { allPeers.push_back(*p); } } /** * Get info about a path * * The supplied result variables are not modified if no special config info is found. * * @param physicalAddress Physical endpoint address * @param mtu Variable set to MTU * @param trustedPathId Variable set to trusted path ID */ ZT_ALWAYS_INLINE void getOutboundPathInfo(const InetAddress &physicalAddress,unsigned int &mtu,uint64_t &trustedPathId) { for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i cpaths; for(unsigned int i=0,j=_numConfiguredPhysicalPaths;i ZT_MAX_PHYSMTU) pc.mtu = ZT_MAX_PHYSMTU; cpaths[*(reinterpret_cast(pathNetwork))] = pc; } else { cpaths.erase(*(reinterpret_cast(pathNetwork))); } unsigned int cnt = 0; for(std::map::const_iterator i(cpaths.begin());((i!=cpaths.end())&&(cntfirst; _physicalPathConfig[cnt].second = i->second; ++cnt; } _numConfiguredPhysicalPaths = cnt; } } private: const RuntimeEnvironment *const RR; const Identity _myIdentity; std::pair _physicalPathConfig[ZT_MAX_CONFIGURABLE_PATHS]; unsigned int _numConfiguredPhysicalPaths; Hashtable< Address,SharedPtr > _peers; Hashtable< Path::HashKey,SharedPtr > _paths; Hashtable< Identity,std::vector > _staticRoots; Hashtable< Str,Locator > _dynamicRoots; Mutex _peers_l; Mutex _paths_l; Mutex _staticRoots_l; Mutex _dynamicRoots_l; }; } // namespace ZeroTier #endif