mirror of
https://github.com/zerotier/ZeroTierOne.git
synced 2024-12-24 07:06:39 +00:00
Merge branch 'dev' into jh-zerotier-multithreaded
This commit is contained in:
commit
6bc785e771
@ -373,6 +373,7 @@ SharedPtr<Path> Bond::getAppropriatePath(int64_t now, int32_t flowId)
|
|||||||
*/
|
*/
|
||||||
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
||||||
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS && _paths[_abPathIdx].p) {
|
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS && _paths[_abPathIdx].p) {
|
||||||
|
//fprintf(stderr, "trying to send via (_abPathIdx=%d) %s\n", _abPathIdx, pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
return _paths[_abPathIdx].p;
|
return _paths[_abPathIdx].p;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1032,6 +1033,13 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
|||||||
bool satisfiedUpDelay = (now - _paths[i].lastAliveToggle) >= _upDelay;
|
bool satisfiedUpDelay = (now - _paths[i].lastAliveToggle) >= _upDelay;
|
||||||
// How long since the last QoS was received (Must be less than ZT_PEER_PATH_EXPIRATION since the remote peer's _qosSendInterval isn't known)
|
// How long since the last QoS was received (Must be less than ZT_PEER_PATH_EXPIRATION since the remote peer's _qosSendInterval isn't known)
|
||||||
bool acceptableQoSAge = (_paths[i].lastQoSReceived == 0 && inTrial) || ((now - _paths[i].lastQoSReceived) < ZT_PEER_EXPIRED_PATH_TRIAL_PERIOD);
|
bool acceptableQoSAge = (_paths[i].lastQoSReceived == 0 && inTrial) || ((now - _paths[i].lastQoSReceived) < ZT_PEER_EXPIRED_PATH_TRIAL_PERIOD);
|
||||||
|
|
||||||
|
// Allow active-backup to operate without the receipt of QoS records
|
||||||
|
// This may be expanded to the other modes as an option
|
||||||
|
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
||||||
|
acceptableQoSAge = true;
|
||||||
|
}
|
||||||
|
|
||||||
currEligibility = _paths[i].allowed() && ((acceptableAge && satisfiedUpDelay && acceptableQoSAge) || inTrial);
|
currEligibility = _paths[i].allowed() && ((acceptableAge && satisfiedUpDelay && acceptableQoSAge) || inTrial);
|
||||||
|
|
||||||
if (currEligibility) {
|
if (currEligibility) {
|
||||||
@ -1043,12 +1051,11 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
|||||||
*/
|
*/
|
||||||
if (currEligibility != _paths[i].eligible) {
|
if (currEligibility != _paths[i].eligible) {
|
||||||
if (currEligibility == 0) {
|
if (currEligibility == 0) {
|
||||||
log("link %s is no longer eligible", pathToStr(_paths[i].p).c_str());
|
log("link %s is no longer eligible (reason: allowed=%d, age=%d, ud=%d, qos=%d, trial=%d)", pathToStr(_paths[i].p).c_str(), _paths[i].allowed(), acceptableAge, satisfiedUpDelay, acceptableQoSAge, inTrial);
|
||||||
}
|
}
|
||||||
if (currEligibility == 1) {
|
if (currEligibility == 1) {
|
||||||
log("link %s is eligible", pathToStr(_paths[i].p).c_str());
|
log("link %s is eligible", pathToStr(_paths[i].p).c_str());
|
||||||
}
|
}
|
||||||
debug("\t[%d] allowed=%d, age=%d, qa=%d, ud=%d, trial=%d", i, _paths[i].allowed(), acceptableAge, acceptableQoSAge, satisfiedUpDelay, inTrial);
|
|
||||||
dumpPathStatus(now, i);
|
dumpPathStatus(now, i);
|
||||||
if (currEligibility) {
|
if (currEligibility) {
|
||||||
rebuildBond = true;
|
rebuildBond = true;
|
||||||
@ -1496,7 +1503,8 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
{
|
{
|
||||||
int prevActiveBackupPathIdx = _abPathIdx;
|
int prevActiveBackupPathIdx = _abPathIdx;
|
||||||
int nonPreferredPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
int nonPreferredPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
bool bFoundPrimaryLink = false;
|
bool foundPathOnPrimaryLink = false;
|
||||||
|
bool foundPreferredPath = false;
|
||||||
|
|
||||||
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS && ! _paths[_abPathIdx].p) {
|
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS && ! _paths[_abPathIdx].p) {
|
||||||
_abPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
_abPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
@ -1559,15 +1567,16 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
if (! _paths[i].preferred()) {
|
if (! _paths[i].preferred()) {
|
||||||
// Found path on primary link, take note in case we don't find a preferred path
|
// Found path on primary link, take note in case we don't find a preferred path
|
||||||
nonPreferredPathIdx = i;
|
nonPreferredPathIdx = i;
|
||||||
bFoundPrimaryLink = true;
|
foundPathOnPrimaryLink = true;
|
||||||
}
|
}
|
||||||
if (_paths[i].preferred()) {
|
if (_paths[i].preferred()) {
|
||||||
_abPathIdx = i;
|
_abPathIdx = i;
|
||||||
bFoundPrimaryLink = true;
|
foundPathOnPrimaryLink = true;
|
||||||
if (_paths[_abPathIdx].p) {
|
if (_paths[_abPathIdx].p) {
|
||||||
SharedPtr<Link> abLink = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
SharedPtr<Link> abLink = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
||||||
if (abLink) {
|
if (abLink) {
|
||||||
log("found preferred primary link %s", pathToStr(_paths[_abPathIdx].p).c_str());
|
log("found preferred primary link (_abPathIdx=%d), %s", _abPathIdx, pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
|
foundPreferredPath = true;
|
||||||
}
|
}
|
||||||
break; // Found preferred path on primary link
|
break; // Found preferred path on primary link
|
||||||
}
|
}
|
||||||
@ -1575,8 +1584,8 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (bFoundPrimaryLink && (nonPreferredPathIdx != ZT_MAX_PEER_NETWORK_PATHS)) {
|
if (!foundPreferredPath && foundPathOnPrimaryLink && (nonPreferredPathIdx != ZT_MAX_PEER_NETWORK_PATHS)) {
|
||||||
log("found non-preferred primary link");
|
log("found non-preferred primary link (_abPathIdx=%d)", _abPathIdx);
|
||||||
_abPathIdx = nonPreferredPathIdx;
|
_abPathIdx = nonPreferredPathIdx;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1614,10 +1623,10 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
}
|
}
|
||||||
if (_paths[(*it)].p && ! _paths[(*it)].eligible) {
|
if (_paths[(*it)].p && ! _paths[(*it)].eligible) {
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[(*it)].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[(*it)].p->localSocket());
|
||||||
it = _abFailoverQueue.erase(it);
|
|
||||||
if (link) {
|
if (link) {
|
||||||
log("link %s is ineligible, removing from failover queue (%zu links remain in queue)", pathToStr(_paths[_abPathIdx].p).c_str(), _abFailoverQueue.size());
|
log("link %s is ineligible, removing from failover queue (%zu links remain in queue)", pathToStr(_paths[(*it)].p).c_str(), _abFailoverQueue.size());
|
||||||
}
|
}
|
||||||
|
it = _abFailoverQueue.erase(it);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
@ -1684,7 +1693,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (! bFoundPathInQueue) {
|
if (! bFoundPathInQueue) {
|
||||||
_abFailoverQueue.push_front(i);
|
_abFailoverQueue.push_back(i);
|
||||||
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
||||||
addPathToBond(i, 0);
|
addPathToBond(i, 0);
|
||||||
}
|
}
|
||||||
@ -1734,13 +1743,14 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (! bFoundPathInQueue) {
|
if (! bFoundPathInQueue) {
|
||||||
_abFailoverQueue.push_front(i);
|
_abFailoverQueue.push_back(i);
|
||||||
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
||||||
addPathToBond(i, 0);
|
addPathToBond(i, 0);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
/*
|
||||||
// Sort queue based on performance
|
// Sort queue based on performance
|
||||||
if (! _abFailoverQueue.empty()) {
|
if (! _abFailoverQueue.empty()) {
|
||||||
for (int i = 0; i < _abFailoverQueue.size(); i++) {
|
for (int i = 0; i < _abFailoverQueue.size(); i++) {
|
||||||
@ -1752,7 +1762,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
|||||||
}
|
}
|
||||||
_abFailoverQueue[hole_position] = value_to_insert;
|
_abFailoverQueue[hole_position] = value_to_insert;
|
||||||
}
|
}
|
||||||
}
|
}*/
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Short-circuit if we have no queued paths
|
* Short-circuit if we have no queued paths
|
||||||
@ -1902,7 +1912,7 @@ void Bond::setBondParameters(int policy, SharedPtr<Bond> templateBond, bool useT
|
|||||||
* Policy defaults
|
* Policy defaults
|
||||||
*/
|
*/
|
||||||
_abPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
_abPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
_abLinkSelectMethod = ZT_BOND_RESELECTION_POLICY_OPTIMIZE;
|
_abLinkSelectMethod = ZT_BOND_RESELECTION_POLICY_ALWAYS;
|
||||||
_rrPacketsSentOnCurrLink = 0;
|
_rrPacketsSentOnCurrLink = 0;
|
||||||
_rrIdx = 0;
|
_rrIdx = 0;
|
||||||
_packetsPerLink = 64;
|
_packetsPerLink = 64;
|
||||||
@ -2021,7 +2031,8 @@ void Bond::dumpInfo(int64_t now, bool force)
|
|||||||
_lastSummaryDump = now;
|
_lastSummaryDump = now;
|
||||||
float overhead = (_overheadBytes / (timeSinceLastDump / 1000.0f) / 1000.0f);
|
float overhead = (_overheadBytes / (timeSinceLastDump / 1000.0f) / 1000.0f);
|
||||||
_overheadBytes = 0;
|
_overheadBytes = 0;
|
||||||
log("bond: bp=%d, fi=%" PRIu64 ", mi=%d, ud=%d, dd=%d, flows=%zu, leaf=%d, overhead=%f KB/s, links=(%d/%d)",
|
log("bond: ready=%d, bp=%d, fi=%" PRIu64 ", mi=%d, ud=%d, dd=%d, flows=%zu, leaf=%d, overhead=%f KB/s, links=(%d/%d)",
|
||||||
|
isReady(),
|
||||||
_policy,
|
_policy,
|
||||||
_failoverInterval,
|
_failoverInterval,
|
||||||
_monitorInterval,
|
_monitorInterval,
|
||||||
|
@ -1144,6 +1144,7 @@ class Bond {
|
|||||||
__attribute__((format(printf, 2, 3)))
|
__attribute__((format(printf, 2, 3)))
|
||||||
#endif
|
#endif
|
||||||
{
|
{
|
||||||
|
//if (_peerId != 0x0 && _peerId != 0x0) { return; }
|
||||||
#ifdef ZT_TRACE
|
#ifdef ZT_TRACE
|
||||||
time_t rawtime;
|
time_t rawtime;
|
||||||
struct tm* timeinfo;
|
struct tm* timeinfo;
|
||||||
@ -1175,6 +1176,7 @@ class Bond {
|
|||||||
__attribute__((format(printf, 2, 3)))
|
__attribute__((format(printf, 2, 3)))
|
||||||
#endif
|
#endif
|
||||||
{
|
{
|
||||||
|
//if (_peerId != 0x0 && _peerId != 0x0) { return; }
|
||||||
#ifdef ZT_DEBUG
|
#ifdef ZT_DEBUG
|
||||||
time_t rawtime;
|
time_t rawtime;
|
||||||
struct tm* timeinfo;
|
struct tm* timeinfo;
|
||||||
|
@ -335,7 +335,6 @@ bool IncomingPacket::_doACK(const RuntimeEnvironment* RR, void* tPtr, const Shar
|
|||||||
bool IncomingPacket::_doQOS_MEASUREMENT(const RuntimeEnvironment* RR, void* tPtr, const SharedPtr<Peer>& peer)
|
bool IncomingPacket::_doQOS_MEASUREMENT(const RuntimeEnvironment* RR, void* tPtr, const SharedPtr<Peer>& peer)
|
||||||
{
|
{
|
||||||
Metrics::pkt_qos_in++;
|
Metrics::pkt_qos_in++;
|
||||||
SharedPtr<Bond> bond = peer->bond();
|
|
||||||
if (! peer->rateGateQoS(RR->node->now(), _path)) {
|
if (! peer->rateGateQoS(RR->node->now(), _path)) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
@ -2547,7 +2547,7 @@ public:
|
|||||||
}
|
}
|
||||||
_node->bondController()->addCustomLink(customPolicyStr, new Link(linkNameStr,ipvPref,mtu,capacity,enabled,linkMode,failoverToStr));
|
_node->bondController()->addCustomLink(customPolicyStr, new Link(linkNameStr,ipvPref,mtu,capacity,enabled,linkMode,failoverToStr));
|
||||||
}
|
}
|
||||||
std::string linkSelectMethodStr(OSUtils::jsonString(customPolicy["activeReselect"],"optimize"));
|
std::string linkSelectMethodStr(OSUtils::jsonString(customPolicy["activeReselect"],"always"));
|
||||||
if (linkSelectMethodStr == "always") {
|
if (linkSelectMethodStr == "always") {
|
||||||
newTemplateBond->setLinkSelectMethod(ZT_BOND_RESELECTION_POLICY_ALWAYS);
|
newTemplateBond->setLinkSelectMethod(ZT_BOND_RESELECTION_POLICY_ALWAYS);
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user