|
|
|
|
@@ -8,6 +8,8 @@
|
|
|
|
|
#define _CRT_SECURE_NO_WARNINGS
|
|
|
|
|
#define _CRT_SECURE_NO_DEPRECATE
|
|
|
|
|
#define NOMINMAX
|
|
|
|
|
#include <chrono>
|
|
|
|
|
#include <thread>
|
|
|
|
|
#include <limits>
|
|
|
|
|
#include <utility>
|
|
|
|
|
#include "rabit/internal/io.h"
|
|
|
|
|
@@ -19,6 +21,7 @@
|
|
|
|
|
|
|
|
|
|
namespace rabit {
|
|
|
|
|
namespace engine {
|
|
|
|
|
|
|
|
|
|
AllreduceRobust::AllreduceRobust(void) {
|
|
|
|
|
num_local_replica = 0;
|
|
|
|
|
num_global_replica = 5;
|
|
|
|
|
@@ -38,7 +41,7 @@ bool AllreduceRobust::Init(int argc, char* argv[]) {
|
|
|
|
|
if (AllreduceBase::Init(argc, argv)) {
|
|
|
|
|
// chenqin: alert user opted in experimental feature.
|
|
|
|
|
if (rabit_bootstrap_cache) utils::HandleLogInfo(
|
|
|
|
|
"[EXPERIMENTAL] rabit bootstrap cache has been enabled\n");
|
|
|
|
|
"[EXPERIMENTAL] bootstrap cache has been enabled\n");
|
|
|
|
|
checkpoint_loaded = false;
|
|
|
|
|
if (num_global_replica == 0) {
|
|
|
|
|
result_buffer_round = -1;
|
|
|
|
|
@@ -55,24 +58,31 @@ bool AllreduceRobust::Shutdown(void) {
|
|
|
|
|
try {
|
|
|
|
|
// need to sync the exec before we shutdown, do a pesudo check point
|
|
|
|
|
// execute checkpoint, note: when checkpoint existing, load will not happen
|
|
|
|
|
utils::Assert(RecoverExec(NULL, 0, ActionSummary::kCheckPoint, ActionSummary::kSpecialOp,
|
|
|
|
|
_assert(RecoverExec(NULL, 0, ActionSummary::kCheckPoint, ActionSummary::kSpecialOp,
|
|
|
|
|
cur_cache_seq), "Shutdown: check point must return true");
|
|
|
|
|
// reset result buffer
|
|
|
|
|
resbuf.Clear(); seq_counter = 0;
|
|
|
|
|
cachebuf.Clear(); cur_cache_seq = 0;
|
|
|
|
|
lookupbuf.Clear();
|
|
|
|
|
// execute check ack step, load happens here
|
|
|
|
|
utils::Assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
_assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
ActionSummary::kSpecialOp, cur_cache_seq), "Shutdown: check ack must return true");
|
|
|
|
|
// travis ci only osx test hang
|
|
|
|
|
#if defined (__APPLE__)
|
|
|
|
|
sleep(1);
|
|
|
|
|
#endif
|
|
|
|
|
shutdown_timeout = true;
|
|
|
|
|
if (rabit_timeout_task.valid()) {
|
|
|
|
|
rabit_timeout_task.wait();
|
|
|
|
|
_assert(rabit_timeout_task.get(), "expect timeout task return\n");
|
|
|
|
|
}
|
|
|
|
|
return AllreduceBase::Shutdown();
|
|
|
|
|
} catch (const std::exception& e) {
|
|
|
|
|
fprintf(stderr, "%s\n", e.what());
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*!
|
|
|
|
|
* \brief set parameters to the engine
|
|
|
|
|
* \param name parameter name
|
|
|
|
|
@@ -98,8 +108,8 @@ int AllreduceRobust::SetBootstrapCache(const std::string &key, const void *buf,
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
utils::Assert(index == -1, "immutable cache key already exists");
|
|
|
|
|
utils::Assert(type_nbytes*count > 0, "can't set empty cache");
|
|
|
|
|
_assert(index == -1, "immutable cache key already exists");
|
|
|
|
|
_assert(type_nbytes*count > 0, "can't set empty cache");
|
|
|
|
|
void* temp = cachebuf.AllocTemp(type_nbytes, count);
|
|
|
|
|
cachebuf.PushTemp(cur_cache_seq, type_nbytes, count);
|
|
|
|
|
std::memcpy(temp, buf, type_nbytes*count);
|
|
|
|
|
@@ -133,9 +143,9 @@ int AllreduceRobust::GetBootstrapCache(const std::string &key, void* buf,
|
|
|
|
|
|
|
|
|
|
size_t siz = 0;
|
|
|
|
|
void* temp = cachebuf.Query(index, &siz);
|
|
|
|
|
utils::Assert(cur_cache_seq > index, "cur_cache_seq is smaller than lookup cache seq index");
|
|
|
|
|
utils::Assert(siz == type_nbytes*count, "cache size stored expected to be same as requested");
|
|
|
|
|
utils::Assert(siz > 0, "cache size should be greater than 0");
|
|
|
|
|
_assert(cur_cache_seq > index, "cur_cache_seq is smaller than lookup cache seq index");
|
|
|
|
|
_assert(siz == type_nbytes*count, "cache size stored expected to be same as requested");
|
|
|
|
|
_assert(siz > 0, "cache size should be greater than 0");
|
|
|
|
|
std::memcpy(buf, temp, type_nbytes*count);
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
@@ -317,7 +327,7 @@ int AllreduceRobust::LoadCheckPoint(Serializable *global_model,
|
|
|
|
|
local_rptr[local_chkpt_version][1]);
|
|
|
|
|
local_model->Load(&fs);
|
|
|
|
|
} else {
|
|
|
|
|
utils::Assert(nlocal == 0, "[%d] local model inconsistent, nlocal=%d", rank, nlocal);
|
|
|
|
|
_assert(nlocal == 0, "[%d] local model inconsistent, nlocal=%d", rank, nlocal);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
// reset result buffer
|
|
|
|
|
@@ -327,14 +337,14 @@ int AllreduceRobust::LoadCheckPoint(Serializable *global_model,
|
|
|
|
|
if (global_checkpoint.length() == 0) {
|
|
|
|
|
version_number = 0;
|
|
|
|
|
} else {
|
|
|
|
|
utils::Assert(fs.Read(&version_number, sizeof(version_number)) != 0,
|
|
|
|
|
_assert(fs.Read(&version_number, sizeof(version_number)) != 0,
|
|
|
|
|
"read in version number");
|
|
|
|
|
global_model->Load(&fs);
|
|
|
|
|
utils::Assert(local_model == NULL || nlocal == num_local_replica + 1,
|
|
|
|
|
_assert(local_model == NULL || nlocal == num_local_replica + 1,
|
|
|
|
|
"local model inconsistent, nlocal=%d", nlocal);
|
|
|
|
|
}
|
|
|
|
|
// run another phase of check ack, if recovered from data
|
|
|
|
|
utils::Assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
_assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
ActionSummary::kSpecialOp, cur_cache_seq), "check ack must return true");
|
|
|
|
|
|
|
|
|
|
if (!RecoverExec(NULL, 0, ActionSummary::kLoadBootstrapCache, seq_counter, cur_cache_seq)) {
|
|
|
|
|
@@ -433,7 +443,7 @@ void AllreduceRobust::CheckPoint_(const Serializable *global_model,
|
|
|
|
|
local_chkpt_version = !local_chkpt_version;
|
|
|
|
|
}
|
|
|
|
|
// execute checkpoint, note: when checkpoint existing, load will not happen
|
|
|
|
|
utils::Assert(RecoverExec(NULL, 0, ActionSummary::kCheckPoint,
|
|
|
|
|
_assert(RecoverExec(NULL, 0, ActionSummary::kCheckPoint,
|
|
|
|
|
ActionSummary::kSpecialOp, cur_cache_seq),
|
|
|
|
|
"check point must return true");
|
|
|
|
|
// this is the critical region where we will change all the stored models
|
|
|
|
|
@@ -460,7 +470,7 @@ void AllreduceRobust::CheckPoint_(const Serializable *global_model,
|
|
|
|
|
// reset result buffer, mark boostrap phase complete
|
|
|
|
|
resbuf.Clear(); seq_counter = 0;
|
|
|
|
|
// execute check ack step, load happens here
|
|
|
|
|
utils::Assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
_assert(RecoverExec(NULL, 0, ActionSummary::kCheckAck,
|
|
|
|
|
ActionSummary::kSpecialOp, cur_cache_seq), "check ack must return true");
|
|
|
|
|
|
|
|
|
|
delta = utils::GetTime() - start;
|
|
|
|
|
@@ -533,7 +543,7 @@ AllreduceRobust::ReturnType AllreduceRobust::TryResetLinks(void) {
|
|
|
|
|
if (all_links[i].size_read == 0) {
|
|
|
|
|
int atmark = all_links[i].sock.AtMark();
|
|
|
|
|
if (atmark < 0) {
|
|
|
|
|
utils::Assert(all_links[i].sock.BadSocket(), "must already gone bad");
|
|
|
|
|
_assert(all_links[i].sock.BadSocket(), "must already gone bad");
|
|
|
|
|
} else if (atmark > 0) {
|
|
|
|
|
all_links[i].size_read = 1;
|
|
|
|
|
} else {
|
|
|
|
|
@@ -555,10 +565,10 @@ AllreduceRobust::ReturnType AllreduceRobust::TryResetLinks(void) {
|
|
|
|
|
if (len == 0) {
|
|
|
|
|
all_links[i].sock.Close(); continue;
|
|
|
|
|
} else if (len > 0) {
|
|
|
|
|
utils::Assert(oob_mark == kResetMark, "wrong oob msg");
|
|
|
|
|
utils::Assert(all_links[i].sock.AtMark() != 1, "should already read past mark");
|
|
|
|
|
_assert(oob_mark == kResetMark, "wrong oob msg");
|
|
|
|
|
_assert(all_links[i].sock.AtMark() != 1, "should already read past mark");
|
|
|
|
|
} else {
|
|
|
|
|
utils::Assert(errno != EAGAIN|| errno != EWOULDBLOCK, "BUG");
|
|
|
|
|
_assert(errno != EAGAIN|| errno != EWOULDBLOCK, "BUG");
|
|
|
|
|
}
|
|
|
|
|
// send out ack
|
|
|
|
|
char ack = kResetAck;
|
|
|
|
|
@@ -579,9 +589,9 @@ AllreduceRobust::ReturnType AllreduceRobust::TryResetLinks(void) {
|
|
|
|
|
if (len == 0) {
|
|
|
|
|
all_links[i].sock.Close(); continue;
|
|
|
|
|
} else if (len > 0) {
|
|
|
|
|
utils::Assert(ack == kResetAck, "wrong Ack MSG");
|
|
|
|
|
_assert(ack == kResetAck, "wrong Ack MSG");
|
|
|
|
|
} else {
|
|
|
|
|
utils::Assert(errno != EAGAIN|| errno != EWOULDBLOCK, "BUG");
|
|
|
|
|
_assert(errno != EAGAIN|| errno != EWOULDBLOCK, "BUG");
|
|
|
|
|
}
|
|
|
|
|
// set back to nonblock mode
|
|
|
|
|
all_links[i].sock.SetNonBlock(true);
|
|
|
|
|
@@ -600,14 +610,44 @@ AllreduceRobust::ReturnType AllreduceRobust::TryResetLinks(void) {
|
|
|
|
|
* \return true if err_type is kSuccess, false otherwise
|
|
|
|
|
*/
|
|
|
|
|
bool AllreduceRobust::CheckAndRecover(ReturnType err_type) {
|
|
|
|
|
shutdown_timeout = err_type == kSuccess;
|
|
|
|
|
if (err_type == kSuccess) return true;
|
|
|
|
|
utils::Assert(err_link != NULL, "must know the error source");
|
|
|
|
|
recover_counter += 1;
|
|
|
|
|
|
|
|
|
|
_assert(err_link != NULL, "must know the error link");
|
|
|
|
|
recover_counter += 1;
|
|
|
|
|
// async launch timeout task if enable_rabit_timeout is set
|
|
|
|
|
if (rabit_timeout && !rabit_timeout_task.valid()) {
|
|
|
|
|
utils::Printf("[EXPERIMENTAL] timeout thread expires in %d second(s)\n", timeout_sec);
|
|
|
|
|
rabit_timeout_task = std::async(std::launch::async, [=]() {
|
|
|
|
|
if (rabit_debug) {
|
|
|
|
|
utils::Printf("[%d] timeout thread %ld starts\n", rank,
|
|
|
|
|
std::this_thread::get_id());
|
|
|
|
|
}
|
|
|
|
|
int time = 0;
|
|
|
|
|
// check if rabit recovered every 100ms
|
|
|
|
|
while (time++ < 10 * timeout_sec) {
|
|
|
|
|
std::this_thread::sleep_for(std::chrono::milliseconds(100));
|
|
|
|
|
if (shutdown_timeout.load()) {
|
|
|
|
|
if (rabit_debug) {
|
|
|
|
|
utils::Printf("[%d] timeout task thread %ld exits\n",
|
|
|
|
|
rank, std::this_thread::get_id());
|
|
|
|
|
}
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
// print on tracker to help debuging
|
|
|
|
|
TrackerPrint("[ERROR] rank " + std::to_string(rank) + "@"+
|
|
|
|
|
host_uri + ":" +std::to_string(port) + " timeout\n");
|
|
|
|
|
_error("[%d] exit due to time out %d s\n", rank, timeout_sec);
|
|
|
|
|
return false;
|
|
|
|
|
});
|
|
|
|
|
}
|
|
|
|
|
// simple way, shutdown all links
|
|
|
|
|
for (size_t i = 0; i < all_links.size(); ++i) {
|
|
|
|
|
if (!all_links[i].sock.BadSocket()) all_links[i].sock.Close();
|
|
|
|
|
}
|
|
|
|
|
// smooth out traffic to tracker
|
|
|
|
|
std::this_thread::sleep_for(std::chrono::milliseconds(10*rank));
|
|
|
|
|
ReConnectLinks("recover");
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
@@ -724,8 +764,8 @@ AllreduceRobust::TryDecideRouting(AllreduceRobust::RecoverType role,
|
|
|
|
|
// set p_req_in
|
|
|
|
|
(*p_req_in)[i] = (req_in[i] != 0);
|
|
|
|
|
if (req_out[i] != 0) {
|
|
|
|
|
utils::Assert(req_in[i] == 0, "cannot get and receive request");
|
|
|
|
|
utils::Assert(static_cast<int>(i) == best_link, "request result inconsistent");
|
|
|
|
|
_assert(req_in[i] == 0, "cannot get and receive request");
|
|
|
|
|
_assert(static_cast<int>(i) == best_link, "request result inconsistent");
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
*p_recvlink = best_link;
|
|
|
|
|
@@ -755,20 +795,20 @@ AllreduceRobust::TryRecoverData(RecoverType role,
|
|
|
|
|
RefLinkVector &links = tree_links;
|
|
|
|
|
// no need to run recovery for zero size messages
|
|
|
|
|
if (links.size() == 0 || size == 0) return kSuccess;
|
|
|
|
|
utils::Assert(req_in.size() == links.size(), "TryRecoverData");
|
|
|
|
|
_assert(req_in.size() == links.size(), "TryRecoverData");
|
|
|
|
|
const int nlink = static_cast<int>(links.size());
|
|
|
|
|
{
|
|
|
|
|
bool req_data = role == kRequestData;
|
|
|
|
|
for (int i = 0; i < nlink; ++i) {
|
|
|
|
|
if (req_in[i]) {
|
|
|
|
|
utils::Assert(i != recv_link, "TryDecideRouting");
|
|
|
|
|
_assert(i != recv_link, "TryDecideRouting");
|
|
|
|
|
req_data = true;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
// do not need to provide data or receive data, directly exit
|
|
|
|
|
if (!req_data) return kSuccess;
|
|
|
|
|
}
|
|
|
|
|
utils::Assert(recv_link >= 0 || role == kHaveData, "recv_link must be active");
|
|
|
|
|
_assert(recv_link >= 0 || role == kHaveData, "recv_link must be active");
|
|
|
|
|
if (role == kPassData) {
|
|
|
|
|
links[recv_link].InitBuffer(1, size, reduce_buffer_size);
|
|
|
|
|
}
|
|
|
|
|
@@ -835,7 +875,7 @@ AllreduceRobust::TryRecoverData(RecoverType role,
|
|
|
|
|
for (int i = 0; i < nlink; ++i) {
|
|
|
|
|
if (req_in[i]) min_write = std::min(links[i].size_write, min_write);
|
|
|
|
|
}
|
|
|
|
|
utils::Assert(min_write <= links[pid].size_read, "boundary check");
|
|
|
|
|
_assert(min_write <= links[pid].size_read, "boundary check");
|
|
|
|
|
ReturnType ret = links[pid].ReadToRingBuffer(min_write, size);
|
|
|
|
|
if (ret != kSuccess) {
|
|
|
|
|
return ReportError(&links[pid], ret);
|
|
|
|
|
@@ -869,7 +909,7 @@ AllreduceRobust::ReturnType AllreduceRobust::TryRestoreCache(bool requester,
|
|
|
|
|
const int min_seq, const int max_seq) {
|
|
|
|
|
// clear requester and rebuild from those with most cache entries
|
|
|
|
|
if (requester) {
|
|
|
|
|
utils::Assert(cur_cache_seq <= max_seq, "requester is expected to have fewer cache entries");
|
|
|
|
|
_assert(cur_cache_seq <= max_seq, "requester is expected to have fewer cache entries");
|
|
|
|
|
cachebuf.Clear();
|
|
|
|
|
lookupbuf.Clear();
|
|
|
|
|
cur_cache_seq = 0;
|
|
|
|
|
@@ -998,7 +1038,7 @@ AllreduceRobust::TryGetResult(void *sendrecvbuf, size_t size, int seqno, bool re
|
|
|
|
|
int new_version = !local_chkpt_version;
|
|
|
|
|
int nlocal = std::max(static_cast<int>(local_rptr[new_version].size()) - 1, 0);
|
|
|
|
|
// if we goes to this place, use must have already setup the state once
|
|
|
|
|
utils::Assert(nlocal == 1 || nlocal == num_local_replica + 1,
|
|
|
|
|
_assert(nlocal == 1 || nlocal == num_local_replica + 1,
|
|
|
|
|
"TryGetResult::Checkpoint");
|
|
|
|
|
return TryRecoverLocalState(&local_rptr[new_version], &local_chkpt[new_version]);
|
|
|
|
|
}
|
|
|
|
|
@@ -1048,13 +1088,13 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
// kLoadBootstrapCache should be treated similar as allreduce
|
|
|
|
|
// when loadcheck/check/checkack runs in other nodes
|
|
|
|
|
if (flag != 0 && flag != ActionSummary::kLoadBootstrapCache) {
|
|
|
|
|
utils::Assert(seqno == ActionSummary::kSpecialOp, "must only set seqno for normal operations");
|
|
|
|
|
_assert(seqno == ActionSummary::kSpecialOp, "must only set seqno for normal operations");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
std::string msg = std::string(caller) + " pass negative seqno "
|
|
|
|
|
+ std::to_string(seqno) + " flag " + std::to_string(flag)
|
|
|
|
|
+ " version " + std::to_string(version_number);
|
|
|
|
|
utils::Assert(seqno >=0, msg.c_str());
|
|
|
|
|
_assert(seqno >=0, msg.c_str());
|
|
|
|
|
|
|
|
|
|
ActionSummary req(flag, flag, seqno, cache_seqno);
|
|
|
|
|
|
|
|
|
|
@@ -1068,7 +1108,7 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
if (act.check_ack()) {
|
|
|
|
|
if (act.check_point()) {
|
|
|
|
|
// if we also have check_point, do check point first
|
|
|
|
|
utils::Assert(!act.diff_seq(),
|
|
|
|
|
_assert(!act.diff_seq(),
|
|
|
|
|
"check ack & check pt cannot occur together with normal ops");
|
|
|
|
|
// if we requested checkpoint, we are free to go
|
|
|
|
|
if (req.check_point()) return true;
|
|
|
|
|
@@ -1087,7 +1127,7 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
} else {
|
|
|
|
|
if (act.check_point()) {
|
|
|
|
|
if (act.diff_seq()) {
|
|
|
|
|
utils::Assert(act.seqno() != ActionSummary::kSpecialOp, "min seq bug");
|
|
|
|
|
_assert(act.seqno() != ActionSummary::kSpecialOp, "min seq bug");
|
|
|
|
|
// print checkpoint consensus flag if user turn on debug
|
|
|
|
|
if (rabit_debug) {
|
|
|
|
|
req.print_flags(rank, "checkpoint req");
|
|
|
|
|
@@ -1112,16 +1152,16 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
if (!act.load_cache()) {
|
|
|
|
|
if (act.seqno() > 0) {
|
|
|
|
|
if (!requester) {
|
|
|
|
|
utils::Assert(req.check_point(), "checkpoint node should be KHaveData role");
|
|
|
|
|
_assert(req.check_point(), "checkpoint node should be KHaveData role");
|
|
|
|
|
buf = resbuf.Query(act.seqno(), &size);
|
|
|
|
|
utils::Assert(buf != NULL, "buf should have data from resbuf");
|
|
|
|
|
utils::Assert(size > 0, "buf size should be greater than 0");
|
|
|
|
|
_assert(buf != NULL, "buf should have data from resbuf");
|
|
|
|
|
_assert(size > 0, "buf size should be greater than 0");
|
|
|
|
|
}
|
|
|
|
|
if (!CheckAndRecover(TryGetResult(buf, size, act.seqno(), requester))) continue;
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
// cache seq no should be smaller than kSpecialOp
|
|
|
|
|
utils::Assert(act.seqno(SeqType::kCache) != ActionSummary::kSpecialOp,
|
|
|
|
|
_assert(act.seqno(SeqType::kCache) != ActionSummary::kSpecialOp,
|
|
|
|
|
"checkpoint with kSpecialOp");
|
|
|
|
|
int max_cache_seq = cur_cache_seq;
|
|
|
|
|
if (TryAllreduce(&max_cache_seq, sizeof(max_cache_seq), 1,
|
|
|
|
|
@@ -1153,11 +1193,11 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
act.print_flags(rank, "loadcache act");
|
|
|
|
|
}
|
|
|
|
|
// load cache should not running in parralel with other states
|
|
|
|
|
utils::Assert(!act.load_check(),
|
|
|
|
|
_assert(!act.load_check(),
|
|
|
|
|
"load cache state expect no nodes doing load checkpoint");
|
|
|
|
|
utils::Assert(!act.check_point() ,
|
|
|
|
|
_assert(!act.check_point() ,
|
|
|
|
|
"load cache state expect no nodes doing checkpoint");
|
|
|
|
|
utils::Assert(!act.check_ack(),
|
|
|
|
|
_assert(!act.check_ack(),
|
|
|
|
|
"load cache state expect no nodes doing checkpoint ack");
|
|
|
|
|
|
|
|
|
|
// if all nodes are requester in load cache, skip
|
|
|
|
|
@@ -1176,10 +1216,10 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// assert no req with load cache set goes into seq catch up
|
|
|
|
|
utils::Assert(!req.load_cache(), "load cache not interacte with rest states");
|
|
|
|
|
_assert(!req.load_cache(), "load cache not interacte with rest states");
|
|
|
|
|
|
|
|
|
|
// no special flags, no checkpoint, check ack, load_check
|
|
|
|
|
utils::Assert(act.seqno() != ActionSummary::kSpecialOp, "min seq bug");
|
|
|
|
|
_assert(act.seqno() != ActionSummary::kSpecialOp, "min seq bug");
|
|
|
|
|
if (act.diff_seq()) {
|
|
|
|
|
bool requester = req.seqno() == act.seqno();
|
|
|
|
|
if (!CheckAndRecover(TryGetResult(buf, size, act.seqno(), requester))) continue;
|
|
|
|
|
@@ -1194,7 +1234,7 @@ bool AllreduceRobust::RecoverExec(void *buf, size_t size, int flag, int seqno,
|
|
|
|
|
// something is still incomplete try next round
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
utils::Assert(false, "RecoverExec: should not reach here");
|
|
|
|
|
_assert(false, "RecoverExec: should not reach here");
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
/*!
|
|
|
|
|
@@ -1222,13 +1262,13 @@ AllreduceRobust::TryRecoverLocalState(std::vector<size_t> *p_local_rptr,
|
|
|
|
|
std::string &chkpt = *p_local_chkpt;
|
|
|
|
|
if (rptr.size() == 0) {
|
|
|
|
|
rptr.push_back(0);
|
|
|
|
|
utils::Assert(chkpt.length() == 0, "local chkpt space inconsistent");
|
|
|
|
|
_assert(chkpt.length() == 0, "local chkpt space inconsistent");
|
|
|
|
|
}
|
|
|
|
|
const int n = num_local_replica;
|
|
|
|
|
{
|
|
|
|
|
// backward passing, passing state in backward direction of the ring
|
|
|
|
|
const int nlocal = static_cast<int>(rptr.size() - 1);
|
|
|
|
|
utils::Assert(nlocal <= n + 1, "invalid local replica");
|
|
|
|
|
_assert(nlocal <= n + 1, "invalid local replica");
|
|
|
|
|
std::vector<int> msg_back(n + 1);
|
|
|
|
|
msg_back[0] = nlocal;
|
|
|
|
|
// backward passing one hop the request
|
|
|
|
|
@@ -1282,7 +1322,7 @@ AllreduceRobust::TryRecoverLocalState(std::vector<size_t> *p_local_rptr,
|
|
|
|
|
{
|
|
|
|
|
// forward passing, passing state in forward direction of the ring
|
|
|
|
|
const int nlocal = static_cast<int>(rptr.size() - 1);
|
|
|
|
|
utils::Assert(nlocal <= n + 1, "invalid local replica");
|
|
|
|
|
_assert(nlocal <= n + 1, "invalid local replica");
|
|
|
|
|
std::vector<int> msg_forward(n + 1);
|
|
|
|
|
msg_forward[0] = nlocal;
|
|
|
|
|
// backward passing one hop the request
|
|
|
|
|
@@ -1367,7 +1407,7 @@ AllreduceRobust::TryCheckinLocalState(std::vector<size_t> *p_local_rptr,
|
|
|
|
|
if (num_local_replica == 0) return kSuccess;
|
|
|
|
|
std::vector<size_t> &rptr = *p_local_rptr;
|
|
|
|
|
std::string &chkpt = *p_local_chkpt;
|
|
|
|
|
utils::Assert(rptr.size() == 2,
|
|
|
|
|
_assert(rptr.size() == 2,
|
|
|
|
|
"TryCheckinLocalState must have exactly 1 state");
|
|
|
|
|
const int n = num_local_replica;
|
|
|
|
|
std::vector<size_t> sizes(n + 1);
|
|
|
|
|
@@ -1423,10 +1463,10 @@ AllreduceRobust::RingPassing(void *sendrecvbuf_,
|
|
|
|
|
LinkRecord *read_link,
|
|
|
|
|
LinkRecord *write_link) {
|
|
|
|
|
if (read_link == NULL || write_link == NULL || read_end == 0) return kSuccess;
|
|
|
|
|
utils::Assert(write_end <= read_end,
|
|
|
|
|
_assert(write_end <= read_end,
|
|
|
|
|
"RingPassing: boundary check1");
|
|
|
|
|
utils::Assert(read_ptr <= read_end, "RingPassing: boundary check2");
|
|
|
|
|
utils::Assert(write_ptr <= write_end, "RingPassing: boundary check3");
|
|
|
|
|
_assert(read_ptr <= read_end, "RingPassing: boundary check2");
|
|
|
|
|
_assert(write_ptr <= write_end, "RingPassing: boundary check3");
|
|
|
|
|
// take reference
|
|
|
|
|
LinkRecord &prev = *read_link, &next = *write_link;
|
|
|
|
|
// send recv buffer
|
|
|
|
|
|