Avoid rabit calls in learner configuration (#5581)

This commit is contained in:
Rory Mitchell
2020-04-24 14:59:29 +12:00
committed by GitHub
parent 92913aaf7f
commit 660be66207
3 changed files with 26 additions and 20 deletions

View File

@@ -20,8 +20,9 @@ num_round = 20
bst = xgb.train(param, dtrain, num_round, watchlist, early_stopping_rounds=2)
# Save the model, only ask process 0 to save the model.
bst.save_model("test.model{}".format(xgb.rabit.get_rank()))
xgb.rabit.tracker_print("Finished training\n")
if xgb.rabit.get_rank() == 0:
bst.save_model("test.model")
xgb.rabit.tracker_print("Finished training\n")
# Notify the tracker all training has been successful
# This is only needed in distributed training.

View File

@@ -70,8 +70,9 @@ watchlist = [(dtrain,'train')]
num_round = 2
bst = xgb.train(param, dtrain, num_round, watchlist)
bst.save_model("test_issue3402.model{}".format(xgb.rabit.get_rank()))
xgb.rabit.tracker_print("Finished training\n")
if xgb.rabit.get_rank() == 0:
bst.save_model("test_issue3402.model")
xgb.rabit.tracker_print("Finished training\n")
# Notify the tracker all training has been successful
# This is only needed in distributed training.