Replaced std::vector-based interfaces with HostDeviceVector-based interfaces. (#3116)

* Replaced std::vector-based interfaces with HostDeviceVector-based interfaces.

- replacement was performed in the learner, boosters, predictors,
  updaters, and objective functions
- only interfaces used in training were replaced;
  interfaces like PredictInstance() still use std::vector
- refactoring necessary for replacement of interfaces was also performed,
  such as using HostDeviceVector in prediction cache

* HostDeviceVector-based interfaces for custom objective function example plugin.
This commit is contained in:
Andrew V. Adinetz
2018-02-28 01:00:04 +01:00
committed by Rory Mitchell
parent 11bfa8584d
commit d5992dd881
38 changed files with 371 additions and 519 deletions

View File

@@ -46,10 +46,11 @@ TEST(Objective, LogisticRegressionBasic) {
<< "Expected error when base_score not in range [0,1f] for LogisticRegression";
// test PredTransform
std::vector<xgboost::bst_float> preds = {0, 0.1f, 0.5f, 0.9f, 1};
xgboost::HostDeviceVector<xgboost::bst_float> io_preds = {0, 0.1f, 0.5f, 0.9f, 1};
std::vector<xgboost::bst_float> out_preds = {0.5f, 0.524f, 0.622f, 0.710f, 0.731f};
obj->PredTransform(&preds);
for (int i = 0; i < static_cast<int>(preds.size()); ++i) {
obj->PredTransform(&io_preds);
auto& preds = io_preds.data_h();
for (int i = 0; i < static_cast<int>(io_preds.size()); ++i) {
EXPECT_NEAR(preds[i], out_preds[i], 0.01f);
}
}
@@ -94,10 +95,11 @@ TEST(Objective, PoissonRegressionBasic) {
EXPECT_NEAR(obj->ProbToMargin(0.9f), -0.10f, 0.01f);
// test PredTransform
std::vector<xgboost::bst_float> preds = {0, 0.1f, 0.5f, 0.9f, 1};
xgboost::HostDeviceVector<xgboost::bst_float> io_preds = {0, 0.1f, 0.5f, 0.9f, 1};
std::vector<xgboost::bst_float> out_preds = {1, 1.10f, 1.64f, 2.45f, 2.71f};
obj->PredTransform(&preds);
for (int i = 0; i < static_cast<int>(preds.size()); ++i) {
obj->PredTransform(&io_preds);
auto& preds = io_preds.data_h();
for (int i = 0; i < static_cast<int>(io_preds.size()); ++i) {
EXPECT_NEAR(preds[i], out_preds[i], 0.01f);
}
}
@@ -129,10 +131,11 @@ TEST(Objective, GammaRegressionBasic) {
EXPECT_NEAR(obj->ProbToMargin(0.9f), -0.10f, 0.01f);
// test PredTransform
std::vector<xgboost::bst_float> preds = {0, 0.1f, 0.5f, 0.9f, 1};
xgboost::HostDeviceVector<xgboost::bst_float> io_preds = {0, 0.1f, 0.5f, 0.9f, 1};
std::vector<xgboost::bst_float> out_preds = {1, 1.10f, 1.64f, 2.45f, 2.71f};
obj->PredTransform(&preds);
for (int i = 0; i < static_cast<int>(preds.size()); ++i) {
obj->PredTransform(&io_preds);
auto& preds = io_preds.data_h();
for (int i = 0; i < static_cast<int>(io_preds.size()); ++i) {
EXPECT_NEAR(preds[i], out_preds[i], 0.01f);
}
}
@@ -165,10 +168,11 @@ TEST(Objective, TweedieRegressionBasic) {
EXPECT_NEAR(obj->ProbToMargin(0.9f), 0.89f, 0.01f);
// test PredTransform
std::vector<xgboost::bst_float> preds = {0, 0.1f, 0.5f, 0.9f, 1};
xgboost::HostDeviceVector<xgboost::bst_float> io_preds = {0, 0.1f, 0.5f, 0.9f, 1};
std::vector<xgboost::bst_float> out_preds = {1, 1.10f, 1.64f, 2.45f, 2.71f};
obj->PredTransform(&preds);
for (int i = 0; i < static_cast<int>(preds.size()); ++i) {
obj->PredTransform(&io_preds);
auto& preds = io_preds.data_h();
for (int i = 0; i < static_cast<int>(io_preds.size()); ++i) {
EXPECT_NEAR(preds[i], out_preds[i], 0.01f);
}
}

View File

@@ -48,10 +48,11 @@ TEST(Objective, GPULogisticRegressionBasic) {
<< "Expected error when base_score not in range [0,1f] for LogisticRegression";
// test PredTransform
std::vector<xgboost::bst_float> preds = {0, 0.1f, 0.5f, 0.9f, 1};
xgboost::HostDeviceVector<xgboost::bst_float> io_preds = {0, 0.1f, 0.5f, 0.9f, 1};
std::vector<xgboost::bst_float> out_preds = {0.5f, 0.524f, 0.622f, 0.710f, 0.731f};
obj->PredTransform(&preds);
for (int i = 0; i < static_cast<int>(preds.size()); ++i) {
obj->PredTransform(&io_preds);
auto& preds = io_preds.data_h();
for (int i = 0; i < static_cast<int>(io_preds.size()); ++i) {
EXPECT_NEAR(preds[i], out_preds[i], 0.01f);
}
}