Use context in SetInfo. (#7687)
* Use the name `Context`. * Pass a context object into `SetInfo`. * Add context to proxy matrix. * Add context to iterative DMatrix. This is to remove the use of the default number of threads during `SetInfo` as a follow-up on removing the global omp variable while preparing for CUDA stream semantic. Currently, XGBoost uses the legacy CUDA stream, we will gradually remove them in the future in favor of non-blocking streams.
This commit is contained in:
@@ -25,14 +25,13 @@ std::string PrepareData(std::string typestr, thrust::device_vector<T>* out, cons
|
||||
|
||||
std::vector<Json> j_shape {Json(Integer(static_cast<Integer::Int>(kRows)))};
|
||||
column["shape"] = Array(j_shape);
|
||||
column["strides"] = Array(std::vector<Json>{Json(Integer(static_cast<Integer::Int>(sizeof(T))))});
|
||||
column["strides"] = Array(std::vector<Json>{Json(Integer{static_cast<Integer::Int>(sizeof(T))})});
|
||||
column["version"] = 3;
|
||||
column["typestr"] = String(typestr);
|
||||
|
||||
auto p_d_data = d_data.data().get();
|
||||
std::vector<Json> j_data {
|
||||
Json(Integer(reinterpret_cast<Integer::Int>(p_d_data))),
|
||||
Json(Boolean(false))};
|
||||
std::vector<Json> j_data{Json(Integer{reinterpret_cast<Integer::Int>(p_d_data)}),
|
||||
Json(Boolean(false))};
|
||||
column["data"] = j_data;
|
||||
column["stream"] = nullptr;
|
||||
Json array(std::vector<Json>{column});
|
||||
@@ -45,12 +44,13 @@ std::string PrepareData(std::string typestr, thrust::device_vector<T>* out, cons
|
||||
|
||||
TEST(MetaInfo, FromInterface) {
|
||||
cudaSetDevice(0);
|
||||
Context ctx;
|
||||
thrust::device_vector<float> d_data;
|
||||
|
||||
std::string str = PrepareData<float>("<f4", &d_data);
|
||||
|
||||
MetaInfo info;
|
||||
info.SetInfo("label", str.c_str());
|
||||
info.SetInfo(ctx, "label", str.c_str());
|
||||
|
||||
auto const& h_label = info.labels.HostView();
|
||||
ASSERT_EQ(h_label.Size(), d_data.size());
|
||||
@@ -58,13 +58,13 @@ TEST(MetaInfo, FromInterface) {
|
||||
ASSERT_EQ(h_label(i), d_data[i]);
|
||||
}
|
||||
|
||||
info.SetInfo("weight", str.c_str());
|
||||
info.SetInfo(ctx, "weight", str.c_str());
|
||||
auto const& h_weight = info.weights_.HostVector();
|
||||
for (size_t i = 0; i < d_data.size(); ++i) {
|
||||
ASSERT_EQ(h_weight[i], d_data[i]);
|
||||
}
|
||||
|
||||
info.SetInfo("base_margin", str.c_str());
|
||||
info.SetInfo(ctx, "base_margin", str.c_str());
|
||||
auto const h_base_margin = info.base_margin_.View(GenericParameter::kCpuId);
|
||||
ASSERT_EQ(h_base_margin.Size(), d_data.size());
|
||||
for (size_t i = 0; i < d_data.size(); ++i) {
|
||||
@@ -77,7 +77,7 @@ TEST(MetaInfo, FromInterface) {
|
||||
d_group_data[1] = 3;
|
||||
d_group_data[2] = 2;
|
||||
d_group_data[3] = 1;
|
||||
info.SetInfo("group", group_str.c_str());
|
||||
info.SetInfo(ctx, "group", group_str.c_str());
|
||||
std::vector<bst_group_t> expected_group_ptr = {0, 4, 7, 9, 10};
|
||||
EXPECT_EQ(info.group_ptr_, expected_group_ptr);
|
||||
}
|
||||
@@ -89,10 +89,11 @@ TEST(MetaInfo, GPUStridedData) {
|
||||
TEST(MetaInfo, Group) {
|
||||
cudaSetDevice(0);
|
||||
MetaInfo info;
|
||||
Context ctx;
|
||||
|
||||
thrust::device_vector<uint32_t> d_uint;
|
||||
std::string uint_str = PrepareData<uint32_t>("<u4", &d_uint);
|
||||
info.SetInfo("group", uint_str.c_str());
|
||||
info.SetInfo(ctx, "group", uint_str.c_str());
|
||||
auto& h_group = info.group_ptr_;
|
||||
ASSERT_EQ(h_group.size(), d_uint.size() + 1);
|
||||
for (size_t i = 1; i < h_group.size(); ++i) {
|
||||
@@ -102,7 +103,7 @@ TEST(MetaInfo, Group) {
|
||||
thrust::device_vector<int64_t> d_int64;
|
||||
std::string int_str = PrepareData<int64_t>("<i8", &d_int64);
|
||||
info = MetaInfo();
|
||||
info.SetInfo("group", int_str.c_str());
|
||||
info.SetInfo(ctx, "group", int_str.c_str());
|
||||
h_group = info.group_ptr_;
|
||||
ASSERT_EQ(h_group.size(), d_uint.size() + 1);
|
||||
for (size_t i = 1; i < h_group.size(); ++i) {
|
||||
@@ -113,11 +114,12 @@ TEST(MetaInfo, Group) {
|
||||
thrust::device_vector<float> d_float;
|
||||
std::string float_str = PrepareData<float>("<f4", &d_float);
|
||||
info = MetaInfo();
|
||||
EXPECT_ANY_THROW(info.SetInfo("group", float_str.c_str()));
|
||||
EXPECT_ANY_THROW(info.SetInfo(ctx, "group", float_str.c_str()));
|
||||
}
|
||||
|
||||
TEST(MetaInfo, GPUQid) {
|
||||
xgboost::MetaInfo info;
|
||||
Context ctx;
|
||||
info.num_row_ = 100;
|
||||
thrust::device_vector<uint32_t> qid(info.num_row_, 0);
|
||||
for (size_t i = 0; i < qid.size(); ++i) {
|
||||
@@ -127,7 +129,7 @@ TEST(MetaInfo, GPUQid) {
|
||||
Json array{std::vector<Json>{column}};
|
||||
std::string array_str;
|
||||
Json::Dump(array, &array_str);
|
||||
info.SetInfo("qid", array_str.c_str());
|
||||
info.SetInfo(ctx, "qid", array_str.c_str());
|
||||
ASSERT_EQ(info.group_ptr_.size(), info.num_row_ + 1);
|
||||
ASSERT_EQ(info.group_ptr_.front(), 0);
|
||||
ASSERT_EQ(info.group_ptr_.back(), info.num_row_);
|
||||
@@ -142,11 +144,12 @@ TEST(MetaInfo, DeviceExtend) {
|
||||
dh::safe_cuda(cudaSetDevice(0));
|
||||
size_t const kRows = 100;
|
||||
MetaInfo lhs, rhs;
|
||||
Context ctx;
|
||||
|
||||
thrust::device_vector<float> d_data;
|
||||
std::string str = PrepareData<float>("<f4", &d_data, kRows);
|
||||
lhs.SetInfo("label", str.c_str());
|
||||
rhs.SetInfo("label", str.c_str());
|
||||
lhs.SetInfo(ctx, "label", str.c_str());
|
||||
rhs.SetInfo(ctx, "label", str.c_str());
|
||||
ASSERT_FALSE(rhs.labels.Data()->HostCanRead());
|
||||
lhs.num_row_ = kRows;
|
||||
rhs.num_row_ = kRows;
|
||||
|
||||
Reference in New Issue
Block a user