/** * Copyright 2021-2024, XGBoost Contributors */ #include #include // for equal #include // for make_constant_iterator #include // for sequence #include "../../../src/common/cuda_context.cuh" #include "../../../src/common/linalg_op.cuh" #include "../helpers.h" #include "xgboost/context.h" #include "xgboost/linalg.h" namespace xgboost::linalg { namespace { void TestElementWiseKernel() { auto device = DeviceOrd::CUDA(0); Tensor l{{2, 3, 4}, device}; { /** * Non-contiguous */ // GPU view auto t = l.View(device).Slice(linalg::All(), 1, linalg::All()); ASSERT_FALSE(t.CContiguous()); ElementWiseTransformDevice(t, [] __device__(size_t i, float) { return i; }); // CPU view t = l.View(DeviceOrd::CPU()).Slice(linalg::All(), 1, linalg::All()); std::size_t k = 0; for (size_t i = 0; i < l.Shape(0); ++i) { for (size_t j = 0; j < l.Shape(2); ++j) { ASSERT_EQ(k++, t(i, j)); } } t = l.View(device).Slice(linalg::All(), 1, linalg::All()); cuda_impl::ElementWiseKernel( t, [=] XGBOOST_DEVICE(std::size_t i, std::size_t j) mutable { t(i, j) = i + j; }); t = l.Slice(linalg::All(), 1, linalg::All()); for (size_t i = 0; i < l.Shape(0); ++i) { for (size_t j = 0; j < l.Shape(2); ++j) { ASSERT_EQ(i + j, t(i, j)); } } } { /** * Contiguous */ auto t = l.View(device); ElementWiseTransformDevice(t, [] XGBOOST_DEVICE(size_t i, float) { return i; }); ASSERT_TRUE(t.CContiguous()); // CPU view t = l.View(DeviceOrd::CPU()); size_t ind = 0; for (size_t i = 0; i < l.Shape(0); ++i) { for (size_t j = 0; j < l.Shape(1); ++j) { for (size_t k = 0; k < l.Shape(2); ++k) { ASSERT_EQ(ind++, t(i, j, k)); } } } } } void TestSlice() { auto ctx = MakeCUDACtx(1); thrust::device_vector data(2 * 3 * 4); auto t = MakeTensorView(&ctx, dh::ToSpan(data), 2, 3, 4); dh::LaunchN(1, [=] __device__(size_t) { auto s = t.Slice(linalg::All(), linalg::Range(0, 3), linalg::Range(0, 4)); auto all = t.Slice(linalg::All(), linalg::All(), linalg::All()); static_assert(decltype(s)::kDimension == 3); for (size_t i = 0; i < s.Shape(0); ++i) { for (size_t j = 0; j < s.Shape(1); ++j) { for (size_t k = 0; k < s.Shape(2); ++k) { SPAN_CHECK(s(i, j, k) == all(i, j, k)); } } } }); } void TestWriteAccess(CUDAContext const* cuctx, linalg::TensorView t) { thrust::for_each(cuctx->CTP(), linalg::tbegin(t), linalg::tend(t), [=] XGBOOST_DEVICE(double& v) { v = 0; }); auto eq = thrust::equal(cuctx->CTP(), linalg::tcbegin(t), linalg::tcend(t), thrust::make_constant_iterator(0.0), thrust::equal_to<>{}); ASSERT_TRUE(eq); } } // anonymous namespace TEST(Linalg, GPUElementWise) { TestElementWiseKernel(); } TEST(Linalg, GPUTensorView) { TestSlice(); } TEST(Linalg, GPUIter) { auto ctx = MakeCUDACtx(1); auto cuctx = ctx.CUDACtx(); dh::device_vector data(2 * 3 * 4); thrust::sequence(cuctx->CTP(), data.begin(), data.end(), 1.0); auto t = MakeTensorView(&ctx, dh::ToSpan(data), 2, 3, 4); static_assert(!std::is_const_v); static_assert(!std::is_const_v); auto n = std::distance(linalg::tcbegin(t), linalg::tcend(t)); ASSERT_EQ(n, t.Size()); ASSERT_FALSE(t.Empty()); bool eq = thrust::equal(cuctx->CTP(), data.cbegin(), data.cend(), linalg::tcbegin(t)); ASSERT_TRUE(eq); TestWriteAccess(cuctx, t); } } // namespace xgboost::linalg