Skip to content

Commit

Permalink
c++17.
Browse files Browse the repository at this point in the history
  • Loading branch information
trivialfis committed Oct 13, 2024
1 parent 2f15f6d commit 874c24c
Show file tree
Hide file tree
Showing 5 changed files with 22 additions and 43 deletions.
13 changes: 2 additions & 11 deletions src/c_api/c_api.cc
Original file line number Diff line number Diff line change
Expand Up @@ -322,12 +322,7 @@ XGB_DLL int XGDMatrixCreateFromCallback(DataIterHandle iter, DMatrixHandle proxy
xgboost_CHECK_C_ARG_PTR(reset);
xgboost_CHECK_C_ARG_PTR(out);

ExtMemConfig config{.cache = cache,
.on_host = on_host,
.min_cache_page_bytes = min_cache_page_bytes,
.missing = missing,
.n_threads = n_threads};

auto config = ExtMemConfig{cache, on_host, min_cache_page_bytes, missing, n_threads};
*out = new std::shared_ptr<xgboost::DMatrix>{
xgboost::DMatrix::Create(iter, proxy, reset, next, config)};
API_END();
Expand Down Expand Up @@ -403,11 +398,7 @@ XGB_DLL int XGExtMemQuantileDMatrixCreateFromCallback(DataIterHandle iter, DMatr
xgboost_CHECK_C_ARG_PTR(reset);
xgboost_CHECK_C_ARG_PTR(out);

ExtMemConfig config{.cache = cache,
.on_host = on_host,
.min_cache_page_bytes = min_cache_page_bytes,
.missing = missing,
.n_threads = n_threads};
auto config = ExtMemConfig{cache, on_host, min_cache_page_bytes, missing, n_threads};
*out = new std::shared_ptr<xgboost::DMatrix>{
xgboost::DMatrix::Create(iter, proxy, p_ref, reset, next, max_bin, config)};
API_END();
Expand Down
7 changes: 2 additions & 5 deletions src/data/data.cc
Original file line number Diff line number Diff line change
Expand Up @@ -915,11 +915,8 @@ DMatrix* DMatrix::Load(const std::string& uri, bool silent, DataSplitMode data_s
CHECK(data_split_mode != DataSplitMode::kCol)
<< "Column-wise data split is not supported for external memory.";
data::FileIterator iter{fname, static_cast<uint32_t>(partid), static_cast<uint32_t>(npart)};
auto config = ExtMemConfig{.cache = cache_file,
.on_host = false,
.min_cache_page_bytes = cuda_impl::MatchingPageBytes(),
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = 1};
auto config = ExtMemConfig{cache_file, false, cuda_impl::MatchingPageBytes(),
std::numeric_limits<float>::quiet_NaN(), 1};
dmat = new data::SparsePageDMatrix{&iter, iter.Proxy(), data::fileiter::Reset,
data::fileiter::Next, config};
}
Expand Down
7 changes: 2 additions & 5 deletions tests/cpp/c_api/test_c_api.cc
Original file line number Diff line number Diff line change
Expand Up @@ -496,11 +496,8 @@ auto MakeExtMemForTest(bst_idx_t n_samples, bst_feature_t n_features, Json dconf
0);

NumpyArrayIterForTest iter_1{0.0f, n_samples, n_features, n_batches};
auto config = ExtMemConfig{.cache = "",
.on_host = false,
.min_cache_page_bytes = cuda_impl::MatchingPageBytes(),
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = 0};
auto config = ExtMemConfig{"", false, cuda_impl::MatchingPageBytes(),
std::numeric_limits<float>::quiet_NaN(), 0};
auto Xy = std::make_shared<data::SparsePageDMatrix>(&iter_1, iter_1.Proxy(), Reset, Next, config);
MakeLabelForTest(Xy, p_fmat);
return std::pair{p_fmat, Xy};
Expand Down
14 changes: 4 additions & 10 deletions tests/cpp/data/test_sparse_page_dmatrix.cc
Original file line number Diff line number Diff line change
Expand Up @@ -32,11 +32,8 @@ void TestSparseDMatrixLoadFile(Context const* ctx) {
opath += "?indexing_mode=1&format=libsvm";
data::FileIterator iter{opath, 0, 1};
auto n_threads = 0;
auto config = ExtMemConfig{.cache = tmpdir.path + "cache",
.on_host = false,
.min_cache_page_bytes = cuda_impl::MatchingPageBytes(),
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = n_threads};
auto config = ExtMemConfig{tmpdir.path + "cache", false, cuda_impl::MatchingPageBytes(),
std::numeric_limits<float>::quiet_NaN(), n_threads};
data::SparsePageDMatrix m{&iter, iter.Proxy(), data::fileiter::Reset, data::fileiter::Next,
config};
ASSERT_EQ(AllThreadsForTest(), m.Ctx()->Threads());
Expand Down Expand Up @@ -365,11 +362,8 @@ auto TestSparsePageDMatrixDeterminism(int32_t threads) {
CreateBigTestData(filename, 1 << 16);

data::FileIterator iter(filename + "?format=libsvm", 0, 1);
auto config = ExtMemConfig{.cache = filename,
.on_host = false,
.min_cache_page_bytes = cuda_impl::MatchingPageBytes(),
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = threads};
auto config = ExtMemConfig{filename, false, cuda_impl::MatchingPageBytes(),
std::numeric_limits<float>::quiet_NaN(), threads};
std::unique_ptr<DMatrix> sparse{new data::SparsePageDMatrix{
&iter, iter.Proxy(), data::fileiter::Reset, data::fileiter::Next, config}};
CHECK(sparse->Ctx()->Threads() == threads || sparse->Ctx()->Threads() == AllThreadsForTest());
Expand Down
24 changes: 12 additions & 12 deletions tests/cpp/helpers.cc
Original file line number Diff line number Diff line change
Expand Up @@ -448,12 +448,12 @@ void MakeLabels(DeviceOrd device, bst_idx_t n_samples, bst_target_t n_classes,
#endif // defined(XGBOOST_USE_CUDA)
}

ExtMemConfig config{
.cache = prefix,
.on_host = this->on_host_,
.min_cache_page_bytes = this->min_cache_page_bytes_,
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = Context{}.Threads(),
auto config = ExtMemConfig{
prefix,
this->on_host_,
this->min_cache_page_bytes_,
std::numeric_limits<float>::quiet_NaN(),
Context{}.Threads(),
};
std::shared_ptr<DMatrix> p_fmat{
DMatrix::Create(static_cast<DataIterHandle>(iter.get()), iter->Proxy(), Reset, Next, config)};
Expand Down Expand Up @@ -497,12 +497,12 @@ void MakeLabels(DeviceOrd device, bst_idx_t n_samples, bst_target_t n_classes,
}
CHECK(iter);

ExtMemConfig config{
.cache = prefix,
.on_host = this->on_host_,
.min_cache_page_bytes = this->min_cache_page_bytes_,
.missing = std::numeric_limits<float>::quiet_NaN(),
.n_threads = Context{}.Threads(),
auto config = ExtMemConfig{
prefix,
this->on_host_,
this->min_cache_page_bytes_,
std::numeric_limits<float>::quiet_NaN(),
Context{}.Threads(),
};
std::shared_ptr<DMatrix> p_fmat{DMatrix::Create(static_cast<DataIterHandle>(iter.get()),
iter->Proxy(), nullptr, Reset, Next, this->bins_,
Expand Down

0 comments on commit 874c24c

Please sign in to comment.