Compare commits
1252 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
096047c547 | ||
|
|
e75dd75bb2 | ||
|
|
4d387cbfbf | ||
|
|
3fde9361d7 | ||
|
|
b67c2ed96d | ||
|
|
177fd79864 | ||
|
|
06487d3896 | ||
|
|
e50ccc4d3c | ||
|
|
add57f8880 | ||
|
|
a0d3573c74 | ||
|
|
4301558a57 | ||
|
|
68be454cfa | ||
|
|
5188e27513 | ||
|
|
f380c10a93 | ||
|
|
12fe2fc06c | ||
|
|
b2e93d2742 | ||
|
|
c061e3ae50 | ||
|
|
b82e78c169 | ||
|
|
8463107013 | ||
|
|
19b59938b7 | ||
|
|
e3f624d8e7 | ||
|
|
2c84daeca7 | ||
|
|
344f90b67b | ||
|
|
05d7000096 | ||
|
|
f03463c45b | ||
|
|
fd4335d0bf | ||
|
|
801116c307 | ||
|
|
bb56183396 | ||
|
|
bdc1a3c178 | ||
|
|
428f6cbbe2 | ||
|
|
d638535581 | ||
|
|
44bd2981b2 | ||
|
|
9dbb71490c | ||
|
|
4359356d46 | ||
|
|
1caa93221a | ||
|
|
a57371ef7c | ||
|
|
f05a23b41c | ||
|
|
d495a180d8 | ||
|
|
7f854848d3 | ||
|
|
f05294a6f2 | ||
|
|
819098a48f | ||
|
|
c1b2cff874 | ||
|
|
7ce090e775 | ||
|
|
97fd5207dd | ||
|
|
54029a59af | ||
|
|
5bd163aa25 | ||
|
|
7fc57f3974 | ||
|
|
bde1ebc209 | ||
|
|
1aabc690ec | ||
|
|
04c99683c3 | ||
|
|
1332ff787f | ||
|
|
f958e32683 | ||
|
|
7129988847 | ||
|
|
e93a274823 | ||
|
|
c2b85ab68a | ||
|
|
a9da2e244a | ||
|
|
912e341d57 | ||
|
|
8f0efb4ab3 | ||
|
|
7579905e18 | ||
|
|
54579da4d7 | ||
|
|
3a9996173e | ||
|
|
1b657a5513 | ||
|
|
a196443a07 | ||
|
|
851cba931e | ||
|
|
01e00efc53 | ||
|
|
275da176ba | ||
|
|
22b0a55a04 | ||
|
|
0de7c47495 | ||
|
|
dbd5309b55 | ||
|
|
f7f673b00c | ||
|
|
7a0ccfbb49 | ||
|
|
0897477af0 | ||
|
|
e082718c66 | ||
|
|
6e18d3a290 | ||
|
|
2a0ff209ff | ||
|
|
f4fb2be101 | ||
|
|
2caceb157d | ||
|
|
b342ef951b | ||
|
|
0a07900b9f | ||
|
|
16eb41936d | ||
|
|
9da5050643 | ||
|
|
04aff3af8e | ||
|
|
2d0cd2817e | ||
|
|
a1367ea1f8 | ||
|
|
3632242e0b | ||
|
|
97ed944209 | ||
|
|
20c52f07d2 | ||
|
|
c3124813e8 | ||
|
|
59787b23af | ||
|
|
15ca12a77e | ||
|
|
41c6813496 | ||
|
|
b572a39919 | ||
|
|
645037e376 | ||
|
|
6c9c8a9001 | ||
|
|
bb2de1fd5d | ||
|
|
d0916849a6 | ||
|
|
6155394a06 | ||
|
|
e964654b8f | ||
|
|
39390cc2ee | ||
|
|
3a0f787703 | ||
|
|
f90771eec6 | ||
|
|
f4798718c7 | ||
|
|
bc267dd729 | ||
|
|
96c3071a8a | ||
|
|
cfa9c42eb4 | ||
|
|
6efe7c129f | ||
|
|
54da4b3185 | ||
|
|
4066d68261 | ||
|
|
6d22ea793c | ||
|
|
ee6809e642 | ||
|
|
d8beb517ed | ||
|
|
2718ff530c | ||
|
|
0df1272695 | ||
|
|
e70810be8a | ||
|
|
c2f0486d37 | ||
|
|
aad1313154 | ||
|
|
2b76061659 | ||
|
|
152e2fb072 | ||
|
|
ea0deeca68 | ||
|
|
8c1065f645 | ||
|
|
1fcc26a6f8 | ||
|
|
e4dd6051a0 | ||
|
|
2ec2ecf013 | ||
|
|
181dee13e9 | ||
|
|
ff122d61ff | ||
|
|
84d3fcb7ea | ||
|
|
e229692572 | ||
|
|
4a5802ed2c | ||
|
|
0cba2cdbb0 | ||
|
|
fc8110ef79 | ||
|
|
7f9cb921f4 | ||
|
|
a474a66573 | ||
|
|
962a20693f | ||
|
|
3bf0f145bb | ||
|
|
a1fad72ab3 | ||
|
|
288539ac78 | ||
|
|
9fbde21e9d | ||
|
|
db8288121d | ||
|
|
bb2a17b90c | ||
|
|
e93b805a75 | ||
|
|
fa2ab1f021 | ||
|
|
aba4559c4f | ||
|
|
7f20eaed93 | ||
|
|
62e9387cd5 | ||
|
|
17fd3f55e9 | ||
|
|
097f11b6e0 | ||
|
|
6f83d9c69a | ||
|
|
ae7450ce54 | ||
|
|
ddec0f378c | ||
|
|
320323f533 | ||
|
|
03bc6e6427 | ||
|
|
d563d6d8f4 | ||
|
|
a01df102c9 | ||
|
|
8c174ef2d3 | ||
|
|
c5677a2b2c | ||
|
|
053aababd4 | ||
|
|
614f47c477 | ||
|
|
5b69534b43 | ||
|
|
acd363033e | ||
|
|
5d99b441d5 | ||
|
|
e38e94ba4d | ||
|
|
d6d83c818f | ||
|
|
22b0fc0992 | ||
|
|
e67a0b8599 | ||
|
|
3913ff470f | ||
|
|
6274fba0a5 | ||
|
|
caf326d508 | ||
|
|
cb370c4f7d | ||
|
|
7375bd058b | ||
|
|
59edfdb315 | ||
|
|
779b82c098 | ||
|
|
603f8ce2fa | ||
|
|
52311dcec9 | ||
|
|
e4129ed6ee | ||
|
|
2ab6660943 | ||
|
|
d21e7e5f82 | ||
|
|
0cd4382d72 | ||
|
|
09b44915e7 | ||
|
|
85988a3178 | ||
|
|
a075aa24ba | ||
|
|
55968ed3fa | ||
|
|
250b22dd22 | ||
|
|
47b3cb6fb7 | ||
|
|
07b2d5a26d | ||
|
|
08ce495b5d | ||
|
|
1f9a57d17b | ||
|
|
e922004329 | ||
|
|
17ff471616 | ||
|
|
fb941262b4 | ||
|
|
e206b899ef | ||
|
|
0e470ef606 | ||
|
|
fa267ad093 | ||
|
|
96d3f8a6f3 | ||
|
|
511d4996b5 | ||
|
|
101a2e643d | ||
|
|
353ed5339d | ||
|
|
0e7377ba9c | ||
|
|
a320b402a5 | ||
|
|
49ccae7fb9 | ||
|
|
17add4776f | ||
|
|
a2cc78c1fb | ||
|
|
339f21e1bf | ||
|
|
d237378452 | ||
|
|
c512c3f46b | ||
|
|
8dbe0510de | ||
|
|
7032981350 | ||
|
|
3b742dc4f1 | ||
|
|
39b0fde0e7 | ||
|
|
ee84e22c8d | ||
|
|
b908680bec | ||
|
|
a5cd2412de | ||
|
|
a7b3dd3176 | ||
|
|
2acd78b44b | ||
|
|
a84a1fde02 | ||
|
|
564df59204 | ||
|
|
42d100de18 | ||
|
|
ef13dd31b1 | ||
|
|
ba9d24ff7b | ||
|
|
191d0aa5cf | ||
|
|
8e0f320db3 | ||
|
|
fe9dff339c | ||
|
|
2c8d735cb3 | ||
|
|
ebd64f6e22 | ||
|
|
1cf4d93246 | ||
|
|
15e073ca9d | ||
|
|
720a8c3273 | ||
|
|
4caca2947d | ||
|
|
bcb55d3b6a | ||
|
|
bac22734fb | ||
|
|
b647403baa | ||
|
|
cd05e38533 | ||
|
|
d062a9e009 | ||
|
|
d385cc64e2 | ||
|
|
a58055075b | ||
|
|
6676c28cbc | ||
|
|
ff26cd3212 | ||
|
|
401ce5cf5e | ||
|
|
acc110c251 | ||
|
|
c2b3a13e70 | ||
|
|
90645c4957 | ||
|
|
43878b10b6 | ||
|
|
cff50fe3ef | ||
|
|
21a52c7f98 | ||
|
|
bf88dadb61 | ||
|
|
15a2724ff7 | ||
|
|
151882dd26 | ||
|
|
ea04d4c46c | ||
|
|
a551bed803 | ||
|
|
a05799ed39 | ||
|
|
5891f752c8 | ||
|
|
b240f055d3 | ||
|
|
8dc1e4b3ea | ||
|
|
34092d7fd0 | ||
|
|
9b6cc0ed07 | ||
|
|
36263dd109 | ||
|
|
55ed50c860 | ||
|
|
a093770f36 | ||
|
|
26209a42a5 | ||
|
|
a2cdba51ce | ||
|
|
fd016e43c6 | ||
|
|
f186c87cf9 | ||
|
|
72e8331eab | ||
|
|
8685556af2 | ||
|
|
95e2baf7c2 | ||
|
|
910ce580c8 | ||
|
|
c400fa1e8d | ||
|
|
8be6095ece | ||
|
|
9bade7203a | ||
|
|
5ba3509dd3 | ||
|
|
bbee355b45 | ||
|
|
3689695d16 | ||
|
|
36a7396658 | ||
|
|
79efcd37f5 | ||
|
|
2aa838c75e | ||
|
|
6deaec8027 | ||
|
|
54e001bbf4 | ||
|
|
c5c8f643f2 | ||
|
|
5feee8d4a9 | ||
|
|
46dfcc7d22 | ||
|
|
e8a69013e6 | ||
|
|
8c16da8863 | ||
|
|
85c3334c2b | ||
|
|
f236640427 | ||
|
|
f53055f75e | ||
|
|
f7ce0ec0df | ||
|
|
2b2eb0d0f1 | ||
|
|
5eabcae27b | ||
|
|
d06b1fc26e | ||
|
|
ffa5eb2aa4 | ||
|
|
0f6c502d36 | ||
|
|
7eba285a1e | ||
|
|
cad7401783 | ||
|
|
228a46e8ad | ||
|
|
173096a6a7 | ||
|
|
6a892ce281 | ||
|
|
4d665b3fb0 | ||
|
|
2dc22e7aad | ||
|
|
69a50248b7 | ||
|
|
6cef9a08e9 | ||
|
|
803d5e3c4c | ||
|
|
a5852365fd | ||
|
|
7cbaee9916 | ||
|
|
6d8afb2218 | ||
|
|
d54ef56f6f | ||
|
|
d9688f93c7 | ||
|
|
90c0633a28 | ||
|
|
a65ad0bd9c | ||
|
|
40fd3d6d5f | ||
|
|
6ce9a35f55 | ||
|
|
d62daa0b32 | ||
|
|
c0afdb6786 | ||
|
|
cce4af4acf | ||
|
|
282b1729da | ||
|
|
c7c485d052 | ||
|
|
594371e35b | ||
|
|
e62167937b | ||
|
|
74572b5d45 | ||
|
|
f27a7258c6 | ||
|
|
52d0230b58 | ||
|
|
81b2ee1153 | ||
|
|
31d3ec07af | ||
|
|
457f704e3d | ||
|
|
d11a0044cf | ||
|
|
ed91e775ec | ||
|
|
225b3158f6 | ||
|
|
17b709acb9 | ||
|
|
70c9b885ef | ||
|
|
e9c178f402 | ||
|
|
8a16944664 | ||
|
|
ad0ccc6e4f | ||
|
|
199c421d60 | ||
|
|
5f76edd296 | ||
|
|
4ead65a28c | ||
|
|
cbf98cb9c6 | ||
|
|
48cefa012e | ||
|
|
c4802bfcd0 | ||
|
|
7b3d473593 | ||
|
|
28bb01aa22 | ||
|
|
0f37a01dd9 | ||
|
|
a2e433a089 | ||
|
|
7214a45e83 | ||
|
|
66191e9926 | ||
|
|
c1786849e3 | ||
|
|
213b5602d9 | ||
|
|
dd79ab846f | ||
|
|
0e61ba57d6 | ||
|
|
16ef016ba7 | ||
|
|
0d8248ddcd | ||
|
|
1325ba9251 | ||
|
|
3760cede0f | ||
|
|
21a28f2cc5 | ||
|
|
8af98e30fc | ||
|
|
96e6b6beba | ||
|
|
d29e45371f | ||
|
|
0f4d52a864 | ||
|
|
9fb12b20a4 | ||
|
|
34eee56256 | ||
|
|
e49e0998c0 | ||
|
|
7a068af1a3 | ||
|
|
6933240837 | ||
|
|
4416452f94 | ||
|
|
31b9cbab3d | ||
|
|
175986b739 | ||
|
|
78396f8a6e | ||
|
|
980233e648 | ||
|
|
247946a875 | ||
|
|
06ba285f71 | ||
|
|
43152657d4 | ||
|
|
9f598efc3e | ||
|
|
d6018eb4b9 | ||
|
|
0ae8df9a65 | ||
|
|
a9c6199723 | ||
|
|
37d4482e3e | ||
|
|
e227abc57a | ||
|
|
e7d612d22c | ||
|
|
292df67824 | ||
|
|
f7a2f52136 | ||
|
|
07cf3d3e53 | ||
|
|
e27cda7626 | ||
|
|
b2b6a8aa39 | ||
|
|
72ec0c5484 | ||
|
|
cfa994d57f | ||
|
|
badeff1d74 | ||
|
|
1b58d81315 | ||
|
|
4e12f3e1bc | ||
|
|
06a1cb6e03 | ||
|
|
2b88099c74 | ||
|
|
e68a152d9e | ||
|
|
26c9882e23 | ||
|
|
d3ad0524e7 | ||
|
|
beefd28471 | ||
|
|
d308124910 | ||
|
|
fa44a33ee6 | ||
|
|
6eaddaa9c3 | ||
|
|
8d545ab2a2 | ||
|
|
20e6087579 | ||
|
|
dd72af2620 | ||
|
|
9a98c3726c | ||
|
|
b05abfc494 | ||
|
|
3ceeb8c61c | ||
|
|
77b069c25d | ||
|
|
c7e82b5914 | ||
|
|
f489d824ca | ||
|
|
c430ae52f3 | ||
|
|
5bf9e79413 | ||
|
|
c6a8754c62 | ||
|
|
e01639548a | ||
|
|
17ce1f26c8 | ||
|
|
53e6e32718 | ||
|
|
f6effa1734 | ||
|
|
a10e4cba4e | ||
|
|
38887a1876 | ||
|
|
43a647a4dd | ||
|
|
7dc3e95a77 | ||
|
|
0c38ca7f6e | ||
|
|
001e663d42 | ||
|
|
7a07dcf651 | ||
|
|
eac980fbfc | ||
|
|
06ea6c7e79 | ||
|
|
f64871c74a | ||
|
|
40343c8ee1 | ||
|
|
15a88ceef0 | ||
|
|
35d8447282 | ||
|
|
42e6fbb0db | ||
|
|
deb3edf562 | ||
|
|
8824b40961 | ||
|
|
0caf2be684 | ||
|
|
ffee35e0f0 | ||
|
|
fbe40d00d8 | ||
|
|
40a1a2ffa8 | ||
|
|
5aeb8f7009 | ||
|
|
f592a5125b | ||
|
|
27aea6c7b5 | ||
|
|
5540019373 | ||
|
|
8c6630c310 | ||
|
|
b7ffdcdbb9 | ||
|
|
4f1e453ff5 | ||
|
|
3e26107a9c | ||
|
|
05fc6f3ca9 | ||
|
|
e38fe21e0d | ||
|
|
7774bf628e | ||
|
|
4a99c9bdb8 | ||
|
|
d99bdd1b1e | ||
|
|
ed1a4f3205 | ||
|
|
e85c9b987b | ||
|
|
7ac52e674f | ||
|
|
2790e3091f | ||
|
|
0c1769b3a5 | ||
|
|
67752e3967 | ||
|
|
8afcecc025 | ||
|
|
e143a4dd7e | ||
|
|
42c5ee5588 | ||
|
|
e3bf5565ab | ||
|
|
d8544e4d9e | ||
|
|
d8d2eefa63 | ||
|
|
8e8d3ac708 | ||
|
|
3bfe90c183 | ||
|
|
a903241fbf | ||
|
|
f1e9bbcee5 | ||
|
|
78d65a1928 | ||
|
|
c0609b98f1 | ||
|
|
ba0ed255ef | ||
|
|
1d8bb7332f | ||
|
|
dcc92a6703 | ||
|
|
fcafd3a777 | ||
|
|
b23e97f8b0 | ||
|
|
8e41ad24f5 | ||
|
|
f747e05eac | ||
|
|
2546d139d6 | ||
|
|
7c6f2346d3 | ||
|
|
f550109641 | ||
|
|
9a98e79649 | ||
|
|
a8255ea678 | ||
|
|
157e98edf7 | ||
|
|
addaa63732 | ||
|
|
3fc1046fd3 | ||
|
|
d666ba775e | ||
|
|
f2209c1fe4 | ||
|
|
67ea1c3435 | ||
|
|
8f97c92541 | ||
|
|
e07245f110 | ||
|
|
284dcf8d22 | ||
|
|
5f1a6fca0d | ||
|
|
58d211545f | ||
|
|
2dde65f807 | ||
|
|
3b8a0e08f7 | ||
|
|
30b1a26fc0 | ||
|
|
812d577597 | ||
|
|
16f96b6cfb | ||
|
|
0c6266bc4a | ||
|
|
9dd8d70f0e | ||
|
|
0252d504d8 | ||
|
|
a83748eb45 | ||
|
|
4449e30184 | ||
|
|
ca0f7f2714 | ||
|
|
0b36f8fba1 | ||
|
|
8e76f5f595 | ||
|
|
0d3da9869c | ||
|
|
bf8de227a9 | ||
|
|
b1b2524dbb | ||
|
|
99fa8dad2d | ||
|
|
0db903b471 | ||
|
|
917cbc0699 | ||
|
|
2ed3c29c8a | ||
|
|
7174d60ed2 | ||
|
|
a408c34558 | ||
|
|
cfd2a9f872 | ||
|
|
f73520bfff | ||
|
|
ae27e228c4 | ||
|
|
5699f60a88 | ||
|
|
a2593e60bf | ||
|
|
786aa27134 | ||
|
|
cf70864fa3 | ||
|
|
7e53189e7c | ||
|
|
ba9cc43464 | ||
|
|
8bb55949ef | ||
|
|
d0b99bdd95 | ||
|
|
1d2f6de573 | ||
|
|
a3b8bca46a | ||
|
|
bb5e18c29c | ||
|
|
5761f27e5e | ||
|
|
99467f3999 | ||
|
|
28a466ab51 | ||
|
|
5bd849f1b5 | ||
|
|
c69af90319 | ||
|
|
c884b9e888 | ||
|
|
52977f0cdf | ||
|
|
c47c71e34f | ||
|
|
76f95a6667 | ||
|
|
3901f5d9db | ||
|
|
521086d56b | ||
|
|
5647fc6542 | ||
|
|
8f3dee58be | ||
|
|
031d66ec27 | ||
|
|
2176e511fc | ||
|
|
fcddbc9264 | ||
|
|
80e10e02ab | ||
|
|
b3208aac4e | ||
|
|
748d516c50 | ||
|
|
4633b476e9 | ||
|
|
3ef1703553 | ||
|
|
29595102b9 | ||
|
|
2faa744aba | ||
|
|
97a5b088a5 | ||
|
|
ce0382dcb0 | ||
|
|
39afdac3be | ||
|
|
210915c985 | ||
|
|
c68684ff4c | ||
|
|
5545c49cfc | ||
|
|
e1f9f80df2 | ||
|
|
a71421e825 | ||
|
|
d70e59fefc | ||
|
|
50ff8a2623 | ||
|
|
bc7a6ec603 | ||
|
|
c24e9d712c | ||
|
|
8d4038da57 | ||
|
|
66fd9f5207 | ||
|
|
909e49e214 | ||
|
|
668b8a0ea4 | ||
|
|
e47b3a3da3 | ||
|
|
97c3a80a34 | ||
|
|
b2bbf49015 | ||
|
|
d686bf52a6 | ||
|
|
ca0547bb65 | ||
|
|
37886a5dff | ||
|
|
9af99760d4 | ||
|
|
299e5000a4 | ||
|
|
55cf24cc32 | ||
|
|
b14c44ee5e | ||
|
|
cbf3a5f918 | ||
|
|
c91fed083d | ||
|
|
6925b222e0 | ||
|
|
f835368bcf | ||
|
|
6d1452074a | ||
|
|
fcab51aa82 | ||
|
|
1082ccd3cc | ||
|
|
8f77677193 | ||
|
|
4056974e37 | ||
|
|
ff71c69adf | ||
|
|
0cd11b893a | ||
|
|
be5b95e743 | ||
|
|
8d247f0d64 | ||
|
|
ab342af242 | ||
|
|
3fd331f8f2 | ||
|
|
eb7bbee2c9 | ||
|
|
b791446623 | ||
|
|
7d43e74e71 | ||
|
|
fffb1fca52 | ||
|
|
4f42aa5f12 | ||
|
|
520586ffa7 | ||
|
|
70df36c99c | ||
|
|
2e63af6117 | ||
|
|
bdf265076d | ||
|
|
a2686543a9 | ||
|
|
bc818316f2 | ||
|
|
dd44ac91b8 | ||
|
|
23faf656ad | ||
|
|
e888eb2fa9 | ||
|
|
b397d64c96 | ||
|
|
dc07137a2c | ||
|
|
b5eb36f1af | ||
|
|
441ffc017a | ||
|
|
7ee10e3dbd | ||
|
|
ada4a86d1c | ||
|
|
deae99e662 | ||
|
|
b78bc734d9 | ||
|
|
56395d120b | ||
|
|
a868498c18 | ||
|
|
8dac90a593 | ||
|
|
d6e2013c5f | ||
|
|
651f0a8889 | ||
|
|
d03794ce7a | ||
|
|
9b32e6e2dc | ||
|
|
f4628c22a4 | ||
|
|
35ef8abc27 | ||
|
|
ad3bc0edee | ||
|
|
53d2a733b0 | ||
|
|
1703dc330f | ||
|
|
40a10c217d | ||
|
|
93966b0d19 | ||
|
|
a9458fd844 | ||
|
|
fa869eebd9 | ||
|
|
f421c26d35 | ||
|
|
16bca5d4a1 | ||
|
|
36e7c5364d | ||
|
|
20d1bba1bb | ||
|
|
d868126c39 | ||
|
|
570f8ae4ba | ||
|
|
bdb291f1c2 | ||
|
|
446d536c23 | ||
|
|
8fc60b31bc | ||
|
|
9ae547f994 | ||
|
|
03cc3b359c | ||
|
|
bcc8679a05 | ||
|
|
ff471b3fab | ||
|
|
d87f69215e | ||
|
|
2cba1d9fcc | ||
|
|
24c2373080 | ||
|
|
2c70751d1e | ||
|
|
546de5efd2 | ||
|
|
3970e4e6bb | ||
|
|
7785d65c8a | ||
|
|
4a4e5c7c18 | ||
|
|
1be09848a7 | ||
|
|
cb40bbdadd | ||
|
|
703261e78f | ||
|
|
ef11b024e8 | ||
|
|
5156be0f49 | ||
|
|
8bdea72688 | ||
|
|
f23cc92130 | ||
|
|
f801d3cf15 | ||
|
|
2365f82750 | ||
|
|
7a6b711eb8 | ||
|
|
4325178822 | ||
|
|
4083440690 | ||
|
|
e28f6f6657 | ||
|
|
579ab23b10 | ||
|
|
a33f35eecf | ||
|
|
91bb9e2cb3 | ||
|
|
0ce80b7bcf | ||
|
|
7a5586f3db | ||
|
|
8fccc3c4ad | ||
|
|
647d3844dd | ||
|
|
dae7a41baa | ||
|
|
8dd96013f1 | ||
|
|
36cf979b82 | ||
|
|
abaa593aa0 | ||
|
|
937352c78f | ||
|
|
176fec8789 | ||
|
|
8959622836 | ||
|
|
0bdaca25ca | ||
|
|
a5bc8e2c6a | ||
|
|
210eb471e9 | ||
|
|
701f32b227 | ||
|
|
794cbaa60a | ||
|
|
ff1c559084 | ||
|
|
8746f9cddf | ||
|
|
f24bfc7684 | ||
|
|
3af02584c1 | ||
|
|
bc4f802b17 | ||
|
|
2407381c3d | ||
|
|
e88d6e071d | ||
|
|
dcaf580476 | ||
|
|
6eb23353d7 | ||
|
|
f470ad3af9 | ||
|
|
45dc1f818a | ||
|
|
0725fd6081 | ||
|
|
a68580e2a7 | ||
|
|
f0c1b842bf | ||
|
|
142a208a90 | ||
|
|
e44a082620 | ||
|
|
e5ec546da5 | ||
|
|
4a87ea49b8 | ||
|
|
d285d6ba2a | ||
|
|
e58e417603 | ||
|
|
9b0eb66b78 | ||
|
|
637e42a0c0 | ||
|
|
bb47fd8c49 | ||
|
|
8f8bd8147a | ||
|
|
1a33b50a0d | ||
|
|
d48123d23b | ||
|
|
1ced638165 | ||
|
|
b90c6d25e8 | ||
|
|
78694405a6 | ||
|
|
6426449c8b | ||
|
|
31e6902e43 | ||
|
|
6b55150e80 | ||
|
|
13b15e07e8 | ||
|
|
d3429f2ff6 | ||
|
|
545fd4548e | ||
|
|
27c66f12d1 | ||
|
|
5a7dc41351 | ||
|
|
80339c3427 | ||
|
|
6275cdc486 | ||
|
|
cc6d57aa0d | ||
|
|
7a039e03fe | ||
|
|
fbc3d861bb | ||
|
|
47224dd6d3 | ||
|
|
bde4f25794 | ||
|
|
2070afea02 | ||
|
|
18cbebaeb9 | ||
|
|
755d9d4609 | ||
|
|
606be9e663 | ||
|
|
18a38f7ca0 | ||
|
|
d314680a15 | ||
|
|
474366c020 | ||
|
|
f6babc814c | ||
|
|
6f424d8d6c | ||
|
|
f93a727869 | ||
|
|
edf9a9608e | ||
|
|
765097d514 | ||
|
|
19775ffe15 | ||
|
|
5ef33adf68 | ||
|
|
806c92c80b | ||
|
|
71d3b2e036 | ||
|
|
b41cf92dc2 | ||
|
|
77d4a53c32 | ||
|
|
4fcfd9c96e | ||
|
|
1496789561 | ||
|
|
a06d53688c | ||
|
|
4cd14aee5a | ||
|
|
1baad8650c | ||
|
|
ee382c4153 | ||
|
|
af907e2d0d | ||
|
|
11e46e4bc0 | ||
|
|
1b6538b4e5 | ||
|
|
c8f9d4b6e6 | ||
|
|
9fa7ed1743 | ||
|
|
11d65fcb21 | ||
|
|
db80671d6b | ||
|
|
94ca52b7b7 | ||
|
|
8ba4722d04 | ||
|
|
65e6d73b95 | ||
|
|
16ba74d008 | ||
|
|
d2bc0f0f08 | ||
|
|
1823db53f2 | ||
|
|
7ef54e39ec | ||
|
|
14ef38b834 | ||
|
|
46e0bce212 | ||
|
|
8ab5e13b5d | ||
|
|
317d7be6ee | ||
|
|
90cce38236 | ||
|
|
50d854e02e | ||
|
|
1b103e1f5f | ||
|
|
288c52596c | ||
|
|
f7db16add1 | ||
|
|
a94e1b172e | ||
|
|
686caad40c | ||
|
|
fdf533f2b9 | ||
|
|
ad06172c6b | ||
|
|
bef1f939ce | ||
|
|
dc2e699656 | ||
|
|
6ece549a90 | ||
|
|
332380479b | ||
|
|
c45665a55a | ||
|
|
f0f76259c9 | ||
|
|
c13a2a3114 | ||
|
|
c70fa502a5 | ||
|
|
401d451569 | ||
|
|
52d4eda786 | ||
|
|
5815df4c46 | ||
|
|
d0de954af2 | ||
|
|
5dea21273a | ||
|
|
2d83b2ad8f | ||
|
|
6f032b7152 | ||
|
|
1bb1913811 | ||
|
|
56e4baff7c | ||
|
|
3f536b5308 | ||
|
|
4b00c64d96 | ||
|
|
118192f116 | ||
|
|
729d227b89 | ||
|
|
89d6419fd5 | ||
|
|
2454407f3a | ||
|
|
e5ab8f3ebe | ||
|
|
bcce17e688 | ||
|
|
c467e90ac1 | ||
|
|
fd78af404b | ||
|
|
02dd7b6913 | ||
|
|
e8eff3581b | ||
|
|
6fa1afdffc | ||
|
|
522636cb52 | ||
|
|
9150fdbd4d | ||
|
|
d4796482b5 | ||
|
|
a50b84244e | ||
|
|
3c9b04460a | ||
|
|
8b3ecfca25 | ||
|
|
39c5616af2 | ||
|
|
b37ff3d492 | ||
|
|
b3ba0e8708 | ||
|
|
af0cf88921 | ||
|
|
64575591d8 | ||
|
|
f5b20286e2 | ||
|
|
c92ab2ce49 | ||
|
|
66cb4afc6c | ||
|
|
f20ffa8db3 | ||
|
|
4d81c741e9 | ||
|
|
cd55823112 | ||
|
|
996cc705af | ||
|
|
718472dbe2 | ||
|
|
9a400731d9 | ||
|
|
da351621a1 | ||
|
|
e78a38b837 | ||
|
|
613ec36c5a | ||
|
|
6b6849b001 | ||
|
|
98d6faefd6 | ||
|
|
4dafb5fac8 | ||
|
|
04fc575c0e | ||
|
|
55463b76c1 | ||
|
|
a62a3d991d | ||
|
|
68b6d6bbe2 | ||
|
|
a92e0f6240 | ||
|
|
1d468e20a4 | ||
|
|
18a4af63aa | ||
|
|
5eed2990ad | ||
|
|
1b25dd59f9 | ||
|
|
83a66b4994 | ||
|
|
f60d95b0ba | ||
|
|
89aa8ddf52 | ||
|
|
6762c45494 | ||
|
|
49c74a5369 | ||
|
|
e3e6de5ed9 | ||
|
|
c859764d29 | ||
|
|
584bae1fc6 | ||
|
|
e56d1779e1 | ||
|
|
549f3bd781 | ||
|
|
671b3c8d8e | ||
|
|
b2341eab0c | ||
|
|
131858e7cb | ||
|
|
f08c5dcb06 | ||
|
|
b76c5d54bf | ||
|
|
7366d3b20c | ||
|
|
14d61b0141 | ||
|
|
d625dc2047 | ||
|
|
3877043d41 | ||
|
|
711f7f3851 | ||
|
|
12949c6b31 | ||
|
|
0149f81a5a | ||
|
|
93eebe8664 | ||
|
|
0da7d872ef | ||
|
|
0d0abe1845 | ||
|
|
2369d55e9a | ||
|
|
5cd1f71b51 | ||
|
|
b52c4e13b0 | ||
|
|
1f020a6097 | ||
|
|
1441a6cd27 | ||
|
|
2775c2a1ab | ||
|
|
87c01f49d8 | ||
|
|
fe4ce920b2 | ||
|
|
926af9951e | ||
|
|
3e693e4f97 | ||
|
|
d152c59a9c | ||
|
|
34a238ca98 | ||
|
|
f6e6d0b2c0 | ||
|
|
271a7c5d43 | ||
|
|
c621775f34 | ||
|
|
b4340abf56 | ||
|
|
7f738e7f6f | ||
|
|
549bd419bb | ||
|
|
f21301c749 | ||
|
|
81210420c6 | ||
|
|
028bdc1740 | ||
|
|
e060519d4f | ||
|
|
5d7818e75d | ||
|
|
24789429fd | ||
|
|
511805c981 | ||
|
|
6967ef7267 | ||
|
|
5817840858 | ||
|
|
f84291c1e1 | ||
|
|
d262503781 | ||
|
|
ef4dae4c0e | ||
|
|
5ddd4a9d06 | ||
|
|
9fd510faa5 | ||
|
|
529cf8a54a | ||
|
|
ac7a36367c | ||
|
|
b4ec1682c6 | ||
|
|
dac9eb13bd | ||
|
|
9f20a3315e | ||
|
|
bb56bb9a13 | ||
|
|
cc06fab9a7 | ||
|
|
deab0e32ba | ||
|
|
d6ea5cc1ed | ||
|
|
465dc63833 | ||
|
|
a1bcd33a3b | ||
|
|
13b0fa4b97 | ||
|
|
52277cc3da | ||
|
|
e94b766310 | ||
|
|
e5e47c3c99 | ||
|
|
20c0d60ac7 | ||
|
|
2db808021d | ||
|
|
c635d4c46a | ||
|
|
001503186c | ||
|
|
ed95e77752 | ||
|
|
91c1a1c52f | ||
|
|
0df2ae63c7 | ||
|
|
28af6f9abb | ||
|
|
eea094e1bc | ||
|
|
ec56d5869b | ||
|
|
54582f641a | ||
|
|
eb1efb54b5 | ||
|
|
8f0a42a266 | ||
|
|
68cdbc9c16 | ||
|
|
5516281881 | ||
|
|
a4a0ebb85d | ||
|
|
3886c3dd8f | ||
|
|
29bfa94bb6 | ||
|
|
7f399eac8b | ||
|
|
eabec370e4 | ||
|
|
e8c1eb99e4 | ||
|
|
59bd1ab17e | ||
|
|
58a6723eb1 | ||
|
|
9ab73f737e | ||
|
|
b4a1236cfc | ||
|
|
24e25802a7 | ||
|
|
5b1161bb64 | ||
|
|
6f8a4633b7 | ||
|
|
70b12d898a | ||
|
|
a512b4b394 | ||
|
|
05497a9141 | ||
|
|
01152f89ee | ||
|
|
1864fab592 | ||
|
|
021f8bf28b | ||
|
|
eee527d264 | ||
|
|
85cbd32c5a | ||
|
|
6f38f5affa | ||
|
|
557ffc4bf5 | ||
|
|
bf7bb575b4 | ||
|
|
24be04e848 | ||
|
|
619c450a49 | ||
|
|
820e1c01ef | ||
|
|
488f12a996 | ||
|
|
c024c42dce | ||
|
|
96a9848c9e | ||
|
|
b124a27f57 | ||
|
|
5262e933f7 | ||
|
|
0c67685e43 | ||
|
|
0552ca8021 | ||
|
|
176110a22d | ||
|
|
d33854af1b | ||
|
|
9fb4338964 | ||
|
|
7cfb310eb4 | ||
|
|
2adf222fb2 | ||
|
|
b0015fda96 | ||
|
|
55ee272ea8 | ||
|
|
e27f543deb | ||
|
|
d4274bc556 | ||
|
|
a7057fa64c | ||
|
|
d27a11ff87 | ||
|
|
8cc75f1576 | ||
|
|
38ca96c9fc | ||
|
|
46726ec176 | ||
|
|
937fa282b5 | ||
|
|
ca6f980932 | ||
|
|
97d7582457 | ||
|
|
cb685607b2 | ||
|
|
8df0a252b7 | ||
|
|
d7d1b6e3a6 | ||
|
|
6ede12412c | ||
|
|
c968217ca8 | ||
|
|
48aff0eabd | ||
|
|
b06040b6d0 | ||
|
|
232144ca09 | ||
|
|
4100827971 | ||
|
|
ccdabe4512 | ||
|
|
57a4b4ff64 | ||
|
|
e6ab594e14 | ||
|
|
a55d43ccfd | ||
|
|
c74df31bf9 | ||
|
|
154b15060e | ||
|
|
32e673d8c4 | ||
|
|
a13321148a | ||
|
|
6295dc3b67 | ||
|
|
8211e5f341 | ||
|
|
0f7a9b42f1 | ||
|
|
239dbb3c0a | ||
|
|
b81ebbef62 | ||
|
|
c6769488b3 | ||
|
|
45aef75cca | ||
|
|
6b074add66 | ||
|
|
3c4aa9b2ea | ||
|
|
d05754f558 | ||
|
|
ac9bfaa4f2 | ||
|
|
2eee87423c | ||
|
|
b9414b6477 | ||
|
|
d4349426d8 | ||
|
|
a6bcd54b47 | ||
|
|
fd61c61071 | ||
|
|
e36b066344 | ||
|
|
864d236a82 | ||
|
|
31a307cf6b | ||
|
|
7593fa9982 | ||
|
|
d1f00fb0b7 | ||
|
|
74bab6e504 | ||
|
|
8d7c6366d7 | ||
|
|
15685996fc | ||
|
|
f999897615 | ||
|
|
b8e8f0fcd9 | ||
|
|
3b0b74fa94 | ||
|
|
376b448015 | ||
|
|
f53da412aa | ||
|
|
5ff210ed75 | ||
|
|
c42e3fbcf3 | ||
|
|
4fd149b3a2 | ||
|
|
fbb0dc4275 | ||
|
|
fb1a9e6bc5 | ||
|
|
f56e2e9a66 | ||
|
|
8e619010d0 | ||
|
|
6cdcfe8128 | ||
|
|
e6a142fe70 | ||
|
|
4ddf8d001c | ||
|
|
406c70ba0e | ||
|
|
0bd8f21e4e | ||
|
|
a7d0c66457 | ||
|
|
130df8cdda | ||
|
|
5b17bb0031 | ||
|
|
89d87e5331 | ||
|
|
298af6f409 | ||
|
|
69d3b1b8b4 | ||
|
|
578de9f762 | ||
|
|
f7caac2563 | ||
|
|
e2660ab8f3 | ||
|
|
cc459755be | ||
|
|
d8cb395380 | ||
|
|
b3b03200e2 | ||
|
|
2a0368b7ca | ||
|
|
b2d8431aea | ||
|
|
d8a549e6ac | ||
|
|
ca17f8a5fc | ||
|
|
fbd58bf190 | ||
|
|
0ee11dac77 | ||
|
|
d27a427dc5 | ||
|
|
475fd1abec | ||
|
|
9472be7d77 | ||
|
|
4f93e5586a | ||
|
|
18bd16341a | ||
|
|
61a619b5c3 | ||
|
|
e48e05e6e2 | ||
|
|
c735c17f33 | ||
|
|
c311a8c1d8 | ||
|
|
b18f5f61b0 | ||
|
|
38a23f66a8 | ||
|
|
8ad7e8eeb0 | ||
|
|
22d56cebf1 | ||
|
|
32e0858501 | ||
|
|
31c1e13f90 | ||
|
|
9f63d6fead | ||
|
|
0ed979b096 | ||
|
|
2942dc68e4 | ||
|
|
037dd0820d | ||
|
|
d997c967d5 | ||
|
|
3515931305 | ||
|
|
a0dcf6f5c1 | ||
|
|
804b2ac60f | ||
|
|
68a2c7b8d6 | ||
|
|
b12e7f7edd | ||
|
|
3a4f51f39f | ||
|
|
ba69244a94 | ||
|
|
7a1d67f9cb | ||
|
|
e7d7ab6bc3 | ||
|
|
b70e07da1f | ||
|
|
cdfaa705f3 | ||
|
|
3060f0b562 | ||
|
|
9c64618cb6 | ||
|
|
d04312b9c0 | ||
|
|
ee8d1f5ed8 | ||
|
|
3290a4f3ed | ||
|
|
bf562bd33c | ||
|
|
01b7acba30 | ||
|
|
ec849ec335 | ||
|
|
46c46829ce | ||
|
|
6bcbc77226 | ||
|
|
3f38d983a6 | ||
|
|
9600ca83f3 | ||
|
|
149f209af6 | ||
|
|
336af4f974 | ||
|
|
f7003dc819 | ||
|
|
8a84be37b8 | ||
|
|
8ee127469f | ||
|
|
ba47eda61b | ||
|
|
e2c406f5c8 | ||
|
|
7bdedacb54 | ||
|
|
d080b5a953 | ||
|
|
36346f8f56 | ||
|
|
1369133916 | ||
|
|
f1a4a1ac95 | ||
|
|
dfdf0b08fc | ||
|
|
92ae3abc97 | ||
|
|
1a75f43304 | ||
|
|
7017dd5a26 | ||
|
|
48d5de80a2 | ||
|
|
7ee7a95b84 | ||
|
|
e88ac9cc54 | ||
|
|
778135f657 | ||
|
|
41e882f80b | ||
|
|
caa9e527dd | ||
|
|
e6088366df | ||
|
|
e64ee6592f | ||
|
|
9f7f8b976d | ||
|
|
d7c14496d2 | ||
|
|
bd1f3a38f0 | ||
|
|
2f524e9f41 | ||
|
|
abec3dbf6d | ||
|
|
2801d69fb7 | ||
|
|
8e8232fb4c | ||
|
|
345796825f | ||
|
|
1d91f71119 | ||
|
|
77f6cf2d13 | ||
|
|
84d359efb8 | ||
|
|
5d7cdf2e36 | ||
|
|
c766f143ab | ||
|
|
689eb8f620 | ||
|
|
615ab2b03e | ||
|
|
d22b293f2f | ||
|
|
f937f514aa | ||
|
|
116d711815 | ||
|
|
d7e1fa7664 | ||
|
|
ffa66aace0 | ||
|
|
b56d3d5d5c | ||
|
|
93f3acdef9 | ||
|
|
a5d222fcdb | ||
|
|
1cd20efe68 | ||
|
|
1c8fdf2218 | ||
|
|
dd4db347f3 | ||
|
|
8fa32fdda2 | ||
|
|
663136aa08 | ||
|
|
b2d300e727 | ||
|
|
1d4d345634 | ||
|
|
da1ad798ca | ||
|
|
bbfffb444d | ||
|
|
29f8fd6fee | ||
|
|
7968c0d051 | ||
|
|
86715e4cd4 | ||
|
|
7dd29ffd47 | ||
|
|
dcd84b3979 | ||
|
|
d9799b09d0 | ||
|
|
5c2d7a18c9 | ||
|
|
2567404ab6 | ||
|
|
1faad825f4 | ||
|
|
b56614e9b8 | ||
|
|
25514e104a | ||
|
|
f79cc4a7a4 | ||
|
|
72f9daf9b6 | ||
|
|
bd2ca543c4 | ||
|
|
7beb2f7fae | ||
|
|
c4b9f4f622 | ||
|
|
655e6992f6 | ||
|
|
5cdaac00c1 | ||
|
|
05db6a6c29 | ||
|
|
57c732655e | ||
|
|
ee4f51a631 | ||
|
|
816b789bf0 | ||
|
|
55b823b27d | ||
|
|
89a49cf30e | ||
|
|
4cf95a6041 | ||
|
|
ab6fd304c4 | ||
|
|
29d6a5e2b8 | ||
|
|
86e60e3ba8 | ||
|
|
c6d87e5e18 | ||
|
|
a4bc7ecf27 | ||
|
|
6e52aefb37 | ||
|
|
cf06a266a8 | ||
|
|
81bdfb835d | ||
|
|
29c942f2a8 | ||
|
|
2320aa0da2 | ||
|
|
5cb51a191e | ||
|
|
7e846bb965 | ||
|
|
6e104f0570 | ||
|
|
42fc7ca6a0 | ||
|
|
a4886c404a | ||
|
|
f94f479358 | ||
|
|
d245bc891e | ||
|
|
894e9bc5d4 | ||
|
|
44cc9c04ea | ||
|
|
05ac415780 | ||
|
|
3e7e426b36 | ||
|
|
2a9979e256 | ||
|
|
90cd724be1 | ||
|
|
e41619b1fc | ||
|
|
ec6ce08cd0 | ||
|
|
4ddbaeea32 | ||
|
|
b35dd76dca | ||
|
|
37ad60fe25 | ||
|
|
a1d23f6613 | ||
|
|
45ddc39c1d | ||
|
|
34df1f588b | ||
|
|
b31d37eac5 | ||
|
|
db6285fb55 | ||
|
|
4e1a8b1fe5 | ||
|
|
5472ef626c | ||
|
|
20f34d9776 | ||
|
|
ef473b1f09 | ||
|
|
8760ec4827 | ||
|
|
54afa3ac7a | ||
|
|
a2ecbdaa31 | ||
|
|
896aede340 | ||
|
|
74b41637de | ||
|
|
c8cc3eacc9 | ||
|
|
2828da3c4c | ||
|
|
233bdf105f | ||
|
|
71b938f608 | ||
|
|
146549260a | ||
|
|
bec2b4f094 | ||
|
|
2c684ffd32 | ||
|
|
556a83022d | ||
|
|
1b26a2a561 | ||
|
|
a5d7094a45 | ||
|
|
d31a57cf5f | ||
|
|
bccb7e87d1 | ||
|
|
2e8c101b4a | ||
|
|
4224c08cac | ||
|
|
878b990fcd | ||
|
|
dee5ef2dfd | ||
|
|
3d919db0c0 | ||
|
|
b9a4f3336a | ||
|
|
ea7a6a0321 | ||
|
|
f294c4e023 | ||
|
|
b65e3c4444 | ||
|
|
7bcc8b3e5c | ||
|
|
aa0d8f20c1 | ||
|
|
7e06c81894 | ||
|
|
0cced530ea | ||
|
|
b1fdb220f4 | ||
|
|
74f3a2f4b5 | ||
|
|
47b62480af | ||
|
|
a5c852660b | ||
|
|
905fdd3e08 | ||
|
|
ca998df912 | ||
|
|
3039dd194b | ||
|
|
10ae0f9511 | ||
|
|
138fe8516a | ||
|
|
79b8b560d2 | ||
|
|
4aa12e10c0 | ||
|
|
f01af43eb0 |
214
.clang-format
Normal file
214
.clang-format
Normal file
@@ -0,0 +1,214 @@
|
|||||||
|
---
|
||||||
|
Language: Cpp
|
||||||
|
# BasedOnStyle: Google
|
||||||
|
AccessModifierOffset: -1
|
||||||
|
AlignAfterOpenBracket: Align
|
||||||
|
AlignArrayOfStructures: None
|
||||||
|
AlignConsecutiveMacros: None
|
||||||
|
AlignConsecutiveAssignments: None
|
||||||
|
AlignConsecutiveBitFields: None
|
||||||
|
AlignConsecutiveDeclarations: None
|
||||||
|
AlignEscapedNewlines: Left
|
||||||
|
AlignOperands: Align
|
||||||
|
AlignTrailingComments: true
|
||||||
|
AllowAllArgumentsOnNextLine: true
|
||||||
|
AllowAllParametersOfDeclarationOnNextLine: true
|
||||||
|
AllowShortEnumsOnASingleLine: true
|
||||||
|
AllowShortBlocksOnASingleLine: Never
|
||||||
|
AllowShortCaseLabelsOnASingleLine: false
|
||||||
|
AllowShortFunctionsOnASingleLine: All
|
||||||
|
AllowShortLambdasOnASingleLine: All
|
||||||
|
AllowShortIfStatementsOnASingleLine: WithoutElse
|
||||||
|
AllowShortLoopsOnASingleLine: true
|
||||||
|
AlwaysBreakAfterDefinitionReturnType: None
|
||||||
|
AlwaysBreakAfterReturnType: None
|
||||||
|
AlwaysBreakBeforeMultilineStrings: true
|
||||||
|
AlwaysBreakTemplateDeclarations: Yes
|
||||||
|
AttributeMacros:
|
||||||
|
- __capability
|
||||||
|
BinPackArguments: true
|
||||||
|
BinPackParameters: true
|
||||||
|
BraceWrapping:
|
||||||
|
AfterCaseLabel: false
|
||||||
|
AfterClass: false
|
||||||
|
AfterControlStatement: Never
|
||||||
|
AfterEnum: false
|
||||||
|
AfterFunction: false
|
||||||
|
AfterNamespace: false
|
||||||
|
AfterObjCDeclaration: false
|
||||||
|
AfterStruct: false
|
||||||
|
AfterUnion: false
|
||||||
|
AfterExternBlock: false
|
||||||
|
BeforeCatch: false
|
||||||
|
BeforeElse: false
|
||||||
|
BeforeLambdaBody: false
|
||||||
|
BeforeWhile: false
|
||||||
|
IndentBraces: false
|
||||||
|
SplitEmptyFunction: true
|
||||||
|
SplitEmptyRecord: true
|
||||||
|
SplitEmptyNamespace: true
|
||||||
|
BreakBeforeBinaryOperators: None
|
||||||
|
BreakBeforeConceptDeclarations: true
|
||||||
|
BreakBeforeBraces: Attach
|
||||||
|
BreakBeforeInheritanceComma: false
|
||||||
|
BreakInheritanceList: BeforeColon
|
||||||
|
BreakBeforeTernaryOperators: true
|
||||||
|
BreakConstructorInitializersBeforeComma: false
|
||||||
|
BreakConstructorInitializers: BeforeColon
|
||||||
|
BreakAfterJavaFieldAnnotations: false
|
||||||
|
BreakStringLiterals: true
|
||||||
|
ColumnLimit: 100
|
||||||
|
CommentPragmas: '^ IWYU pragma:'
|
||||||
|
QualifierAlignment: Leave
|
||||||
|
CompactNamespaces: false
|
||||||
|
ConstructorInitializerIndentWidth: 4
|
||||||
|
ContinuationIndentWidth: 4
|
||||||
|
Cpp11BracedListStyle: true
|
||||||
|
DeriveLineEnding: true
|
||||||
|
DerivePointerAlignment: true
|
||||||
|
DisableFormat: false
|
||||||
|
EmptyLineAfterAccessModifier: Never
|
||||||
|
EmptyLineBeforeAccessModifier: LogicalBlock
|
||||||
|
ExperimentalAutoDetectBinPacking: false
|
||||||
|
PackConstructorInitializers: NextLine
|
||||||
|
BasedOnStyle: ''
|
||||||
|
ConstructorInitializerAllOnOneLineOrOnePerLine: false
|
||||||
|
AllowAllConstructorInitializersOnNextLine: true
|
||||||
|
FixNamespaceComments: true
|
||||||
|
ForEachMacros:
|
||||||
|
- foreach
|
||||||
|
- Q_FOREACH
|
||||||
|
- BOOST_FOREACH
|
||||||
|
IfMacros:
|
||||||
|
- KJ_IF_MAYBE
|
||||||
|
IncludeBlocks: Regroup
|
||||||
|
IncludeCategories:
|
||||||
|
- Regex: '^<ext/.*\.h>'
|
||||||
|
Priority: 2
|
||||||
|
SortPriority: 0
|
||||||
|
CaseSensitive: false
|
||||||
|
- Regex: '^<.*\.h>'
|
||||||
|
Priority: 1
|
||||||
|
SortPriority: 0
|
||||||
|
CaseSensitive: false
|
||||||
|
- Regex: '^<.*'
|
||||||
|
Priority: 2
|
||||||
|
SortPriority: 0
|
||||||
|
CaseSensitive: false
|
||||||
|
- Regex: '.*'
|
||||||
|
Priority: 3
|
||||||
|
SortPriority: 0
|
||||||
|
CaseSensitive: false
|
||||||
|
IncludeIsMainRegex: '([-_](test|unittest))?$'
|
||||||
|
IncludeIsMainSourceRegex: ''
|
||||||
|
IndentAccessModifiers: false
|
||||||
|
IndentCaseLabels: true
|
||||||
|
IndentCaseBlocks: false
|
||||||
|
IndentGotoLabels: true
|
||||||
|
IndentPPDirectives: None
|
||||||
|
IndentExternBlock: AfterExternBlock
|
||||||
|
IndentRequires: false
|
||||||
|
IndentWidth: 2
|
||||||
|
IndentWrappedFunctionNames: false
|
||||||
|
InsertTrailingCommas: None
|
||||||
|
JavaScriptQuotes: Leave
|
||||||
|
JavaScriptWrapImports: true
|
||||||
|
KeepEmptyLinesAtTheStartOfBlocks: false
|
||||||
|
LambdaBodyIndentation: Signature
|
||||||
|
MacroBlockBegin: ''
|
||||||
|
MacroBlockEnd: ''
|
||||||
|
MaxEmptyLinesToKeep: 1
|
||||||
|
NamespaceIndentation: None
|
||||||
|
ObjCBinPackProtocolList: Never
|
||||||
|
ObjCBlockIndentWidth: 2
|
||||||
|
ObjCBreakBeforeNestedBlockParam: true
|
||||||
|
ObjCSpaceAfterProperty: false
|
||||||
|
ObjCSpaceBeforeProtocolList: true
|
||||||
|
PenaltyBreakAssignment: 2
|
||||||
|
PenaltyBreakBeforeFirstCallParameter: 1
|
||||||
|
PenaltyBreakComment: 300
|
||||||
|
PenaltyBreakFirstLessLess: 120
|
||||||
|
PenaltyBreakString: 1000
|
||||||
|
PenaltyBreakTemplateDeclaration: 10
|
||||||
|
PenaltyExcessCharacter: 1000000
|
||||||
|
PenaltyReturnTypeOnItsOwnLine: 200
|
||||||
|
PenaltyIndentedWhitespace: 0
|
||||||
|
PointerAlignment: Left
|
||||||
|
PPIndentWidth: -1
|
||||||
|
RawStringFormats:
|
||||||
|
- Language: Cpp
|
||||||
|
Delimiters:
|
||||||
|
- cc
|
||||||
|
- CC
|
||||||
|
- cpp
|
||||||
|
- Cpp
|
||||||
|
- CPP
|
||||||
|
- 'c++'
|
||||||
|
- 'C++'
|
||||||
|
CanonicalDelimiter: ''
|
||||||
|
BasedOnStyle: google
|
||||||
|
- Language: TextProto
|
||||||
|
Delimiters:
|
||||||
|
- pb
|
||||||
|
- PB
|
||||||
|
- proto
|
||||||
|
- PROTO
|
||||||
|
EnclosingFunctions:
|
||||||
|
- EqualsProto
|
||||||
|
- EquivToProto
|
||||||
|
- PARSE_PARTIAL_TEXT_PROTO
|
||||||
|
- PARSE_TEST_PROTO
|
||||||
|
- PARSE_TEXT_PROTO
|
||||||
|
- ParseTextOrDie
|
||||||
|
- ParseTextProtoOrDie
|
||||||
|
- ParseTestProto
|
||||||
|
- ParsePartialTestProto
|
||||||
|
CanonicalDelimiter: pb
|
||||||
|
BasedOnStyle: google
|
||||||
|
ReferenceAlignment: Pointer
|
||||||
|
ReflowComments: true
|
||||||
|
ShortNamespaceLines: 1
|
||||||
|
SortIncludes: CaseSensitive
|
||||||
|
SortJavaStaticImport: Before
|
||||||
|
SortUsingDeclarations: true
|
||||||
|
SpaceAfterCStyleCast: false
|
||||||
|
SpaceAfterLogicalNot: false
|
||||||
|
SpaceAfterTemplateKeyword: true
|
||||||
|
SpaceBeforeAssignmentOperators: true
|
||||||
|
SpaceBeforeCaseColon: false
|
||||||
|
SpaceBeforeCpp11BracedList: false
|
||||||
|
SpaceBeforeCtorInitializerColon: true
|
||||||
|
SpaceBeforeInheritanceColon: true
|
||||||
|
SpaceBeforeParens: ControlStatements
|
||||||
|
SpaceAroundPointerQualifiers: Default
|
||||||
|
SpaceBeforeRangeBasedForLoopColon: true
|
||||||
|
SpaceInEmptyBlock: false
|
||||||
|
SpaceInEmptyParentheses: false
|
||||||
|
SpacesBeforeTrailingComments: 2
|
||||||
|
SpacesInAngles: Never
|
||||||
|
SpacesInConditionalStatement: false
|
||||||
|
SpacesInContainerLiterals: true
|
||||||
|
SpacesInCStyleCastParentheses: false
|
||||||
|
SpacesInLineCommentPrefix:
|
||||||
|
Minimum: 1
|
||||||
|
Maximum: -1
|
||||||
|
SpacesInParentheses: false
|
||||||
|
SpacesInSquareBrackets: false
|
||||||
|
SpaceBeforeSquareBrackets: false
|
||||||
|
BitFieldColonSpacing: Both
|
||||||
|
Standard: Auto
|
||||||
|
StatementAttributeLikeMacros:
|
||||||
|
- Q_EMIT
|
||||||
|
StatementMacros:
|
||||||
|
- Q_UNUSED
|
||||||
|
- QT_REQUIRE_VERSION
|
||||||
|
TabWidth: 8
|
||||||
|
UseCRLF: false
|
||||||
|
UseTab: Never
|
||||||
|
WhitespaceSensitiveMacros:
|
||||||
|
- STRINGIZE
|
||||||
|
- PP_STRINGIZE
|
||||||
|
- BOOST_PP_STRINGIZE
|
||||||
|
- NS_SWIFT_NAME
|
||||||
|
- CF_SWIFT_NAME
|
||||||
|
...
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
Checks: 'modernize-*,-modernize-make-*,-modernize-use-auto,-modernize-raw-string-literal,-modernize-avoid-c-arrays,-modernize-use-trailing-return-type,google-*,-google-default-arguments,-clang-diagnostic-#pragma-messages,readability-identifier-naming'
|
Checks: 'modernize-*,-modernize-use-nodiscard,-modernize-concat-nested-namespaces,-modernize-make-*,-modernize-use-auto,-modernize-raw-string-literal,-modernize-avoid-c-arrays,-modernize-use-trailing-return-type,google-*,-google-default-arguments,-clang-diagnostic-#pragma-messages,readability-identifier-naming'
|
||||||
CheckOptions:
|
CheckOptions:
|
||||||
- { key: readability-identifier-naming.ClassCase, value: CamelCase }
|
- { key: readability-identifier-naming.ClassCase, value: CamelCase }
|
||||||
- { key: readability-identifier-naming.StructCase, value: CamelCase }
|
- { key: readability-identifier-naming.StructCase, value: CamelCase }
|
||||||
|
|||||||
18
.gitattributes
vendored
Normal file
18
.gitattributes
vendored
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
* text=auto
|
||||||
|
|
||||||
|
*.c text eol=lf
|
||||||
|
*.h text eol=lf
|
||||||
|
*.cc text eol=lf
|
||||||
|
*.cuh text eol=lf
|
||||||
|
*.cu text eol=lf
|
||||||
|
*.py text eol=lf
|
||||||
|
*.txt text eol=lf
|
||||||
|
*.R text eol=lf
|
||||||
|
*.scala text eol=lf
|
||||||
|
*.java text eol=lf
|
||||||
|
|
||||||
|
*.sh text eol=lf
|
||||||
|
|
||||||
|
*.rst text eol=lf
|
||||||
|
*.md text eol=lf
|
||||||
|
*.csv text eol=lf
|
||||||
31
.github/dependabot.yml
vendored
Normal file
31
.github/dependabot.yml
vendored
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
# To get started with Dependabot version updates, you'll need to specify which
|
||||||
|
# package ecosystems to update and where the package manifests are located.
|
||||||
|
# Please see the documentation for all configuration options:
|
||||||
|
# https://docs.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
|
||||||
|
|
||||||
|
version: 2
|
||||||
|
updates:
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages/xgboost4j"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages/xgboost4j-gpu"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages/xgboost4j-example"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages/xgboost4j-spark"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
|
- package-ecosystem: "maven"
|
||||||
|
directory: "/jvm-packages/xgboost4j-spark-gpu"
|
||||||
|
schedule:
|
||||||
|
interval: "daily"
|
||||||
87
.github/workflows/jvm_tests.yml
vendored
Normal file
87
.github/workflows/jvm_tests.yml
vendored
Normal file
@@ -0,0 +1,87 @@
|
|||||||
|
name: XGBoost-JVM-Tests
|
||||||
|
|
||||||
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test-with-jvm:
|
||||||
|
name: Test JVM on OS ${{ matrix.os }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
os: [windows-latest, ubuntu-latest, macos-11]
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
|
||||||
|
- uses: actions/setup-python@7f80679172b057fc5e90d70d197929d454754a5a # v4.3.0
|
||||||
|
with:
|
||||||
|
python-version: '3.8'
|
||||||
|
architecture: 'x64'
|
||||||
|
|
||||||
|
- uses: actions/setup-java@d202f5dbf7256730fb690ec59f6381650114feb2 # v3.6.0
|
||||||
|
with:
|
||||||
|
java-version: 1.8
|
||||||
|
|
||||||
|
- name: Install Python packages
|
||||||
|
run: |
|
||||||
|
python -m pip install wheel setuptools
|
||||||
|
python -m pip install awscli
|
||||||
|
|
||||||
|
- name: Cache Maven packages
|
||||||
|
uses: actions/cache@6998d139ddd3e68c71e9e398d8e40b71a2f39812 # v3.2.5
|
||||||
|
with:
|
||||||
|
path: ~/.m2
|
||||||
|
key: ${{ runner.os }}-m2-${{ hashFiles('./jvm-packages/pom.xml') }}
|
||||||
|
restore-keys: ${{ runner.os }}-m2-${{ hashFiles('./jvm-packages/pom.xml') }}
|
||||||
|
|
||||||
|
- name: Test XGBoost4J (Core)
|
||||||
|
run: |
|
||||||
|
cd jvm-packages
|
||||||
|
mvn test -B -pl :xgboost4j_2.12
|
||||||
|
|
||||||
|
- name: Extract branch name
|
||||||
|
shell: bash
|
||||||
|
run: echo "##[set-output name=branch;]$(echo ${GITHUB_REF#refs/heads/})"
|
||||||
|
id: extract_branch
|
||||||
|
if: |
|
||||||
|
(github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')) &&
|
||||||
|
matrix.os == 'windows-latest'
|
||||||
|
|
||||||
|
- name: Publish artifact xgboost4j.dll to S3
|
||||||
|
run: |
|
||||||
|
cd lib/
|
||||||
|
Rename-Item -Path xgboost4j.dll -NewName xgboost4j_${{ github.sha }}.dll
|
||||||
|
dir
|
||||||
|
python -m awscli s3 cp xgboost4j_${{ github.sha }}.dll s3://xgboost-nightly-builds/${{ steps.extract_branch.outputs.branch }}/ --acl public-read
|
||||||
|
if: |
|
||||||
|
(github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')) &&
|
||||||
|
matrix.os == 'windows-latest'
|
||||||
|
env:
|
||||||
|
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID_IAM_S3_UPLOADER }}
|
||||||
|
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY_IAM_S3_UPLOADER }}
|
||||||
|
|
||||||
|
|
||||||
|
- name: Test XGBoost4J (Core, Spark, Examples)
|
||||||
|
run: |
|
||||||
|
rm -rfv build/
|
||||||
|
cd jvm-packages
|
||||||
|
mvn -B test
|
||||||
|
if: matrix.os == 'ubuntu-latest' # Distributed training doesn't work on Windows
|
||||||
|
env:
|
||||||
|
RABIT_MOCK: ON
|
||||||
|
|
||||||
|
|
||||||
|
- name: Build and Test XGBoost4J with scala 2.13
|
||||||
|
run: |
|
||||||
|
rm -rfv build/
|
||||||
|
cd jvm-packages
|
||||||
|
mvn -B clean install test -Pdefault,scala-2.13
|
||||||
|
if: matrix.os == 'ubuntu-latest' # Distributed training doesn't work on Windows
|
||||||
|
env:
|
||||||
|
RABIT_MOCK: ON
|
||||||
277
.github/workflows/main.yml
vendored
277
.github/workflows/main.yml
vendored
@@ -6,6 +6,9 @@ name: XGBoost-CI
|
|||||||
# events but only for the master branch
|
# events but only for the master branch
|
||||||
on: [push, pull_request]
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
|
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
|
||||||
jobs:
|
jobs:
|
||||||
gtest-cpu:
|
gtest-cpu:
|
||||||
@@ -14,24 +17,25 @@ jobs:
|
|||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
os: [macos-10.15]
|
os: [macos-11]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
- name: Install system packages
|
- name: Install system packages
|
||||||
run: |
|
run: |
|
||||||
brew install lz4 ninja libomp
|
brew install ninja libomp
|
||||||
- name: Build gtest binary
|
- name: Build gtest binary
|
||||||
run: |
|
run: |
|
||||||
mkdir build
|
mkdir build
|
||||||
cd build
|
cd build
|
||||||
cmake .. -DGOOGLE_TEST=ON -DUSE_OPENMP=ON -DUSE_DMLC_GTEST=ON -DPLUGIN_LZ4=ON -DPLUGIN_DENSE_PARSER=ON -GNinja
|
cmake .. -DGOOGLE_TEST=ON -DUSE_OPENMP=ON -DUSE_DMLC_GTEST=ON -DPLUGIN_DENSE_PARSER=ON -GNinja
|
||||||
ninja -v
|
ninja -v
|
||||||
- name: Run gtest binary
|
- name: Run gtest binary
|
||||||
run: |
|
run: |
|
||||||
cd build
|
cd build
|
||||||
ctest --exclude-regex AllTestsInDMLCUnitTests --extra-verbose
|
./testxgboost
|
||||||
|
ctest -R TestXGBoostCLI --extra-verbose
|
||||||
|
|
||||||
gtest-cpu-nonomp:
|
gtest-cpu-nonomp:
|
||||||
name: Test Google C++ unittest (CPU Non-OMP)
|
name: Test Google C++ unittest (CPU Non-OMP)
|
||||||
@@ -41,7 +45,7 @@ jobs:
|
|||||||
matrix:
|
matrix:
|
||||||
os: [ubuntu-latest]
|
os: [ubuntu-latest]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
- name: Install system packages
|
- name: Install system packages
|
||||||
@@ -59,255 +63,96 @@ jobs:
|
|||||||
cd build
|
cd build
|
||||||
ctest --extra-verbose
|
ctest --extra-verbose
|
||||||
|
|
||||||
python-sdist-test:
|
|
||||||
name: Test installing XGBoost Python source package
|
|
||||||
runs-on: ${{ matrix.os }}
|
|
||||||
strategy:
|
|
||||||
matrix:
|
|
||||||
os: [ubuntu-latest, macos-10.15, windows-latest]
|
|
||||||
python-version: ["3.8"]
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: 'true'
|
|
||||||
- name: Install osx system dependencies
|
|
||||||
if: matrix.os == 'macos-10.15'
|
|
||||||
run: |
|
|
||||||
brew install ninja libomp
|
|
||||||
- name: Install Ubuntu system dependencies
|
|
||||||
if: matrix.os == 'ubuntu-latest'
|
|
||||||
run: |
|
|
||||||
sudo apt-get install -y --no-install-recommends ninja-build
|
|
||||||
- uses: conda-incubator/setup-miniconda@v2
|
|
||||||
with:
|
|
||||||
auto-update-conda: true
|
|
||||||
python-version: ${{ matrix.python-version }}
|
|
||||||
activate-environment: test
|
|
||||||
- name: Display Conda env
|
|
||||||
shell: bash -l {0}
|
|
||||||
run: |
|
|
||||||
conda info
|
|
||||||
conda list
|
|
||||||
- name: Build and install XGBoost
|
|
||||||
shell: bash -l {0}
|
|
||||||
run: |
|
|
||||||
cd python-package
|
|
||||||
python --version
|
|
||||||
python setup.py sdist
|
|
||||||
pip install -v ./dist/xgboost-*.tar.gz
|
|
||||||
cd ..
|
|
||||||
python -c 'import xgboost'
|
|
||||||
|
|
||||||
c-api-demo:
|
c-api-demo:
|
||||||
name: Test installing XGBoost lib + building the C API demo
|
name: Test installing XGBoost lib + building the C API demo
|
||||||
runs-on: ${{ matrix.os }}
|
runs-on: ${{ matrix.os }}
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
shell: bash -l {0}
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
os: ["ubuntu-latest"]
|
os: ["ubuntu-latest"]
|
||||||
python-version: ["3.8"]
|
python-version: ["3.8"]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
- name: Install system packages
|
- uses: mamba-org/provision-with-micromamba@f347426e5745fe3dfc13ec5baf20496990d0281f # v14
|
||||||
run: |
|
|
||||||
sudo apt-get install -y --no-install-recommends ninja-build
|
|
||||||
- uses: conda-incubator/setup-miniconda@v2
|
|
||||||
with:
|
with:
|
||||||
auto-update-conda: true
|
cache-downloads: true
|
||||||
python-version: ${{ matrix.python-version }}
|
cache-env: true
|
||||||
activate-environment: test
|
environment-name: cpp_test
|
||||||
|
environment-file: tests/ci_build/conda_env/cpp_test.yml
|
||||||
- name: Display Conda env
|
- name: Display Conda env
|
||||||
shell: bash -l {0}
|
|
||||||
run: |
|
run: |
|
||||||
conda info
|
conda info
|
||||||
conda list
|
conda list
|
||||||
- name: Build and install XGBoost
|
|
||||||
shell: bash -l {0}
|
- name: Build and install XGBoost static library
|
||||||
run: |
|
run: |
|
||||||
mkdir build
|
mkdir build
|
||||||
cd build
|
cd build
|
||||||
cmake .. -DBUILD_STATIC_LIB=ON -DCMAKE_INSTALL_PREFIX=$CONDA_PREFIX -GNinja
|
cmake .. -DBUILD_STATIC_LIB=ON -DCMAKE_INSTALL_PREFIX=$CONDA_PREFIX -GNinja
|
||||||
ninja -v install
|
ninja -v install
|
||||||
- name: Build and run C API demo
|
cd -
|
||||||
shell: bash -l {0}
|
- name: Build and run C API demo with static
|
||||||
run: |
|
run: |
|
||||||
|
pushd .
|
||||||
cd demo/c-api/
|
cd demo/c-api/
|
||||||
mkdir build
|
mkdir build
|
||||||
cd build
|
cd build
|
||||||
cmake .. -GNinja -DCMAKE_PREFIX_PATH=$CONDA_PREFIX
|
cmake .. -GNinja -DCMAKE_PREFIX_PATH=$CONDA_PREFIX
|
||||||
ninja -v
|
ninja -v
|
||||||
|
ctest
|
||||||
cd ..
|
cd ..
|
||||||
./build/api-demo
|
rm -rf ./build
|
||||||
|
popd
|
||||||
|
|
||||||
test-with-jvm:
|
- name: Build and install XGBoost shared library
|
||||||
name: Test JVM on OS ${{ matrix.os }}
|
|
||||||
runs-on: ${{ matrix.os }}
|
|
||||||
strategy:
|
|
||||||
fail-fast: false
|
|
||||||
matrix:
|
|
||||||
os: [windows-latest, ubuntu-latest]
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: 'true'
|
|
||||||
|
|
||||||
- uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: '3.8'
|
|
||||||
architecture: 'x64'
|
|
||||||
|
|
||||||
- uses: actions/setup-java@v1
|
|
||||||
with:
|
|
||||||
java-version: 1.8
|
|
||||||
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
run: |
|
||||||
python -m pip install wheel setuptools
|
cd build
|
||||||
python -m pip install awscli
|
cmake .. -DBUILD_STATIC_LIB=OFF -DCMAKE_INSTALL_PREFIX=$CONDA_PREFIX -GNinja
|
||||||
|
ninja -v install
|
||||||
- name: Cache Maven packages
|
cd -
|
||||||
uses: actions/cache@v2
|
- name: Build and run C API demo with shared
|
||||||
with:
|
|
||||||
path: ~/.m2
|
|
||||||
key: ${{ runner.os }}-m2-${{ hashFiles('./jvm-packages/pom.xml') }}
|
|
||||||
restore-keys: ${{ runner.os }}-m2
|
|
||||||
|
|
||||||
- name: Test XGBoost4J
|
|
||||||
run: |
|
|
||||||
cd jvm-packages
|
|
||||||
mvn test -B -pl :xgboost4j_2.12
|
|
||||||
|
|
||||||
- name: Extract branch name
|
|
||||||
shell: bash
|
|
||||||
run: echo "##[set-output name=branch;]$(echo ${GITHUB_REF#refs/heads/})"
|
|
||||||
id: extract_branch
|
|
||||||
if: |
|
|
||||||
(github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')) &&
|
|
||||||
matrix.os == 'windows-latest'
|
|
||||||
|
|
||||||
- name: Publish artifact xgboost4j.dll to S3
|
|
||||||
run: |
|
|
||||||
cd lib/
|
|
||||||
Rename-Item -Path xgboost4j.dll -NewName xgboost4j_${{ github.sha }}.dll
|
|
||||||
dir
|
|
||||||
python -m awscli s3 cp xgboost4j_${{ github.sha }}.dll s3://xgboost-nightly-builds/${{ steps.extract_branch.outputs.branch }}/ --acl public-read
|
|
||||||
if: |
|
|
||||||
(github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')) &&
|
|
||||||
matrix.os == 'windows-latest'
|
|
||||||
env:
|
|
||||||
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID_IAM_S3_UPLOADER }}
|
|
||||||
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY_IAM_S3_UPLOADER }}
|
|
||||||
|
|
||||||
|
|
||||||
- name: Test XGBoost4J-Spark
|
|
||||||
run: |
|
|
||||||
rm -rfv build/
|
|
||||||
cd jvm-packages
|
|
||||||
mvn -B test
|
|
||||||
if: matrix.os == 'ubuntu-latest' # Distributed training doesn't work on Windows
|
|
||||||
env:
|
|
||||||
RABIT_MOCK: ON
|
|
||||||
|
|
||||||
lint:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
name: Code linting for Python and C++
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: 'true'
|
|
||||||
- uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: '3.7'
|
|
||||||
architecture: 'x64'
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
|
||||||
python -m pip install wheel setuptools
|
|
||||||
python -m pip install pylint cpplint numpy scipy scikit-learn
|
|
||||||
- name: Run lint
|
|
||||||
run: |
|
|
||||||
make lint
|
|
||||||
|
|
||||||
mypy:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
name: Type checking for Python
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: 'true'
|
|
||||||
- uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: '3.7'
|
|
||||||
architecture: 'x64'
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
|
||||||
python -m pip install wheel setuptools mypy dask[complete] distributed
|
|
||||||
- name: Run mypy
|
|
||||||
run: |
|
|
||||||
make mypy
|
|
||||||
|
|
||||||
doxygen:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
name: Generate C/C++ API doc using Doxygen
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
with:
|
|
||||||
submodules: 'true'
|
|
||||||
- uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: '3.7'
|
|
||||||
architecture: 'x64'
|
|
||||||
- name: Install system packages
|
|
||||||
run: |
|
|
||||||
sudo apt-get install -y --no-install-recommends doxygen graphviz ninja-build
|
|
||||||
python -m pip install wheel setuptools
|
|
||||||
python -m pip install awscli
|
|
||||||
- name: Run Doxygen
|
|
||||||
run: |
|
run: |
|
||||||
|
pushd .
|
||||||
|
cd demo/c-api/
|
||||||
mkdir build
|
mkdir build
|
||||||
cd build
|
cd build
|
||||||
cmake .. -DBUILD_C_DOC=ON -GNinja
|
cmake .. -GNinja -DCMAKE_PREFIX_PATH=$CONDA_PREFIX
|
||||||
ninja -v doc_doxygen
|
ninja -v
|
||||||
- name: Extract branch name
|
ctest
|
||||||
shell: bash
|
popd
|
||||||
run: echo "##[set-output name=branch;]$(echo ${GITHUB_REF#refs/heads/})"
|
./tests/ci_build/verify_link.sh ./demo/c-api/build/basic/api-demo
|
||||||
id: extract_branch
|
./tests/ci_build/verify_link.sh ./demo/c-api/build/external-memory/external-memory-demo
|
||||||
if: github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')
|
|
||||||
- name: Publish
|
|
||||||
run: |
|
|
||||||
cd build/
|
|
||||||
tar cvjf ${{ steps.extract_branch.outputs.branch }}.tar.bz2 doc_doxygen/
|
|
||||||
python -m awscli s3 cp ./${{ steps.extract_branch.outputs.branch }}.tar.bz2 s3://xgboost-docs/ --acl public-read
|
|
||||||
if: github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')
|
|
||||||
env:
|
|
||||||
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID_IAM_S3_UPLOADER }}
|
|
||||||
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY_IAM_S3_UPLOADER }}
|
|
||||||
|
|
||||||
sphinx:
|
cpp-lint:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
name: Build docs using Sphinx
|
name: Code linting for C++
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
- uses: actions/setup-python@v2
|
- uses: actions/setup-python@7f80679172b057fc5e90d70d197929d454754a5a # v4.3.0
|
||||||
with:
|
with:
|
||||||
python-version: '3.8'
|
python-version: "3.8"
|
||||||
architecture: 'x64'
|
architecture: 'x64'
|
||||||
- name: Install system packages
|
- name: Install Python packages
|
||||||
run: |
|
run: |
|
||||||
sudo apt-get install -y --no-install-recommends graphviz
|
python -m pip install wheel setuptools cpplint pylint
|
||||||
python -m pip install wheel setuptools
|
- name: Run lint
|
||||||
python -m pip install -r doc/requirements.txt
|
|
||||||
- name: Extract branch name
|
|
||||||
shell: bash
|
|
||||||
run: echo "##[set-output name=branch;]$(echo ${GITHUB_REF#refs/heads/})"
|
|
||||||
id: extract_branch
|
|
||||||
if: github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')
|
|
||||||
- name: Run Sphinx
|
|
||||||
run: |
|
run: |
|
||||||
make -C doc html
|
python3 dmlc-core/scripts/lint.py xgboost cpp R-package/src
|
||||||
env:
|
|
||||||
SPHINX_GIT_BRANCH: ${{ steps.extract_branch.outputs.branch }}
|
python3 dmlc-core/scripts/lint.py --exclude_path \
|
||||||
|
python-package/xgboost/dmlc-core \
|
||||||
|
python-package/xgboost/include \
|
||||||
|
python-package/xgboost/lib \
|
||||||
|
python-package/xgboost/rabit \
|
||||||
|
python-package/xgboost/src \
|
||||||
|
--pylint-rc python-package/.pylintrc \
|
||||||
|
xgboost \
|
||||||
|
cpp \
|
||||||
|
include src python-package
|
||||||
|
|||||||
298
.github/workflows/python_tests.yml
vendored
Normal file
298
.github/workflows/python_tests.yml
vendored
Normal file
@@ -0,0 +1,298 @@
|
|||||||
|
name: XGBoost-Python-Tests
|
||||||
|
|
||||||
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
shell: bash -l {0}
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
python-mypy-lint:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
name: Type and format checks for the Python package
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest]
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
- uses: mamba-org/provision-with-micromamba@f347426e5745fe3dfc13ec5baf20496990d0281f # v14
|
||||||
|
with:
|
||||||
|
cache-downloads: true
|
||||||
|
cache-env: true
|
||||||
|
environment-name: python_lint
|
||||||
|
environment-file: tests/ci_build/conda_env/python_lint.yml
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
- name: Run mypy
|
||||||
|
run: |
|
||||||
|
python tests/ci_build/lint_python.py --format=0 --type-check=1 --pylint=0
|
||||||
|
- name: Run formatter
|
||||||
|
run: |
|
||||||
|
python tests/ci_build/lint_python.py --format=1 --type-check=0 --pylint=0
|
||||||
|
- name: Run pylint
|
||||||
|
run: |
|
||||||
|
python tests/ci_build/lint_python.py --format=0 --type-check=0 --pylint=1
|
||||||
|
|
||||||
|
python-sdist-test-on-Linux:
|
||||||
|
# Mismatched glibcxx version between system and conda forge.
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
name: Test installing XGBoost Python source package on ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest]
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
- uses: mamba-org/provision-with-micromamba@f347426e5745fe3dfc13ec5baf20496990d0281f # v14
|
||||||
|
with:
|
||||||
|
cache-downloads: true
|
||||||
|
cache-env: true
|
||||||
|
environment-name: sdist_test
|
||||||
|
environment-file: tests/ci_build/conda_env/sdist_test.yml
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
- name: Build and install XGBoost
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
python --version
|
||||||
|
python -m build --sdist
|
||||||
|
pip install -v ./dist/xgboost-*.tar.gz --config-settings use_openmp=False
|
||||||
|
cd ..
|
||||||
|
python -c 'import xgboost'
|
||||||
|
|
||||||
|
python-sdist-test:
|
||||||
|
# Use system toolchain instead of conda toolchain for macos and windows.
|
||||||
|
# MacOS has linker error if clang++ from conda-forge is used
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
name: Test installing XGBoost Python source package on ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
os: [macos-11, windows-latest]
|
||||||
|
python-version: ["3.8"]
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
- name: Install osx system dependencies
|
||||||
|
if: matrix.os == 'macos-11'
|
||||||
|
run: |
|
||||||
|
brew install ninja libomp
|
||||||
|
- uses: conda-incubator/setup-miniconda@35d1405e78aa3f784fe3ce9a2eb378d5eeb62169 # v2.1.1
|
||||||
|
with:
|
||||||
|
auto-update-conda: true
|
||||||
|
python-version: ${{ matrix.python-version }}
|
||||||
|
activate-environment: test
|
||||||
|
- name: Install build
|
||||||
|
run: |
|
||||||
|
conda install -c conda-forge python-build
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
- name: Build and install XGBoost
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
python --version
|
||||||
|
python -m build --sdist
|
||||||
|
pip install -v ./dist/xgboost-*.tar.gz
|
||||||
|
cd ..
|
||||||
|
python -c 'import xgboost'
|
||||||
|
|
||||||
|
python-tests-on-macos:
|
||||||
|
name: Test XGBoost Python package on ${{ matrix.config.os }}
|
||||||
|
runs-on: ${{ matrix.config.os }}
|
||||||
|
timeout-minutes: 60
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
config:
|
||||||
|
- {os: macos-11}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
|
||||||
|
- uses: mamba-org/provision-with-micromamba@f347426e5745fe3dfc13ec5baf20496990d0281f # v14
|
||||||
|
with:
|
||||||
|
cache-downloads: true
|
||||||
|
cache-env: true
|
||||||
|
environment-name: macos_test
|
||||||
|
environment-file: tests/ci_build/conda_env/macos_cpu_test.yml
|
||||||
|
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
|
||||||
|
- name: Build XGBoost on macos
|
||||||
|
run: |
|
||||||
|
brew install ninja
|
||||||
|
|
||||||
|
mkdir build
|
||||||
|
cd build
|
||||||
|
# Set prefix, to use OpenMP library from Conda env
|
||||||
|
# See https://github.com/dmlc/xgboost/issues/7039#issuecomment-1025038228
|
||||||
|
# to learn why we don't use libomp from Homebrew.
|
||||||
|
cmake .. -GNinja -DCMAKE_PREFIX_PATH=$CONDA_PREFIX
|
||||||
|
ninja
|
||||||
|
|
||||||
|
- name: Install Python package
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
python --version
|
||||||
|
pip install -v .
|
||||||
|
|
||||||
|
- name: Test Python package
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/python
|
||||||
|
|
||||||
|
- name: Test Dask Interface
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/test_distributed/test_with_dask
|
||||||
|
|
||||||
|
python-tests-on-win:
|
||||||
|
name: Test XGBoost Python package on ${{ matrix.config.os }}
|
||||||
|
runs-on: ${{ matrix.config.os }}
|
||||||
|
timeout-minutes: 60
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
config:
|
||||||
|
- {os: windows-latest, python-version: '3.8'}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
|
||||||
|
- uses: conda-incubator/setup-miniconda@35d1405e78aa3f784fe3ce9a2eb378d5eeb62169 # v2.1.1
|
||||||
|
with:
|
||||||
|
auto-update-conda: true
|
||||||
|
python-version: ${{ matrix.config.python-version }}
|
||||||
|
activate-environment: win64_env
|
||||||
|
environment-file: tests/ci_build/conda_env/win64_cpu_test.yml
|
||||||
|
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
|
||||||
|
- name: Build XGBoost on Windows
|
||||||
|
run: |
|
||||||
|
mkdir build_msvc
|
||||||
|
cd build_msvc
|
||||||
|
cmake .. -G"Visual Studio 17 2022" -DCMAKE_CONFIGURATION_TYPES="Release" -A x64 -DGOOGLE_TEST=ON -DUSE_DMLC_GTEST=ON
|
||||||
|
cmake --build . --config Release --parallel $(nproc)
|
||||||
|
|
||||||
|
- name: Install Python package
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
python --version
|
||||||
|
pip wheel -v . --wheel-dir dist/
|
||||||
|
pip install ./dist/*.whl
|
||||||
|
|
||||||
|
- name: Test Python package
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/python
|
||||||
|
|
||||||
|
python-tests-on-ubuntu:
|
||||||
|
name: Test XGBoost Python package on ${{ matrix.config.os }}
|
||||||
|
runs-on: ${{ matrix.config.os }}
|
||||||
|
timeout-minutes: 90
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
config:
|
||||||
|
- {os: ubuntu-latest, python-version: "3.8"}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
|
||||||
|
- uses: mamba-org/provision-with-micromamba@f347426e5745fe3dfc13ec5baf20496990d0281f # v14
|
||||||
|
with:
|
||||||
|
cache-downloads: true
|
||||||
|
cache-env: true
|
||||||
|
environment-name: linux_cpu_test
|
||||||
|
environment-file: tests/ci_build/conda_env/linux_cpu_test.yml
|
||||||
|
|
||||||
|
- name: Display Conda env
|
||||||
|
run: |
|
||||||
|
conda info
|
||||||
|
conda list
|
||||||
|
|
||||||
|
- name: Build XGBoost on Ubuntu
|
||||||
|
run: |
|
||||||
|
mkdir build
|
||||||
|
cd build
|
||||||
|
cmake .. -GNinja -DCMAKE_PREFIX_PATH=$CONDA_PREFIX
|
||||||
|
ninja
|
||||||
|
|
||||||
|
- name: Install Python package
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
python --version
|
||||||
|
pip install -v .
|
||||||
|
|
||||||
|
- name: Test Python package
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/python
|
||||||
|
|
||||||
|
- name: Test Dask Interface
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/test_distributed/test_with_dask
|
||||||
|
|
||||||
|
- name: Test PySpark Interface
|
||||||
|
shell: bash -l {0}
|
||||||
|
run: |
|
||||||
|
pytest -s -v -rxXs --durations=0 ./tests/test_distributed/test_with_spark
|
||||||
|
|
||||||
|
python-system-installation-on-ubuntu:
|
||||||
|
name: Test XGBoost Python package System Installation on ${{ matrix.os }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest]
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
|
||||||
|
- name: Set up Python 3.8
|
||||||
|
uses: actions/setup-python@v4
|
||||||
|
with:
|
||||||
|
python-version: 3.8
|
||||||
|
|
||||||
|
- name: Install ninja
|
||||||
|
run: |
|
||||||
|
sudo apt-get update && sudo apt-get install -y ninja-build
|
||||||
|
|
||||||
|
- name: Build XGBoost on Ubuntu
|
||||||
|
run: |
|
||||||
|
mkdir build
|
||||||
|
cd build
|
||||||
|
cmake .. -GNinja
|
||||||
|
ninja
|
||||||
|
|
||||||
|
- name: Copy lib to system lib
|
||||||
|
run: |
|
||||||
|
cp lib/* "$(python -c 'import sys; print(sys.base_prefix)')/lib"
|
||||||
|
|
||||||
|
- name: Install XGBoost in Virtual Environment
|
||||||
|
run: |
|
||||||
|
cd python-package
|
||||||
|
pip install virtualenv
|
||||||
|
virtualenv venv
|
||||||
|
source venv/bin/activate && \
|
||||||
|
pip install -v . --config-settings use_system_libxgboost=True && \
|
||||||
|
python -c 'import xgboost'
|
||||||
41
.github/workflows/python_wheels.yml
vendored
Normal file
41
.github/workflows/python_wheels.yml
vendored
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
name: XGBoost-Python-Wheels
|
||||||
|
|
||||||
|
on: [push, pull_request]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
python-wheels:
|
||||||
|
name: Build wheel for ${{ matrix.platform_id }}
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
include:
|
||||||
|
- os: macos-latest
|
||||||
|
platform_id: macosx_x86_64
|
||||||
|
- os: macos-latest
|
||||||
|
platform_id: macosx_arm64
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
- name: Setup Python
|
||||||
|
uses: actions/setup-python@7f80679172b057fc5e90d70d197929d454754a5a # v4.3.0
|
||||||
|
with:
|
||||||
|
python-version: "3.8"
|
||||||
|
- name: Build wheels
|
||||||
|
run: bash tests/ci_build/build_python_wheels.sh ${{ matrix.platform_id }} ${{ github.sha }}
|
||||||
|
- name: Extract branch name
|
||||||
|
shell: bash
|
||||||
|
run: echo "##[set-output name=branch;]$(echo ${GITHUB_REF#refs/heads/})"
|
||||||
|
id: extract_branch
|
||||||
|
if: github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')
|
||||||
|
- name: Upload Python wheel
|
||||||
|
if: github.ref == 'refs/heads/master' || contains(github.ref, 'refs/heads/release_')
|
||||||
|
run: |
|
||||||
|
python -m pip install awscli
|
||||||
|
python -m awscli s3 cp wheelhouse/*.whl s3://xgboost-nightly-builds/${{ steps.extract_branch.outputs.branch }}/ --acl public-read
|
||||||
|
env:
|
||||||
|
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID_IAM_S3_UPLOADER }}
|
||||||
|
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY_IAM_S3_UPLOADER }}
|
||||||
22
.github/workflows/r_nold.yml
vendored
22
.github/workflows/r_nold.yml
vendored
@@ -1,4 +1,4 @@
|
|||||||
# Run R tests with noLD R. Only triggered by a pull request review
|
# Run expensive R tests with the help of rhub. Only triggered by a pull request review
|
||||||
# See discussion at https://github.com/dmlc/xgboost/pull/6378
|
# See discussion at https://github.com/dmlc/xgboost/pull/6378
|
||||||
|
|
||||||
name: XGBoost-R-noLD
|
name: XGBoost-R-noLD
|
||||||
@@ -7,34 +7,30 @@ on:
|
|||||||
pull_request_review_comment:
|
pull_request_review_comment:
|
||||||
types: [created]
|
types: [created]
|
||||||
|
|
||||||
env:
|
permissions:
|
||||||
R_PACKAGES: c('XML', 'igraph', 'data.table', 'magrittr', 'ggplot2', 'DiagrammeR', 'Ckmeans.1d.dp', 'vcd', 'testthat', 'lintr', 'knitr', 'rmarkdown', 'e1071', 'cplm', 'devtools', 'float', 'titanic')
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
test-R-noLD:
|
test-R-noLD:
|
||||||
if: github.event.comment.body == '/gha run r-nold-test' && contains('OWNER,MEMBER,COLLABORATOR', github.event.comment.author_association)
|
if: github.event.comment.body == '/gha run r-nold-test' && contains('OWNER,MEMBER,COLLABORATOR', github.event.comment.author_association)
|
||||||
timeout-minutes: 120
|
timeout-minutes: 120
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
container: rhub/debian-gcc-devel-nold
|
container:
|
||||||
|
image: rhub/debian-gcc-devel-nold
|
||||||
steps:
|
steps:
|
||||||
- name: Install git and system packages
|
- name: Install git and system packages
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
apt-get update && apt-get install -y git libcurl4-openssl-dev libssl-dev libssh2-1-dev libgit2-dev libxml2-dev
|
apt update && apt install libcurl4-openssl-dev libssl-dev libssh2-1-dev libgit2-dev libglpk-dev libxml2-dev libharfbuzz-dev libfribidi-dev git -y
|
||||||
|
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
shell: bash
|
shell: bash -l {0}
|
||||||
run: |
|
run: |
|
||||||
cat > install_libs.R <<EOT
|
/tmp/R-devel/bin/Rscript -e "source('./R-package/tests/helper_scripts/install_deps.R')"
|
||||||
install.packages(${{ env.R_PACKAGES }},
|
|
||||||
repos = 'http://cloud.r-project.org',
|
|
||||||
dependencies = c('Depends', 'Imports', 'LinkingTo'))
|
|
||||||
EOT
|
|
||||||
/tmp/R-devel/bin/Rscript install_libs.R
|
|
||||||
|
|
||||||
- name: Run R tests
|
- name: Run R tests
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
121
.github/workflows/r_tests.yml
vendored
121
.github/workflows/r_tests.yml
vendored
@@ -3,7 +3,10 @@ name: XGBoost-R-Tests
|
|||||||
on: [push, pull_request]
|
on: [push, pull_request]
|
||||||
|
|
||||||
env:
|
env:
|
||||||
R_PACKAGES: c('XML', 'igraph', 'data.table', 'magrittr', 'ggplot2', 'DiagrammeR', 'Ckmeans.1d.dp', 'vcd', 'testthat', 'lintr', 'knitr', 'rmarkdown', 'e1071', 'cplm', 'devtools', 'float', 'titanic')
|
GITHUB_PAT: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read # to fetch code (actions/checkout)
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
lintr:
|
lintr:
|
||||||
@@ -12,105 +15,121 @@ jobs:
|
|||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
config:
|
config:
|
||||||
- {os: windows-latest, r: 'release', compiler: 'mingw', build: 'autotools'}
|
- {os: ubuntu-latest, r: 'release'}
|
||||||
env:
|
env:
|
||||||
R_REMOTES_NO_ERRORS_FROM_WARNINGS: true
|
R_REMOTES_NO_ERRORS_FROM_WARNINGS: true
|
||||||
RSPM: ${{ matrix.config.rspm }}
|
RSPM: ${{ matrix.config.rspm }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
|
|
||||||
- uses: r-lib/actions/setup-r@master
|
- uses: r-lib/actions/setup-r@11a22a908006c25fe054c4ef0ac0436b1de3edbe # v2.6.4
|
||||||
with:
|
with:
|
||||||
r-version: ${{ matrix.config.r }}
|
r-version: ${{ matrix.config.r }}
|
||||||
|
|
||||||
|
- name: Cache R packages
|
||||||
|
uses: actions/cache@937d24475381cd9c75ae6db12cb4e79714b926ed # v3.0.11
|
||||||
|
with:
|
||||||
|
path: ${{ env.R_LIBS_USER }}
|
||||||
|
key: ${{ runner.os }}-r-${{ matrix.config.r }}-6-${{ hashFiles('R-package/DESCRIPTION') }}
|
||||||
|
restore-keys: ${{ runner.os }}-r-${{ matrix.config.r }}-6-${{ hashFiles('R-package/DESCRIPTION') }}
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
shell: Rscript {0}
|
shell: Rscript {0}
|
||||||
run: |
|
run: |
|
||||||
install.packages(${{ env.R_PACKAGES }},
|
source("./R-package/tests/helper_scripts/install_deps.R")
|
||||||
repos = 'http://cloud.r-project.org',
|
|
||||||
dependencies = c('Depends', 'Imports', 'LinkingTo'))
|
|
||||||
|
|
||||||
- name: Run lintr
|
- name: Run lintr
|
||||||
run: |
|
run: |
|
||||||
cd R-package
|
MAKEFLAGS="-j$(nproc)" R CMD INSTALL R-package/
|
||||||
R.exe CMD INSTALL .
|
Rscript tests/ci_build/lint_r.R $(pwd)
|
||||||
Rscript.exe tests/helper_scripts/run_lint.R
|
|
||||||
|
|
||||||
test-with-R:
|
test-R-on-Windows:
|
||||||
runs-on: ${{ matrix.config.os }}
|
runs-on: ${{ matrix.config.os }}
|
||||||
name: Test R on OS ${{ matrix.config.os }}, R ${{ matrix.config.r }}, Compiler ${{ matrix.config.compiler }}, Build ${{ matrix.config.build }}
|
name: Test R on OS ${{ matrix.config.os }}, R ${{ matrix.config.r }}, Compiler ${{ matrix.config.compiler }}, Build ${{ matrix.config.build }}
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
config:
|
config:
|
||||||
- {os: windows-2016, r: 'release', compiler: 'mingw', build: 'autotools'}
|
- {os: windows-latest, r: 'release', compiler: 'mingw', build: 'autotools'}
|
||||||
- {os: windows-2016, r: 'release', compiler: 'msvc', build: 'cmake'}
|
- {os: windows-latest, r: '4.2.0', compiler: 'msvc', build: 'cmake'}
|
||||||
- {os: windows-2016, r: 'release', compiler: 'mingw', build: 'cmake'}
|
|
||||||
env:
|
env:
|
||||||
R_REMOTES_NO_ERRORS_FROM_WARNINGS: true
|
R_REMOTES_NO_ERRORS_FROM_WARNINGS: true
|
||||||
RSPM: ${{ matrix.config.rspm }}
|
RSPM: ${{ matrix.config.rspm }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
|
|
||||||
- uses: r-lib/actions/setup-r@master
|
- uses: r-lib/actions/setup-r@11a22a908006c25fe054c4ef0ac0436b1de3edbe # v2.6.4
|
||||||
with:
|
with:
|
||||||
r-version: ${{ matrix.config.r }}
|
r-version: ${{ matrix.config.r }}
|
||||||
|
|
||||||
|
- name: Cache R packages
|
||||||
|
uses: actions/cache@937d24475381cd9c75ae6db12cb4e79714b926ed # v3.0.11
|
||||||
|
with:
|
||||||
|
path: ${{ env.R_LIBS_USER }}
|
||||||
|
key: ${{ runner.os }}-r-${{ matrix.config.r }}-6-${{ hashFiles('R-package/DESCRIPTION') }}
|
||||||
|
restore-keys: ${{ runner.os }}-r-${{ matrix.config.r }}-6-${{ hashFiles('R-package/DESCRIPTION') }}
|
||||||
|
|
||||||
|
- uses: actions/setup-python@7f80679172b057fc5e90d70d197929d454754a5a # v4.3.0
|
||||||
|
with:
|
||||||
|
python-version: "3.8"
|
||||||
|
architecture: 'x64'
|
||||||
|
|
||||||
|
- uses: r-lib/actions/setup-tinytex@v2
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
shell: Rscript {0}
|
shell: Rscript {0}
|
||||||
run: |
|
run: |
|
||||||
install.packages(${{ env.R_PACKAGES }},
|
source("./R-package/tests/helper_scripts/install_deps.R")
|
||||||
repos = 'http://cloud.r-project.org',
|
|
||||||
dependencies = c('Depends', 'Imports', 'LinkingTo'))
|
|
||||||
|
|
||||||
- uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: '3.7'
|
|
||||||
architecture: 'x64'
|
|
||||||
|
|
||||||
- name: Test R
|
- name: Test R
|
||||||
run: |
|
run: |
|
||||||
python tests/ci_build/test_r_package.py --compiler='${{ matrix.config.compiler }}' --build-tool='${{ matrix.config.build }}'
|
python tests/ci_build/test_r_package.py --compiler='${{ matrix.config.compiler }}' --build-tool="${{ matrix.config.build }}" --task=check
|
||||||
|
|
||||||
test-R-CRAN:
|
test-R-on-Debian:
|
||||||
|
name: Test R package on Debian
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
container:
|
||||||
strategy:
|
image: rhub/debian-gcc-devel
|
||||||
fail-fast: false
|
|
||||||
matrix:
|
|
||||||
config:
|
|
||||||
- {r: 'release'}
|
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- name: Install system dependencies
|
||||||
|
run: |
|
||||||
|
# Must run before checkout to have the latest git installed.
|
||||||
|
# No need to add pandoc, the container has it figured out.
|
||||||
|
apt update && apt install libcurl4-openssl-dev libssl-dev libssh2-1-dev libgit2-dev libglpk-dev libxml2-dev libharfbuzz-dev libfribidi-dev git -y
|
||||||
|
|
||||||
|
- name: Trust git cloning project sources
|
||||||
|
run: |
|
||||||
|
git config --global --add safe.directory "${GITHUB_WORKSPACE}"
|
||||||
|
|
||||||
|
- uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v2.5.0
|
||||||
with:
|
with:
|
||||||
submodules: 'true'
|
submodules: 'true'
|
||||||
|
|
||||||
- uses: r-lib/actions/setup-r@master
|
|
||||||
with:
|
|
||||||
r-version: ${{ matrix.config.r }}
|
|
||||||
|
|
||||||
- uses: r-lib/actions/setup-tinytex@master
|
|
||||||
|
|
||||||
- name: Install system packages
|
|
||||||
run: |
|
|
||||||
sudo apt-get update && sudo apt-get install libcurl4-openssl-dev libssl-dev libssh2-1-dev libgit2-dev
|
|
||||||
|
|
||||||
- name: Install dependencies
|
- name: Install dependencies
|
||||||
shell: Rscript {0}
|
shell: bash -l {0}
|
||||||
run: |
|
run: |
|
||||||
install.packages(${{ env.R_PACKAGES }},
|
/tmp/R-devel/bin/Rscript -e "source('./R-package/tests/helper_scripts/install_deps.R')"
|
||||||
repos = 'http://cloud.r-project.org',
|
|
||||||
dependencies = c('Depends', 'Imports', 'LinkingTo'))
|
|
||||||
|
|
||||||
- name: Check R Package
|
- name: Test R
|
||||||
|
shell: bash -l {0}
|
||||||
run: |
|
run: |
|
||||||
# Print stacktrace upon success of failure
|
python3 tests/ci_build/test_r_package.py --r=/tmp/R-devel/bin/R --build-tool=autotools --task=check
|
||||||
make Rcheck || tests/ci_build/print_r_stacktrace.sh fail
|
|
||||||
tests/ci_build/print_r_stacktrace.sh success
|
- uses: dorny/paths-filter@v2
|
||||||
|
id: changes
|
||||||
|
with:
|
||||||
|
filters: |
|
||||||
|
r_package:
|
||||||
|
- 'R-package/**'
|
||||||
|
|
||||||
|
- name: Run document check
|
||||||
|
if: steps.changes.outputs.r_package == 'true'
|
||||||
|
run: |
|
||||||
|
python3 tests/ci_build/test_r_package.py --r=/tmp/R-devel/bin/R --task=doc
|
||||||
|
|||||||
54
.github/workflows/scorecards.yml
vendored
Normal file
54
.github/workflows/scorecards.yml
vendored
Normal file
@@ -0,0 +1,54 @@
|
|||||||
|
name: Scorecards supply-chain security
|
||||||
|
on:
|
||||||
|
# Only the default branch is supported.
|
||||||
|
branch_protection_rule:
|
||||||
|
schedule:
|
||||||
|
- cron: '17 2 * * 6'
|
||||||
|
push:
|
||||||
|
branches: [ "master" ]
|
||||||
|
|
||||||
|
# Declare default permissions as read only.
|
||||||
|
permissions: read-all
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
analysis:
|
||||||
|
name: Scorecards analysis
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
# Needed to upload the results to code-scanning dashboard.
|
||||||
|
security-events: write
|
||||||
|
# Used to receive a badge.
|
||||||
|
id-token: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: "Checkout code"
|
||||||
|
uses: actions/checkout@a12a3943b4bdde767164f792f33f40b04645d846 # tag=v3.0.0
|
||||||
|
with:
|
||||||
|
persist-credentials: false
|
||||||
|
|
||||||
|
- name: "Run analysis"
|
||||||
|
uses: ossf/scorecard-action@08b4669551908b1024bb425080c797723083c031 # tag=v2.2.0
|
||||||
|
with:
|
||||||
|
results_file: results.sarif
|
||||||
|
results_format: sarif
|
||||||
|
|
||||||
|
# Publish the results for public repositories to enable scorecard badges. For more details, see
|
||||||
|
# https://github.com/ossf/scorecard-action#publishing-results.
|
||||||
|
# For private repositories, `publish_results` will automatically be set to `false`, regardless
|
||||||
|
# of the value entered here.
|
||||||
|
publish_results: true
|
||||||
|
|
||||||
|
# Upload the results as artifacts (optional). Commenting out will disable uploads of run results in SARIF
|
||||||
|
# format to the repository Actions tab.
|
||||||
|
- name: "Upload artifact"
|
||||||
|
uses: actions/upload-artifact@0b7f8abb1508181956e8e162db84b466c27e18ce # tag=v3.1.2
|
||||||
|
with:
|
||||||
|
name: SARIF file
|
||||||
|
path: results.sarif
|
||||||
|
retention-days: 5
|
||||||
|
|
||||||
|
# Upload the results to GitHub's code scanning dashboard.
|
||||||
|
- name: "Upload to code-scanning"
|
||||||
|
uses: github/codeql-action/upload-sarif@7b6664fa89524ee6e3c3e9749402d5afd69b3cd8 # tag=v2.14.1
|
||||||
|
with:
|
||||||
|
sarif_file: results.sarif
|
||||||
44
.github/workflows/update_rapids.yml
vendored
Normal file
44
.github/workflows/update_rapids.yml
vendored
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
name: update-rapids
|
||||||
|
|
||||||
|
on:
|
||||||
|
workflow_dispatch:
|
||||||
|
schedule:
|
||||||
|
- cron: "0 20 * * *" # Run once daily
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
pull-requests: write
|
||||||
|
contents: write
|
||||||
|
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
shell: bash -l {0}
|
||||||
|
|
||||||
|
concurrency:
|
||||||
|
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
|
|
||||||
|
env:
|
||||||
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} # To use GitHub CLI
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
update-rapids:
|
||||||
|
name: Check latest RAPIDS
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
submodules: 'true'
|
||||||
|
- name: Check latest RAPIDS and update conftest.sh
|
||||||
|
run: |
|
||||||
|
bash tests/buildkite/update-rapids.sh
|
||||||
|
- name: Create Pull Request
|
||||||
|
uses: peter-evans/create-pull-request@v5
|
||||||
|
if: github.ref == 'refs/heads/master'
|
||||||
|
with:
|
||||||
|
add-paths: |
|
||||||
|
tests/buildkite
|
||||||
|
branch: create-pull-request/update-rapids
|
||||||
|
base: master
|
||||||
|
title: "[CI] Update RAPIDS to latest stable"
|
||||||
|
commit-message: "[CI] Update RAPIDS to latest stable"
|
||||||
|
|
||||||
29
.gitignore
vendored
29
.gitignore
vendored
@@ -48,10 +48,13 @@ Debug
|
|||||||
*.Rproj
|
*.Rproj
|
||||||
./xgboost.mpi
|
./xgboost.mpi
|
||||||
./xgboost.mock
|
./xgboost.mock
|
||||||
|
*.bak
|
||||||
#.Rbuildignore
|
#.Rbuildignore
|
||||||
R-package.Rproj
|
R-package.Rproj
|
||||||
*.cache*
|
*.cache*
|
||||||
.mypy_cache/
|
.mypy_cache/
|
||||||
|
doxygen
|
||||||
|
|
||||||
# java
|
# java
|
||||||
java/xgboost4j/target
|
java/xgboost4j/target
|
||||||
java/xgboost4j/tmp
|
java/xgboost4j/tmp
|
||||||
@@ -63,6 +66,7 @@ nb-configuration*
|
|||||||
# Eclipse
|
# Eclipse
|
||||||
.project
|
.project
|
||||||
.cproject
|
.cproject
|
||||||
|
.classpath
|
||||||
.pydevproject
|
.pydevproject
|
||||||
.settings/
|
.settings/
|
||||||
build
|
build
|
||||||
@@ -96,8 +100,11 @@ metastore_db
|
|||||||
R-package/src/Makevars
|
R-package/src/Makevars
|
||||||
*.lib
|
*.lib
|
||||||
|
|
||||||
# Visual Studio Code
|
# Visual Studio
|
||||||
/.vscode/
|
.vs/
|
||||||
|
CMakeSettings.json
|
||||||
|
*.ilk
|
||||||
|
*.pdb
|
||||||
|
|
||||||
# IntelliJ/CLion
|
# IntelliJ/CLion
|
||||||
.idea
|
.idea
|
||||||
@@ -125,3 +132,21 @@ credentials.csv
|
|||||||
*.pub
|
*.pub
|
||||||
*.rdp
|
*.rdp
|
||||||
*_rsa
|
*_rsa
|
||||||
|
|
||||||
|
# Visual Studio code + extensions
|
||||||
|
.vscode
|
||||||
|
.metals
|
||||||
|
.bloop
|
||||||
|
|
||||||
|
# python tests
|
||||||
|
demo/**/*.txt
|
||||||
|
*.dmatrix
|
||||||
|
.hypothesis
|
||||||
|
__MACOSX/
|
||||||
|
model*.json
|
||||||
|
|
||||||
|
# R tests
|
||||||
|
*.libsvm
|
||||||
|
*.rds
|
||||||
|
Rplots.pdf
|
||||||
|
*.zip
|
||||||
|
|||||||
3
.gitmodules
vendored
3
.gitmodules
vendored
@@ -2,9 +2,6 @@
|
|||||||
path = dmlc-core
|
path = dmlc-core
|
||||||
url = https://github.com/dmlc/dmlc-core
|
url = https://github.com/dmlc/dmlc-core
|
||||||
branch = main
|
branch = main
|
||||||
[submodule "cub"]
|
|
||||||
path = cub
|
|
||||||
url = https://github.com/NVlabs/cub
|
|
||||||
[submodule "gputreeshap"]
|
[submodule "gputreeshap"]
|
||||||
path = gputreeshap
|
path = gputreeshap
|
||||||
url = https://github.com/rapidsai/gputreeshap.git
|
url = https://github.com/rapidsai/gputreeshap.git
|
||||||
|
|||||||
34
.readthedocs.yaml
Normal file
34
.readthedocs.yaml
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
# .readthedocs.yaml
|
||||||
|
# Read the Docs configuration file
|
||||||
|
# See https://docs.readthedocs.io/en/stable/config-file/v2.html for details
|
||||||
|
|
||||||
|
# Required
|
||||||
|
version: 2
|
||||||
|
|
||||||
|
submodules:
|
||||||
|
include: all
|
||||||
|
|
||||||
|
# Set the version of Python and other tools you might need
|
||||||
|
build:
|
||||||
|
os: ubuntu-22.04
|
||||||
|
tools:
|
||||||
|
python: "3.8"
|
||||||
|
apt_packages:
|
||||||
|
- graphviz
|
||||||
|
- cmake
|
||||||
|
- g++
|
||||||
|
- doxygen
|
||||||
|
- ninja-build
|
||||||
|
|
||||||
|
# Build documentation in the docs/ directory with Sphinx
|
||||||
|
sphinx:
|
||||||
|
configuration: doc/conf.py
|
||||||
|
|
||||||
|
# If using Sphinx, optionally build your docs in additional formats such as PDF
|
||||||
|
formats:
|
||||||
|
- pdf
|
||||||
|
|
||||||
|
# Optionally declare the Python requirements required to build your docs
|
||||||
|
python:
|
||||||
|
install:
|
||||||
|
- requirements: doc/requirements.txt
|
||||||
68
.travis.yml
68
.travis.yml
@@ -1,68 +0,0 @@
|
|||||||
sudo: required
|
|
||||||
|
|
||||||
dist: bionic
|
|
||||||
|
|
||||||
env:
|
|
||||||
global:
|
|
||||||
- secure: "PR16i9F8QtNwn99C5NDp8nptAS+97xwDtXEJJfEiEVhxPaaRkOp0MPWhogCaK0Eclxk1TqkgWbdXFknwGycX620AzZWa/A1K3gAs+GrpzqhnPMuoBJ0Z9qxXTbSJvCyvMbYwVrjaxc/zWqdMU8waWz8A7iqKGKs/SqbQ3rO6v7c="
|
|
||||||
- secure: "dAGAjBokqm/0nVoLMofQni/fWIBcYSmdq4XvCBX1ZAMDsWnuOfz/4XCY6h2lEI1rVHZQ+UdZkc9PioOHGPZh5BnvE49/xVVWr9c4/61lrDOlkD01ZjSAeoV0fAZq+93V/wPl4QV+MM+Sem9hNNzFSbN5VsQLAiWCSapWsLdKzqA="
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
include:
|
|
||||||
- os: osx
|
|
||||||
arch: amd64
|
|
||||||
osx_image: xcode10.2
|
|
||||||
env: TASK=python_test
|
|
||||||
- os: osx
|
|
||||||
arch: amd64
|
|
||||||
osx_image: xcode10.2
|
|
||||||
env: TASK=java_test
|
|
||||||
|
|
||||||
# dependent brew packages
|
|
||||||
addons:
|
|
||||||
homebrew:
|
|
||||||
packages:
|
|
||||||
- cmake
|
|
||||||
- libomp
|
|
||||||
- graphviz
|
|
||||||
- openssl
|
|
||||||
- libgit2
|
|
||||||
- lz4
|
|
||||||
- wget
|
|
||||||
- r
|
|
||||||
update: true
|
|
||||||
apt:
|
|
||||||
packages:
|
|
||||||
- snapd
|
|
||||||
- unzip
|
|
||||||
|
|
||||||
before_install:
|
|
||||||
- source tests/travis/travis_setup_env.sh
|
|
||||||
- if [ "${TASK}" != "python_sdist_test" ]; then export PYTHONPATH=${PYTHONPATH}:${PWD}/python-package; fi
|
|
||||||
- echo "MAVEN_OPTS='-Xmx2g -XX:MaxPermSize=1024m -XX:ReservedCodeCacheSize=512m -Dorg.slf4j.simpleLogger.defaultLogLevel=error'" > ~/.mavenrc
|
|
||||||
|
|
||||||
install:
|
|
||||||
- source tests/travis/setup.sh
|
|
||||||
|
|
||||||
script:
|
|
||||||
- tests/travis/run_test.sh
|
|
||||||
|
|
||||||
cache:
|
|
||||||
directories:
|
|
||||||
- ${HOME}/.cache/usr
|
|
||||||
- ${HOME}/.cache/pip
|
|
||||||
|
|
||||||
before_cache:
|
|
||||||
- tests/travis/travis_before_cache.sh
|
|
||||||
|
|
||||||
after_failure:
|
|
||||||
- tests/travis/travis_after_failure.sh
|
|
||||||
|
|
||||||
after_success:
|
|
||||||
- tree build
|
|
||||||
- bash <(curl -s https://codecov.io/bash) -a '-o src/ src/*.c'
|
|
||||||
|
|
||||||
notifications:
|
|
||||||
email:
|
|
||||||
on_success: change
|
|
||||||
on_failure: always
|
|
||||||
1
CITATION
1
CITATION
@@ -15,4 +15,3 @@
|
|||||||
address = {New York, NY, USA},
|
address = {New York, NY, USA},
|
||||||
keywords = {large-scale machine learning},
|
keywords = {large-scale machine learning},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
179
CMakeLists.txt
179
CMakeLists.txt
@@ -1,9 +1,10 @@
|
|||||||
cmake_minimum_required(VERSION 3.13)
|
cmake_minimum_required(VERSION 3.18 FATAL_ERROR)
|
||||||
project(xgboost LANGUAGES CXX C VERSION 1.4.0)
|
project(xgboost LANGUAGES CXX C VERSION 2.0.0)
|
||||||
include(cmake/Utils.cmake)
|
include(cmake/Utils.cmake)
|
||||||
list(APPEND CMAKE_MODULE_PATH "${xgboost_SOURCE_DIR}/cmake/modules")
|
list(APPEND CMAKE_MODULE_PATH "${xgboost_SOURCE_DIR}/cmake/modules")
|
||||||
cmake_policy(SET CMP0022 NEW)
|
cmake_policy(SET CMP0022 NEW)
|
||||||
cmake_policy(SET CMP0079 NEW)
|
cmake_policy(SET CMP0079 NEW)
|
||||||
|
cmake_policy(SET CMP0076 NEW)
|
||||||
set(CMAKE_POLICY_DEFAULT_CMP0063 NEW)
|
set(CMAKE_POLICY_DEFAULT_CMP0063 NEW)
|
||||||
cmake_policy(SET CMP0063 NEW)
|
cmake_policy(SET CMP0063 NEW)
|
||||||
|
|
||||||
@@ -13,8 +14,24 @@ endif ((${CMAKE_VERSION} VERSION_GREATER 3.13) OR (${CMAKE_VERSION} VERSION_EQUA
|
|||||||
|
|
||||||
message(STATUS "CMake version ${CMAKE_VERSION}")
|
message(STATUS "CMake version ${CMAKE_VERSION}")
|
||||||
|
|
||||||
if (CMAKE_COMPILER_IS_GNUCC AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS 5.0)
|
# Check compiler versions
|
||||||
message(FATAL_ERROR "GCC version must be at least 5.0!")
|
# Use recent compilers to ensure that std::filesystem is available
|
||||||
|
if(MSVC)
|
||||||
|
if(MSVC_VERSION LESS 1920)
|
||||||
|
message(FATAL_ERROR "Need Visual Studio 2019 or newer to build XGBoost")
|
||||||
|
endif()
|
||||||
|
elseif(CMAKE_CXX_COMPILER_ID STREQUAL "GNU")
|
||||||
|
if(CMAKE_CXX_COMPILER_VERSION VERSION_LESS "8.1")
|
||||||
|
message(FATAL_ERROR "Need GCC 8.1 or newer to build XGBoost")
|
||||||
|
endif()
|
||||||
|
elseif(CMAKE_CXX_COMPILER_ID STREQUAL "AppleClang")
|
||||||
|
if(CMAKE_CXX_COMPILER_VERSION VERSION_LESS "11.0")
|
||||||
|
message(FATAL_ERROR "Need Xcode 11.0 (AppleClang 11.0) or newer to build XGBoost")
|
||||||
|
endif()
|
||||||
|
elseif(CMAKE_CXX_COMPILER_ID STREQUAL "Clang")
|
||||||
|
if(CMAKE_CXX_COMPILER_VERSION VERSION_LESS "9.0")
|
||||||
|
message(FATAL_ERROR "Need Clang 9.0 or newer to build XGBoost")
|
||||||
|
endif()
|
||||||
endif()
|
endif()
|
||||||
|
|
||||||
include(${xgboost_SOURCE_DIR}/cmake/FindPrefetchIntrinsics.cmake)
|
include(${xgboost_SOURCE_DIR}/cmake/FindPrefetchIntrinsics.cmake)
|
||||||
@@ -28,6 +45,7 @@ set_default_configuration_release()
|
|||||||
option(BUILD_C_DOC "Build documentation for C APIs using Doxygen." OFF)
|
option(BUILD_C_DOC "Build documentation for C APIs using Doxygen." OFF)
|
||||||
option(USE_OPENMP "Build with OpenMP support." ON)
|
option(USE_OPENMP "Build with OpenMP support." ON)
|
||||||
option(BUILD_STATIC_LIB "Build static library" OFF)
|
option(BUILD_STATIC_LIB "Build static library" OFF)
|
||||||
|
option(FORCE_SHARED_CRT "Build with dynamic CRT on Windows (/MD)" OFF)
|
||||||
option(RABIT_BUILD_MPI "Build MPI" OFF)
|
option(RABIT_BUILD_MPI "Build MPI" OFF)
|
||||||
## Bindings
|
## Bindings
|
||||||
option(JVM_BINDINGS "Build JVM bindings" OFF)
|
option(JVM_BINDINGS "Build JVM bindings" OFF)
|
||||||
@@ -45,8 +63,10 @@ option(USE_NVTX "Build with cuda profiling annotations. Developers only." OFF)
|
|||||||
set(NVTX_HEADER_DIR "" CACHE PATH "Path to the stand-alone nvtx header")
|
set(NVTX_HEADER_DIR "" CACHE PATH "Path to the stand-alone nvtx header")
|
||||||
option(RABIT_MOCK "Build rabit with mock" OFF)
|
option(RABIT_MOCK "Build rabit with mock" OFF)
|
||||||
option(HIDE_CXX_SYMBOLS "Build shared library and hide all C++ symbols" OFF)
|
option(HIDE_CXX_SYMBOLS "Build shared library and hide all C++ symbols" OFF)
|
||||||
|
option(KEEP_BUILD_ARTIFACTS_IN_BINARY_DIR "Output build artifacts in CMake binary dir" OFF)
|
||||||
## CUDA
|
## CUDA
|
||||||
option(USE_CUDA "Build with GPU acceleration" OFF)
|
option(USE_CUDA "Build with GPU acceleration" OFF)
|
||||||
|
option(USE_PER_THREAD_DEFAULT_STREAM "Build with per-thread default stream" ON)
|
||||||
option(USE_NCCL "Build with NCCL to enable distributed GPU support." OFF)
|
option(USE_NCCL "Build with NCCL to enable distributed GPU support." OFF)
|
||||||
option(BUILD_WITH_SHARED_NCCL "Build with shared NCCL library." OFF)
|
option(BUILD_WITH_SHARED_NCCL "Build with shared NCCL library." OFF)
|
||||||
set(GPU_COMPUTE_VER "" CACHE STRING
|
set(GPU_COMPUTE_VER "" CACHE STRING
|
||||||
@@ -62,9 +82,9 @@ set(ENABLED_SANITIZERS "address" "leak" CACHE STRING
|
|||||||
"Semicolon separated list of sanitizer names. E.g 'address;leak'. Supported sanitizers are
|
"Semicolon separated list of sanitizer names. E.g 'address;leak'. Supported sanitizers are
|
||||||
address, leak, undefined and thread.")
|
address, leak, undefined and thread.")
|
||||||
## Plugins
|
## Plugins
|
||||||
option(PLUGIN_LZ4 "Build lz4 plugin" OFF)
|
|
||||||
option(PLUGIN_DENSE_PARSER "Build dense parser plugin" OFF)
|
option(PLUGIN_DENSE_PARSER "Build dense parser plugin" OFF)
|
||||||
option(PLUGIN_RMM "Build with RAPIDS Memory Manager (RMM)" OFF)
|
option(PLUGIN_RMM "Build with RAPIDS Memory Manager (RMM)" OFF)
|
||||||
|
option(PLUGIN_FEDERATED "Build with Federated Learning" OFF)
|
||||||
## TODO: 1. Add check if DPC++ compiler is used for building
|
## TODO: 1. Add check if DPC++ compiler is used for building
|
||||||
option(PLUGIN_UPDATER_ONEAPI "DPC++ updater" OFF)
|
option(PLUGIN_UPDATER_ONEAPI "DPC++ updater" OFF)
|
||||||
option(ADD_PKGCONFIG "Add xgboost.pc into system." ON)
|
option(ADD_PKGCONFIG "Add xgboost.pc into system." ON)
|
||||||
@@ -92,6 +112,9 @@ endif (R_LIB AND GOOGLE_TEST)
|
|||||||
if (USE_AVX)
|
if (USE_AVX)
|
||||||
message(SEND_ERROR "The option 'USE_AVX' is deprecated as experimental AVX features have been removed from XGBoost.")
|
message(SEND_ERROR "The option 'USE_AVX' is deprecated as experimental AVX features have been removed from XGBoost.")
|
||||||
endif (USE_AVX)
|
endif (USE_AVX)
|
||||||
|
if (PLUGIN_LZ4)
|
||||||
|
message(SEND_ERROR "The option 'PLUGIN_LZ4' is removed from XGBoost.")
|
||||||
|
endif (PLUGIN_LZ4)
|
||||||
if (PLUGIN_RMM AND NOT (USE_CUDA))
|
if (PLUGIN_RMM AND NOT (USE_CUDA))
|
||||||
message(SEND_ERROR "`PLUGIN_RMM` must be enabled with `USE_CUDA` flag.")
|
message(SEND_ERROR "`PLUGIN_RMM` must be enabled with `USE_CUDA` flag.")
|
||||||
endif (PLUGIN_RMM AND NOT (USE_CUDA))
|
endif (PLUGIN_RMM AND NOT (USE_CUDA))
|
||||||
@@ -109,6 +132,20 @@ endif (ENABLE_ALL_WARNINGS)
|
|||||||
if (BUILD_STATIC_LIB AND (R_LIB OR JVM_BINDINGS))
|
if (BUILD_STATIC_LIB AND (R_LIB OR JVM_BINDINGS))
|
||||||
message(SEND_ERROR "Cannot build a static library libxgboost.a when R or JVM packages are enabled.")
|
message(SEND_ERROR "Cannot build a static library libxgboost.a when R or JVM packages are enabled.")
|
||||||
endif (BUILD_STATIC_LIB AND (R_LIB OR JVM_BINDINGS))
|
endif (BUILD_STATIC_LIB AND (R_LIB OR JVM_BINDINGS))
|
||||||
|
if (PLUGIN_FEDERATED)
|
||||||
|
if (CMAKE_CROSSCOMPILING)
|
||||||
|
message(SEND_ERROR "Cannot cross compile with federated learning support")
|
||||||
|
endif ()
|
||||||
|
if (BUILD_STATIC_LIB)
|
||||||
|
message(SEND_ERROR "Cannot build static lib with federated learning support")
|
||||||
|
endif ()
|
||||||
|
if (R_LIB OR JVM_BINDINGS)
|
||||||
|
message(SEND_ERROR "Cannot enable federated learning support when R or JVM packages are enabled.")
|
||||||
|
endif ()
|
||||||
|
if (WIN32)
|
||||||
|
message(SEND_ERROR "Federated learning not supported for Windows platform")
|
||||||
|
endif ()
|
||||||
|
endif ()
|
||||||
|
|
||||||
#-- Sanitizer
|
#-- Sanitizer
|
||||||
if (USE_SANITIZER)
|
if (USE_SANITIZER)
|
||||||
@@ -117,18 +154,20 @@ if (USE_SANITIZER)
|
|||||||
endif (USE_SANITIZER)
|
endif (USE_SANITIZER)
|
||||||
|
|
||||||
if (USE_CUDA)
|
if (USE_CUDA)
|
||||||
SET(USE_OPENMP ON CACHE BOOL "CUDA requires OpenMP" FORCE)
|
set(USE_OPENMP ON CACHE BOOL "CUDA requires OpenMP" FORCE)
|
||||||
# `export CXX=' is ignored by CMake CUDA.
|
# `export CXX=' is ignored by CMake CUDA.
|
||||||
set(CMAKE_CUDA_HOST_COMPILER ${CMAKE_CXX_COMPILER})
|
set(CMAKE_CUDA_HOST_COMPILER ${CMAKE_CXX_COMPILER})
|
||||||
message(STATUS "Configured CUDA host compiler: ${CMAKE_CUDA_HOST_COMPILER}")
|
message(STATUS "Configured CUDA host compiler: ${CMAKE_CUDA_HOST_COMPILER}")
|
||||||
|
|
||||||
enable_language(CUDA)
|
enable_language(CUDA)
|
||||||
if (${CMAKE_CUDA_COMPILER_VERSION} VERSION_LESS 10.0)
|
if (${CMAKE_CUDA_COMPILER_VERSION} VERSION_LESS 11.0)
|
||||||
message(FATAL_ERROR "CUDA version must be at least 10.0!")
|
message(FATAL_ERROR "CUDA version must be at least 11.0!")
|
||||||
endif()
|
endif()
|
||||||
set(GEN_CODE "")
|
set(GEN_CODE "")
|
||||||
format_gencode_flags("${GPU_COMPUTE_VER}" GEN_CODE)
|
format_gencode_flags("${GPU_COMPUTE_VER}" GEN_CODE)
|
||||||
add_subdirectory(${PROJECT_SOURCE_DIR}/gputreeshap)
|
add_subdirectory(${PROJECT_SOURCE_DIR}/gputreeshap)
|
||||||
|
|
||||||
|
find_package(CUDAToolkit REQUIRED)
|
||||||
endif (USE_CUDA)
|
endif (USE_CUDA)
|
||||||
|
|
||||||
if (FORCE_COLORED_OUTPUT AND (CMAKE_GENERATOR STREQUAL "Ninja") AND
|
if (FORCE_COLORED_OUTPUT AND (CMAKE_GENERATOR STREQUAL "Ninja") AND
|
||||||
@@ -141,34 +180,54 @@ find_package(Threads REQUIRED)
|
|||||||
|
|
||||||
if (USE_OPENMP)
|
if (USE_OPENMP)
|
||||||
if (APPLE)
|
if (APPLE)
|
||||||
# Require CMake 3.16+ on Mac OSX, as previous versions of CMake had trouble locating
|
find_package(OpenMP)
|
||||||
# OpenMP on Mac. See https://github.com/dmlc/xgboost/pull/5146#issuecomment-568312706
|
if (NOT OpenMP_FOUND)
|
||||||
cmake_minimum_required(VERSION 3.16)
|
# Try again with extra path info; required for libomp 15+ from Homebrew
|
||||||
endif (APPLE)
|
execute_process(COMMAND brew --prefix libomp
|
||||||
|
OUTPUT_VARIABLE HOMEBREW_LIBOMP_PREFIX
|
||||||
|
OUTPUT_STRIP_TRAILING_WHITESPACE)
|
||||||
|
set(OpenMP_C_FLAGS
|
||||||
|
"-Xpreprocessor -fopenmp -I${HOMEBREW_LIBOMP_PREFIX}/include")
|
||||||
|
set(OpenMP_CXX_FLAGS
|
||||||
|
"-Xpreprocessor -fopenmp -I${HOMEBREW_LIBOMP_PREFIX}/include")
|
||||||
|
set(OpenMP_C_LIB_NAMES omp)
|
||||||
|
set(OpenMP_CXX_LIB_NAMES omp)
|
||||||
|
set(OpenMP_omp_LIBRARY ${HOMEBREW_LIBOMP_PREFIX}/lib/libomp.dylib)
|
||||||
find_package(OpenMP REQUIRED)
|
find_package(OpenMP REQUIRED)
|
||||||
|
endif ()
|
||||||
|
else ()
|
||||||
|
find_package(OpenMP REQUIRED)
|
||||||
|
endif ()
|
||||||
endif (USE_OPENMP)
|
endif (USE_OPENMP)
|
||||||
|
#Add for IBM i
|
||||||
|
if (${CMAKE_SYSTEM_NAME} MATCHES "OS400")
|
||||||
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -pthread")
|
||||||
|
set(CMAKE_CXX_ARCHIVE_CREATE "<CMAKE_AR> -X64 qc <TARGET> <OBJECTS>")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if (USE_NCCL)
|
||||||
|
find_package(Nccl REQUIRED)
|
||||||
|
endif (USE_NCCL)
|
||||||
|
|
||||||
# dmlc-core
|
# dmlc-core
|
||||||
msvc_use_static_runtime()
|
msvc_use_static_runtime()
|
||||||
|
if (FORCE_SHARED_CRT)
|
||||||
|
set(DMLC_FORCE_SHARED_CRT ON)
|
||||||
|
endif ()
|
||||||
add_subdirectory(${xgboost_SOURCE_DIR}/dmlc-core)
|
add_subdirectory(${xgboost_SOURCE_DIR}/dmlc-core)
|
||||||
set_target_properties(dmlc PROPERTIES
|
|
||||||
CXX_STANDARD 14
|
|
||||||
CXX_STANDARD_REQUIRED ON
|
|
||||||
POSITION_INDEPENDENT_CODE ON)
|
|
||||||
if (MSVC)
|
if (MSVC)
|
||||||
target_compile_options(dmlc PRIVATE
|
|
||||||
-D_CRT_SECURE_NO_WARNINGS -D_CRT_SECURE_NO_DEPRECATE)
|
|
||||||
if (TARGET dmlc_unit_tests)
|
if (TARGET dmlc_unit_tests)
|
||||||
target_compile_options(dmlc_unit_tests PRIVATE
|
target_compile_options(dmlc_unit_tests PRIVATE
|
||||||
-D_CRT_SECURE_NO_WARNINGS -D_CRT_SECURE_NO_DEPRECATE)
|
-D_CRT_SECURE_NO_WARNINGS -D_CRT_SECURE_NO_DEPRECATE)
|
||||||
endif (TARGET dmlc_unit_tests)
|
endif (TARGET dmlc_unit_tests)
|
||||||
endif (MSVC)
|
endif (MSVC)
|
||||||
if (ENABLE_ALL_WARNINGS)
|
|
||||||
target_compile_options(dmlc PRIVATE -Wall -Wextra)
|
|
||||||
endif (ENABLE_ALL_WARNINGS)
|
|
||||||
|
|
||||||
# rabit
|
# rabit
|
||||||
add_subdirectory(rabit)
|
add_subdirectory(rabit)
|
||||||
|
if (RABIT_BUILD_MPI)
|
||||||
|
find_package(MPI REQUIRED)
|
||||||
|
endif (RABIT_BUILD_MPI)
|
||||||
|
|
||||||
# core xgboost
|
# core xgboost
|
||||||
add_subdirectory(${xgboost_SOURCE_DIR}/src)
|
add_subdirectory(${xgboost_SOURCE_DIR}/src)
|
||||||
@@ -179,9 +238,27 @@ if (R_LIB)
|
|||||||
add_subdirectory(${xgboost_SOURCE_DIR}/R-package)
|
add_subdirectory(${xgboost_SOURCE_DIR}/R-package)
|
||||||
endif (R_LIB)
|
endif (R_LIB)
|
||||||
|
|
||||||
|
# This creates its own shared library `xgboost4j'.
|
||||||
|
if (JVM_BINDINGS)
|
||||||
|
add_subdirectory(${xgboost_SOURCE_DIR}/jvm-packages)
|
||||||
|
endif (JVM_BINDINGS)
|
||||||
|
|
||||||
# Plugin
|
# Plugin
|
||||||
add_subdirectory(${xgboost_SOURCE_DIR}/plugin)
|
add_subdirectory(${xgboost_SOURCE_DIR}/plugin)
|
||||||
|
|
||||||
|
if (PLUGIN_RMM)
|
||||||
|
find_package(rmm REQUIRED)
|
||||||
|
|
||||||
|
# Patch the rmm targets so they reference the static cudart
|
||||||
|
# Remove this patch once RMM stops specifying cudart requirement
|
||||||
|
# (since RMM is a header-only library, it should not specify cudart in its CMake config)
|
||||||
|
get_target_property(rmm_link_libs rmm::rmm INTERFACE_LINK_LIBRARIES)
|
||||||
|
list(REMOVE_ITEM rmm_link_libs CUDA::cudart)
|
||||||
|
list(APPEND rmm_link_libs CUDA::cudart_static)
|
||||||
|
set_target_properties(rmm::rmm PROPERTIES INTERFACE_LINK_LIBRARIES "${rmm_link_libs}")
|
||||||
|
get_target_property(rmm_link_libs rmm::rmm INTERFACE_LINK_LIBRARIES)
|
||||||
|
endif (PLUGIN_RMM)
|
||||||
|
|
||||||
#-- library
|
#-- library
|
||||||
if (BUILD_STATIC_LIB)
|
if (BUILD_STATIC_LIB)
|
||||||
add_library(xgboost STATIC)
|
add_library(xgboost STATIC)
|
||||||
@@ -189,50 +266,44 @@ else (BUILD_STATIC_LIB)
|
|||||||
add_library(xgboost SHARED)
|
add_library(xgboost SHARED)
|
||||||
endif (BUILD_STATIC_LIB)
|
endif (BUILD_STATIC_LIB)
|
||||||
target_link_libraries(xgboost PRIVATE objxgboost)
|
target_link_libraries(xgboost PRIVATE objxgboost)
|
||||||
|
|
||||||
if (USE_CUDA)
|
|
||||||
xgboost_set_cuda_flags(xgboost)
|
|
||||||
endif (USE_CUDA)
|
|
||||||
|
|
||||||
#-- Hide all C++ symbols
|
|
||||||
if (HIDE_CXX_SYMBOLS)
|
|
||||||
foreach(target objxgboost xgboost dmlc)
|
|
||||||
set_target_properties(${target} PROPERTIES CXX_VISIBILITY_PRESET hidden)
|
|
||||||
endforeach()
|
|
||||||
endif (HIDE_CXX_SYMBOLS)
|
|
||||||
|
|
||||||
target_include_directories(xgboost
|
target_include_directories(xgboost
|
||||||
INTERFACE
|
INTERFACE
|
||||||
$<INSTALL_INTERFACE:$<INSTALL_PREFIX>/include>
|
$<INSTALL_INTERFACE:$<INSTALL_PREFIX>/include>
|
||||||
$<BUILD_INTERFACE:${CMAKE_CURRENT_LIST_DIR}/include>)
|
$<BUILD_INTERFACE:${CMAKE_CURRENT_LIST_DIR}/include>)
|
||||||
|
|
||||||
# This creates its own shared library `xgboost4j'.
|
|
||||||
if (JVM_BINDINGS)
|
|
||||||
add_subdirectory(${xgboost_SOURCE_DIR}/jvm-packages)
|
|
||||||
endif (JVM_BINDINGS)
|
|
||||||
#-- End shared library
|
#-- End shared library
|
||||||
|
|
||||||
#-- CLI for xgboost
|
#-- CLI for xgboost
|
||||||
add_executable(runxgboost ${xgboost_SOURCE_DIR}/src/cli_main.cc)
|
add_executable(runxgboost ${xgboost_SOURCE_DIR}/src/cli_main.cc)
|
||||||
target_link_libraries(runxgboost PRIVATE objxgboost)
|
target_link_libraries(runxgboost PRIVATE objxgboost)
|
||||||
if (USE_NVTX)
|
|
||||||
enable_nvtx(runxgboost)
|
|
||||||
endif (USE_NVTX)
|
|
||||||
|
|
||||||
target_include_directories(runxgboost
|
target_include_directories(runxgboost
|
||||||
PRIVATE
|
PRIVATE
|
||||||
${xgboost_SOURCE_DIR}/include
|
${xgboost_SOURCE_DIR}/include
|
||||||
${xgboost_SOURCE_DIR}/dmlc-core/include
|
${xgboost_SOURCE_DIR}/dmlc-core/include
|
||||||
${xgboost_SOURCE_DIR}/rabit/include)
|
${xgboost_SOURCE_DIR}/rabit/include
|
||||||
set_target_properties(
|
)
|
||||||
runxgboost PROPERTIES
|
set_target_properties(runxgboost PROPERTIES OUTPUT_NAME xgboost)
|
||||||
OUTPUT_NAME xgboost
|
|
||||||
CXX_STANDARD 14
|
|
||||||
CXX_STANDARD_REQUIRED ON)
|
|
||||||
#-- End CLI for xgboost
|
#-- End CLI for xgboost
|
||||||
|
|
||||||
|
# Common setup for all targets
|
||||||
|
foreach(target xgboost objxgboost dmlc runxgboost)
|
||||||
|
xgboost_target_properties(${target})
|
||||||
|
xgboost_target_link_libraries(${target})
|
||||||
|
xgboost_target_defs(${target})
|
||||||
|
endforeach()
|
||||||
|
|
||||||
|
if (JVM_BINDINGS)
|
||||||
|
xgboost_target_properties(xgboost4j)
|
||||||
|
xgboost_target_link_libraries(xgboost4j)
|
||||||
|
xgboost_target_defs(xgboost4j)
|
||||||
|
endif (JVM_BINDINGS)
|
||||||
|
|
||||||
|
if (KEEP_BUILD_ARTIFACTS_IN_BINARY_DIR)
|
||||||
|
set_output_directory(runxgboost ${xgboost_BINARY_DIR})
|
||||||
|
set_output_directory(xgboost ${xgboost_BINARY_DIR}/lib)
|
||||||
|
else ()
|
||||||
set_output_directory(runxgboost ${xgboost_SOURCE_DIR})
|
set_output_directory(runxgboost ${xgboost_SOURCE_DIR})
|
||||||
set_output_directory(xgboost ${xgboost_SOURCE_DIR}/lib)
|
set_output_directory(xgboost ${xgboost_SOURCE_DIR}/lib)
|
||||||
|
endif ()
|
||||||
# Ensure these two targets do not build simultaneously, as they produce outputs with conflicting names
|
# Ensure these two targets do not build simultaneously, as they produce outputs with conflicting names
|
||||||
add_dependencies(xgboost runxgboost)
|
add_dependencies(xgboost runxgboost)
|
||||||
|
|
||||||
@@ -255,6 +326,8 @@ if (BUILD_C_DOC)
|
|||||||
run_doxygen()
|
run_doxygen()
|
||||||
endif (BUILD_C_DOC)
|
endif (BUILD_C_DOC)
|
||||||
|
|
||||||
|
include(CPack)
|
||||||
|
|
||||||
include(GNUInstallDirs)
|
include(GNUInstallDirs)
|
||||||
# Install all headers. Please note that currently the C++ headers does not form an "API".
|
# Install all headers. Please note that currently the C++ headers does not form an "API".
|
||||||
install(DIRECTORY ${xgboost_SOURCE_DIR}/include/xgboost
|
install(DIRECTORY ${xgboost_SOURCE_DIR}/include/xgboost
|
||||||
@@ -295,7 +368,7 @@ write_basic_package_version_file(
|
|||||||
COMPATIBILITY AnyNewerVersion)
|
COMPATIBILITY AnyNewerVersion)
|
||||||
install(
|
install(
|
||||||
FILES
|
FILES
|
||||||
${CMAKE_BINARY_DIR}/cmake/xgboost-config.cmake
|
${CMAKE_CURRENT_BINARY_DIR}/cmake/xgboost-config.cmake
|
||||||
${CMAKE_BINARY_DIR}/cmake/xgboost-config-version.cmake
|
${CMAKE_BINARY_DIR}/cmake/xgboost-config-version.cmake
|
||||||
DESTINATION ${CMAKE_INSTALL_LIBDIR}/cmake/xgboost)
|
DESTINATION ${CMAKE_INSTALL_LIBDIR}/cmake/xgboost)
|
||||||
|
|
||||||
@@ -303,12 +376,18 @@ install(
|
|||||||
if (GOOGLE_TEST)
|
if (GOOGLE_TEST)
|
||||||
enable_testing()
|
enable_testing()
|
||||||
# Unittests.
|
# Unittests.
|
||||||
|
add_executable(testxgboost)
|
||||||
|
target_link_libraries(testxgboost PRIVATE objxgboost)
|
||||||
|
xgboost_target_properties(testxgboost)
|
||||||
|
xgboost_target_link_libraries(testxgboost)
|
||||||
|
xgboost_target_defs(testxgboost)
|
||||||
|
|
||||||
add_subdirectory(${xgboost_SOURCE_DIR}/tests/cpp)
|
add_subdirectory(${xgboost_SOURCE_DIR}/tests/cpp)
|
||||||
|
|
||||||
add_test(
|
add_test(
|
||||||
NAME TestXGBoostLib
|
NAME TestXGBoostLib
|
||||||
COMMAND testxgboost
|
COMMAND testxgboost
|
||||||
WORKING_DIRECTORY ${xgboost_BINARY_DIR})
|
WORKING_DIRECTORY ${xgboost_BINARY_DIR})
|
||||||
|
|
||||||
# CLI tests
|
# CLI tests
|
||||||
configure_file(
|
configure_file(
|
||||||
${xgboost_SOURCE_DIR}/tests/cli/machine.conf.in
|
${xgboost_SOURCE_DIR}/tests/cli/machine.conf.in
|
||||||
|
|||||||
@@ -10,8 +10,8 @@ The Project Management Committee(PMC) consists group of active committers that m
|
|||||||
- Tianqi is a Ph.D. student working on large-scale machine learning. He is the creator of the project.
|
- Tianqi is a Ph.D. student working on large-scale machine learning. He is the creator of the project.
|
||||||
* [Michael Benesty](https://github.com/pommedeterresautee)
|
* [Michael Benesty](https://github.com/pommedeterresautee)
|
||||||
- Michael is a lawyer and data scientist in France. He is the creator of XGBoost interactive analysis module in R.
|
- Michael is a lawyer and data scientist in France. He is the creator of XGBoost interactive analysis module in R.
|
||||||
* [Yuan Tang](https://github.com/terrytangyuan), Ant Group
|
* [Yuan Tang](https://github.com/terrytangyuan), Akuity
|
||||||
- Yuan is a software engineer in Ant Group. He contributed mostly in R and Python packages.
|
- Yuan is a founding engineer at Akuity. He contributed mostly in R and Python packages.
|
||||||
* [Nan Zhu](https://github.com/CodingCat), Uber
|
* [Nan Zhu](https://github.com/CodingCat), Uber
|
||||||
- Nan is a software engineer in Uber. He contributed mostly in JVM packages.
|
- Nan is a software engineer in Uber. He contributed mostly in JVM packages.
|
||||||
* [Jiaming Yuan](https://github.com/trivialfis)
|
* [Jiaming Yuan](https://github.com/trivialfis)
|
||||||
@@ -59,7 +59,7 @@ List of Contributors
|
|||||||
* [Skipper Seabold](https://github.com/jseabold)
|
* [Skipper Seabold](https://github.com/jseabold)
|
||||||
- Skipper is the major contributor to the scikit-learn module of XGBoost.
|
- Skipper is the major contributor to the scikit-learn module of XGBoost.
|
||||||
* [Zygmunt Zając](https://github.com/zygmuntz)
|
* [Zygmunt Zając](https://github.com/zygmuntz)
|
||||||
- Zygmunt is the master behind the early stopping feature frequently used by kagglers.
|
- Zygmunt is the master behind the early stopping feature frequently used by Kagglers.
|
||||||
* [Ajinkya Kale](https://github.com/ajkl)
|
* [Ajinkya Kale](https://github.com/ajkl)
|
||||||
* [Boliang Chen](https://github.com/cblsjtu)
|
* [Boliang Chen](https://github.com/cblsjtu)
|
||||||
* [Yangqing Men](https://github.com/yanqingmen)
|
* [Yangqing Men](https://github.com/yanqingmen)
|
||||||
@@ -91,7 +91,7 @@ List of Contributors
|
|||||||
* [Henry Gouk](https://github.com/henrygouk)
|
* [Henry Gouk](https://github.com/henrygouk)
|
||||||
* [Pierre de Sahb](https://github.com/pdesahb)
|
* [Pierre de Sahb](https://github.com/pdesahb)
|
||||||
* [liuliang01](https://github.com/liuliang01)
|
* [liuliang01](https://github.com/liuliang01)
|
||||||
- liuliang01 added support for the qid column for LibSVM input format. This makes ranking task easier in distributed setting.
|
- liuliang01 added support for the qid column for LIBSVM input format. This makes ranking task easier in distributed setting.
|
||||||
* [Andrew Thia](https://github.com/BlueTea88)
|
* [Andrew Thia](https://github.com/BlueTea88)
|
||||||
- Andrew Thia implemented feature interaction constraints
|
- Andrew Thia implemented feature interaction constraints
|
||||||
* [Wei Tian](https://github.com/weitian)
|
* [Wei Tian](https://github.com/weitian)
|
||||||
|
|||||||
456
Jenkinsfile
vendored
456
Jenkinsfile
vendored
@@ -1,456 +0,0 @@
|
|||||||
#!/usr/bin/groovy
|
|
||||||
// -*- mode: groovy -*-
|
|
||||||
// Jenkins pipeline
|
|
||||||
// See documents at https://jenkins.io/doc/book/pipeline/jenkinsfile/
|
|
||||||
|
|
||||||
// Command to run command inside a docker container
|
|
||||||
dockerRun = 'tests/ci_build/ci_build.sh'
|
|
||||||
|
|
||||||
// Which CUDA version to use when building reference distribution wheel
|
|
||||||
ref_cuda_ver = '10.0'
|
|
||||||
|
|
||||||
import groovy.transform.Field
|
|
||||||
|
|
||||||
@Field
|
|
||||||
def commit_id // necessary to pass a variable from one stage to another
|
|
||||||
|
|
||||||
pipeline {
|
|
||||||
// Each stage specify its own agent
|
|
||||||
agent none
|
|
||||||
|
|
||||||
environment {
|
|
||||||
DOCKER_CACHE_ECR_ID = '492475357299'
|
|
||||||
DOCKER_CACHE_ECR_REGION = 'us-west-2'
|
|
||||||
}
|
|
||||||
|
|
||||||
// Setup common job properties
|
|
||||||
options {
|
|
||||||
ansiColor('xterm')
|
|
||||||
timestamps()
|
|
||||||
timeout(time: 240, unit: 'MINUTES')
|
|
||||||
buildDiscarder(logRotator(numToKeepStr: '10'))
|
|
||||||
preserveStashes()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build stages
|
|
||||||
stages {
|
|
||||||
stage('Jenkins Linux: Initialize') {
|
|
||||||
agent { label 'job_initializer' }
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
def buildNumber = env.BUILD_NUMBER as int
|
|
||||||
if (buildNumber > 1) milestone(buildNumber - 1)
|
|
||||||
milestone(buildNumber)
|
|
||||||
|
|
||||||
checkoutSrcs()
|
|
||||||
commit_id = "${GIT_COMMIT}"
|
|
||||||
}
|
|
||||||
sh 'python3 tests/jenkins_get_approval.py'
|
|
||||||
stash name: 'srcs'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stage('Jenkins Linux: Build') {
|
|
||||||
agent none
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
parallel ([
|
|
||||||
'clang-tidy': { ClangTidy() },
|
|
||||||
'build-cpu': { BuildCPU() },
|
|
||||||
'build-cpu-arm64': { BuildCPUARM64() },
|
|
||||||
'build-cpu-rabit-mock': { BuildCPUMock() },
|
|
||||||
// Build reference, distribution-ready Python wheel with CUDA 10.0
|
|
||||||
// using CentOS 6 image
|
|
||||||
'build-gpu-cuda10.0': { BuildCUDA(cuda_version: '10.0') },
|
|
||||||
// The build-gpu-* builds below use Ubuntu image
|
|
||||||
'build-gpu-cuda10.1': { BuildCUDA(cuda_version: '10.1') },
|
|
||||||
'build-gpu-cuda10.2': { BuildCUDA(cuda_version: '10.2', build_rmm: true) },
|
|
||||||
'build-gpu-cuda11.0': { BuildCUDA(cuda_version: '11.0') },
|
|
||||||
'build-gpu-rpkg': { BuildRPackageWithCUDA(cuda_version: '10.0') },
|
|
||||||
'build-jvm-packages-gpu-cuda10.0': { BuildJVMPackagesWithCUDA(spark_version: '3.0.0', cuda_version: '10.0') },
|
|
||||||
'build-jvm-packages': { BuildJVMPackages(spark_version: '3.0.0') },
|
|
||||||
'build-jvm-doc': { BuildJVMDoc() }
|
|
||||||
])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stage('Jenkins Linux: Test') {
|
|
||||||
agent none
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
parallel ([
|
|
||||||
'test-python-cpu': { TestPythonCPU() },
|
|
||||||
'test-python-cpu-arm64': { TestPythonCPUARM64() },
|
|
||||||
// artifact_cuda_version doesn't apply to RMM tests; RMM tests will always match CUDA version between artifact and host env
|
|
||||||
'test-python-gpu-cuda10.2': { TestPythonGPU(artifact_cuda_version: '10.0', host_cuda_version: '10.2', test_rmm: true) },
|
|
||||||
'test-python-gpu-cuda11.0-cross': { TestPythonGPU(artifact_cuda_version: '10.0', host_cuda_version: '11.0') },
|
|
||||||
'test-python-gpu-cuda11.0': { TestPythonGPU(artifact_cuda_version: '11.0', host_cuda_version: '11.0') },
|
|
||||||
'test-python-mgpu-cuda10.2': { TestPythonGPU(artifact_cuda_version: '10.0', host_cuda_version: '10.2', multi_gpu: true, test_rmm: true) },
|
|
||||||
'test-cpp-gpu-cuda10.2': { TestCppGPU(artifact_cuda_version: '10.2', host_cuda_version: '10.2', test_rmm: true) },
|
|
||||||
'test-cpp-gpu-cuda11.0': { TestCppGPU(artifact_cuda_version: '11.0', host_cuda_version: '11.0') },
|
|
||||||
'test-jvm-jdk8': { CrossTestJVMwithJDK(jdk_version: '8', spark_version: '3.0.0') },
|
|
||||||
'test-jvm-jdk11': { CrossTestJVMwithJDK(jdk_version: '11') },
|
|
||||||
'test-jvm-jdk12': { CrossTestJVMwithJDK(jdk_version: '12') }
|
|
||||||
])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stage('Jenkins Linux: Deploy') {
|
|
||||||
agent none
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
parallel ([
|
|
||||||
'deploy-jvm-packages': { DeployJVMPackages(spark_version: '3.0.0') }
|
|
||||||
])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// check out source code from git
|
|
||||||
def checkoutSrcs() {
|
|
||||||
retry(5) {
|
|
||||||
try {
|
|
||||||
timeout(time: 2, unit: 'MINUTES') {
|
|
||||||
checkout scm
|
|
||||||
sh 'git submodule update --init'
|
|
||||||
}
|
|
||||||
} catch (exc) {
|
|
||||||
deleteDir()
|
|
||||||
error "Failed to fetch source codes"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def GetCUDABuildContainerType(cuda_version) {
|
|
||||||
return (cuda_version == ref_cuda_ver) ? 'gpu_build_centos6' : 'gpu_build'
|
|
||||||
}
|
|
||||||
|
|
||||||
def ClangTidy() {
|
|
||||||
node('linux && cpu_build') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Running clang-tidy job..."
|
|
||||||
def container_type = "clang_tidy"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
def dockerArgs = "--build-arg CUDA_VERSION_ARG=10.1"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${dockerArgs} python3 tests/ci_build/tidy.py
|
|
||||||
"""
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildCPU() {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build CPU"
|
|
||||||
def container_type = "cpu"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} rm -fv dmlc-core/include/dmlc/build_config_default.h
|
|
||||||
# This step is not necessary, but here we include it, to ensure that DMLC_CORE_USE_CMAKE flag is correctly propagated
|
|
||||||
# We want to make sure that we use the configured header build/dmlc/build_config.h instead of include/dmlc/build_config_default.h.
|
|
||||||
# See discussion at https://github.com/dmlc/xgboost/issues/5510
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_via_cmake.sh -DPLUGIN_LZ4=ON -DPLUGIN_DENSE_PARSER=ON
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} bash -c "cd build && ctest --extra-verbose"
|
|
||||||
"""
|
|
||||||
// Sanitizer test
|
|
||||||
def docker_extra_params = "CI_DOCKER_EXTRA_PARAMS_INIT='-e ASAN_SYMBOLIZER_PATH=/usr/bin/llvm-symbolizer -e ASAN_OPTIONS=symbolize=1 -e UBSAN_OPTIONS=print_stacktrace=1:log_path=ubsan_error.log --cap-add SYS_PTRACE'"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_via_cmake.sh -DUSE_SANITIZER=ON -DENABLED_SANITIZERS="address;leak;undefined" \
|
|
||||||
-DCMAKE_BUILD_TYPE=Debug -DSANITIZER_PATH=/usr/lib/x86_64-linux-gnu/
|
|
||||||
${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} bash -c "cd build && ctest --exclude-regex AllTestsInDMLCUnitTests --extra-verbose"
|
|
||||||
"""
|
|
||||||
|
|
||||||
stash name: 'xgboost_cli', includes: 'xgboost'
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildCPUARM64() {
|
|
||||||
node('linux && arm64') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build CPU ARM64"
|
|
||||||
def container_type = "aarch64"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
def wheel_tag = "manylinux2014_aarch64"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_via_cmake.sh --conda-env=aarch64_test -DOPEN_MP:BOOL=ON -DHIDE_CXX_SYMBOL=ON
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} bash -c "cd build && ctest --extra-verbose"
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} bash -c "cd python-package && rm -rf dist/* && python setup.py bdist_wheel --universal"
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} python tests/ci_build/rename_whl.py python-package/dist/*.whl ${commit_id} ${wheel_tag}
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} auditwheel repair --plat ${wheel_tag} python-package/dist/*.whl
|
|
||||||
mv -v wheelhouse/*.whl python-package/dist/
|
|
||||||
# Make sure that libgomp.so is vendored in the wheel
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} bash -c "unzip -l python-package/dist/*.whl | grep libgomp || exit -1"
|
|
||||||
"""
|
|
||||||
echo 'Stashing Python wheel...'
|
|
||||||
stash name: "xgboost_whl_arm64_cpu", includes: 'python-package/dist/*.whl'
|
|
||||||
if (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release')) {
|
|
||||||
echo 'Uploading Python wheel...'
|
|
||||||
path = ("${BRANCH_NAME}" == 'master') ? '' : "${BRANCH_NAME}/"
|
|
||||||
s3Upload bucket: 'xgboost-nightly-builds', path: path, acl: 'PublicRead', workingDir: 'python-package/dist', includePathPattern:'**/*.whl'
|
|
||||||
}
|
|
||||||
stash name: 'xgboost_cli_arm64', includes: 'xgboost'
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildCPUMock() {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build CPU with rabit mock"
|
|
||||||
def container_type = "cpu"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_mock_cmake.sh
|
|
||||||
"""
|
|
||||||
echo 'Stashing rabit C++ test executable (xgboost)...'
|
|
||||||
stash name: 'xgboost_rabit_tests', includes: 'xgboost'
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildCUDA(args) {
|
|
||||||
node('linux && cpu_build') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build with CUDA ${args.cuda_version}"
|
|
||||||
def container_type = GetCUDABuildContainerType(args.cuda_version)
|
|
||||||
def docker_binary = "docker"
|
|
||||||
def docker_args = "--build-arg CUDA_VERSION_ARG=${args.cuda_version}"
|
|
||||||
def arch_flag = ""
|
|
||||||
if (env.BRANCH_NAME != 'master' && !(env.BRANCH_NAME.startsWith('release'))) {
|
|
||||||
arch_flag = "-DGPU_COMPUTE_VER=75"
|
|
||||||
}
|
|
||||||
def wheel_tag = "manylinux2010_x86_64"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/build_via_cmake.sh -DUSE_CUDA=ON -DUSE_NCCL=ON -DOPEN_MP:BOOL=ON -DHIDE_CXX_SYMBOLS=ON ${arch_flag}
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} bash -c "cd python-package && rm -rf dist/* && python setup.py bdist_wheel --universal"
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} python tests/ci_build/rename_whl.py python-package/dist/*.whl ${commit_id} ${wheel_tag}
|
|
||||||
"""
|
|
||||||
if (args.cuda_version == ref_cuda_ver) {
|
|
||||||
sh """
|
|
||||||
${dockerRun} auditwheel_x86_64 ${docker_binary} auditwheel repair --plat ${wheel_tag} python-package/dist/*.whl
|
|
||||||
mv -v wheelhouse/*.whl python-package/dist/
|
|
||||||
# Make sure that libgomp.so is vendored in the wheel
|
|
||||||
${dockerRun} auditwheel_x86_64 ${docker_binary} bash -c "unzip -l python-package/dist/*.whl | grep libgomp || exit -1"
|
|
||||||
"""
|
|
||||||
}
|
|
||||||
echo 'Stashing Python wheel...'
|
|
||||||
stash name: "xgboost_whl_cuda${args.cuda_version}", includes: 'python-package/dist/*.whl'
|
|
||||||
if (args.cuda_version == ref_cuda_ver && (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release'))) {
|
|
||||||
echo 'Uploading Python wheel...'
|
|
||||||
path = ("${BRANCH_NAME}" == 'master') ? '' : "${BRANCH_NAME}/"
|
|
||||||
s3Upload bucket: 'xgboost-nightly-builds', path: path, acl: 'PublicRead', workingDir: 'python-package/dist', includePathPattern:'**/*.whl'
|
|
||||||
}
|
|
||||||
echo 'Stashing C++ test executable (testxgboost)...'
|
|
||||||
stash name: "xgboost_cpp_tests_cuda${args.cuda_version}", includes: 'build/testxgboost'
|
|
||||||
if (args.build_rmm) {
|
|
||||||
echo "Build with CUDA ${args.cuda_version} and RMM"
|
|
||||||
container_type = "rmm"
|
|
||||||
docker_binary = "docker"
|
|
||||||
docker_args = "--build-arg CUDA_VERSION_ARG=${args.cuda_version}"
|
|
||||||
sh """
|
|
||||||
rm -rf build/
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/build_via_cmake.sh --conda-env=gpu_test -DUSE_CUDA=ON -DUSE_NCCL=ON -DPLUGIN_RMM=ON ${arch_flag}
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} bash -c "cd python-package && rm -rf dist/* && python setup.py bdist_wheel --universal"
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} python tests/ci_build/rename_whl.py python-package/dist/*.whl ${commit_id} manylinux2010_x86_64
|
|
||||||
"""
|
|
||||||
echo 'Stashing Python wheel...'
|
|
||||||
stash name: "xgboost_whl_rmm_cuda${args.cuda_version}", includes: 'python-package/dist/*.whl'
|
|
||||||
echo 'Stashing C++ test executable (testxgboost)...'
|
|
||||||
stash name: "xgboost_cpp_tests_rmm_cuda${args.cuda_version}", includes: 'build/testxgboost'
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildRPackageWithCUDA(args) {
|
|
||||||
node('linux && cpu_build') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
def container_type = 'gpu_build_r_centos6'
|
|
||||||
def docker_binary = "docker"
|
|
||||||
def docker_args = "--build-arg CUDA_VERSION_ARG=10.0"
|
|
||||||
if (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release')) {
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/build_r_pkg_with_cuda.sh ${commit_id}
|
|
||||||
"""
|
|
||||||
echo 'Uploading R tarball...'
|
|
||||||
path = ("${BRANCH_NAME}" == 'master') ? '' : "${BRANCH_NAME}/"
|
|
||||||
s3Upload bucket: 'xgboost-nightly-builds', path: path, acl: 'PublicRead', includePathPattern:'xgboost_r_gpu_linux_*.tar.gz'
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildJVMPackagesWithCUDA(args) {
|
|
||||||
node('linux && mgpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build XGBoost4J-Spark with Spark ${args.spark_version}, CUDA ${args.cuda_version}"
|
|
||||||
def container_type = "jvm_gpu_build"
|
|
||||||
def docker_binary = "nvidia-docker"
|
|
||||||
def docker_args = "--build-arg CUDA_VERSION_ARG=${args.cuda_version}"
|
|
||||||
def arch_flag = ""
|
|
||||||
if (env.BRANCH_NAME != 'master' && !(env.BRANCH_NAME.startsWith('release'))) {
|
|
||||||
arch_flag = "-DGPU_COMPUTE_VER=75"
|
|
||||||
}
|
|
||||||
// Use only 4 CPU cores
|
|
||||||
def docker_extra_params = "CI_DOCKER_EXTRA_PARAMS_INIT='--cpuset-cpus 0-3'"
|
|
||||||
sh """
|
|
||||||
${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/build_jvm_packages.sh ${args.spark_version} -Duse.cuda=ON $arch_flag
|
|
||||||
"""
|
|
||||||
echo "Stashing XGBoost4J JAR with CUDA ${args.cuda_version} ..."
|
|
||||||
stash name: 'xgboost4j_jar_gpu', includes: "jvm-packages/xgboost4j-gpu/target/*.jar,jvm-packages/xgboost4j-spark-gpu/target/*.jar"
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildJVMPackages(args) {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Build XGBoost4J-Spark with Spark ${args.spark_version}"
|
|
||||||
def container_type = "jvm"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
// Use only 4 CPU cores
|
|
||||||
def docker_extra_params = "CI_DOCKER_EXTRA_PARAMS_INIT='--cpuset-cpus 0-3'"
|
|
||||||
sh """
|
|
||||||
${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_jvm_packages.sh ${args.spark_version}
|
|
||||||
"""
|
|
||||||
echo 'Stashing XGBoost4J JAR...'
|
|
||||||
stash name: 'xgboost4j_jar', includes: "jvm-packages/xgboost4j/target/*.jar,jvm-packages/xgboost4j-spark/target/*.jar,jvm-packages/xgboost4j-example/target/*.jar"
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildJVMDoc() {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Building JVM doc..."
|
|
||||||
def container_type = "jvm"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/build_jvm_doc.sh ${BRANCH_NAME}
|
|
||||||
"""
|
|
||||||
if (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release')) {
|
|
||||||
echo 'Uploading doc...'
|
|
||||||
s3Upload file: "jvm-packages/${BRANCH_NAME}.tar.bz2", bucket: 'xgboost-docs', acl: 'PublicRead', path: "${BRANCH_NAME}.tar.bz2"
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def TestPythonCPU() {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: "xgboost_whl_cuda${ref_cuda_ver}"
|
|
||||||
unstash name: 'srcs'
|
|
||||||
unstash name: 'xgboost_cli'
|
|
||||||
echo "Test Python CPU"
|
|
||||||
def container_type = "cpu"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/test_python.sh cpu
|
|
||||||
"""
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def TestPythonCPUARM64() {
|
|
||||||
node('linux && arm64') {
|
|
||||||
unstash name: "xgboost_whl_arm64_cpu"
|
|
||||||
unstash name: 'srcs'
|
|
||||||
unstash name: 'xgboost_cli_arm64'
|
|
||||||
echo "Test Python CPU ARM64"
|
|
||||||
def container_type = "aarch64"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} tests/ci_build/test_python.sh cpu-arm64
|
|
||||||
"""
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def TestPythonGPU(args) {
|
|
||||||
def nodeReq = (args.multi_gpu) ? 'linux && mgpu' : 'linux && gpu'
|
|
||||||
def artifact_cuda_version = (args.artifact_cuda_version) ?: ref_cuda_ver
|
|
||||||
node(nodeReq) {
|
|
||||||
unstash name: "xgboost_whl_cuda${artifact_cuda_version}"
|
|
||||||
unstash name: "xgboost_cpp_tests_cuda${artifact_cuda_version}"
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Test Python GPU: CUDA ${args.host_cuda_version}"
|
|
||||||
def container_type = "gpu"
|
|
||||||
def docker_binary = "nvidia-docker"
|
|
||||||
def docker_args = "--build-arg CUDA_VERSION_ARG=${args.host_cuda_version}"
|
|
||||||
def mgpu_indicator = (args.multi_gpu) ? 'mgpu' : 'gpu'
|
|
||||||
// Allocate extra space in /dev/shm to enable NCCL
|
|
||||||
def docker_extra_params = (args.multi_gpu) ? "CI_DOCKER_EXTRA_PARAMS_INIT='--shm-size=4g'" : ''
|
|
||||||
sh "${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/test_python.sh ${mgpu_indicator}"
|
|
||||||
if (args.test_rmm) {
|
|
||||||
sh "rm -rfv build/ python-package/dist/"
|
|
||||||
unstash name: "xgboost_whl_rmm_cuda${args.host_cuda_version}"
|
|
||||||
unstash name: "xgboost_cpp_tests_rmm_cuda${args.host_cuda_version}"
|
|
||||||
sh "${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/test_python.sh ${mgpu_indicator} --use-rmm-pool"
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def TestCppGPU(args) {
|
|
||||||
def nodeReq = 'linux && mgpu'
|
|
||||||
def artifact_cuda_version = (args.artifact_cuda_version) ?: ref_cuda_ver
|
|
||||||
node(nodeReq) {
|
|
||||||
unstash name: "xgboost_cpp_tests_cuda${artifact_cuda_version}"
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Test C++, CUDA ${args.host_cuda_version}"
|
|
||||||
def container_type = "gpu"
|
|
||||||
def docker_binary = "nvidia-docker"
|
|
||||||
def docker_args = "--build-arg CUDA_VERSION_ARG=${args.host_cuda_version}"
|
|
||||||
sh "${dockerRun} ${container_type} ${docker_binary} ${docker_args} build/testxgboost"
|
|
||||||
if (args.test_rmm) {
|
|
||||||
sh "rm -rfv build/"
|
|
||||||
unstash name: "xgboost_cpp_tests_rmm_cuda${args.host_cuda_version}"
|
|
||||||
echo "Test C++, CUDA ${args.host_cuda_version} with RMM"
|
|
||||||
container_type = "rmm"
|
|
||||||
docker_binary = "nvidia-docker"
|
|
||||||
docker_args = "--build-arg CUDA_VERSION_ARG=${args.host_cuda_version}"
|
|
||||||
sh """
|
|
||||||
${dockerRun} ${container_type} ${docker_binary} ${docker_args} bash -c "source activate gpu_test && build/testxgboost --use-rmm-pool --gtest_filter=-*DeathTest.*"
|
|
||||||
"""
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def CrossTestJVMwithJDK(args) {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'xgboost4j_jar'
|
|
||||||
unstash name: 'srcs'
|
|
||||||
if (args.spark_version != null) {
|
|
||||||
echo "Test XGBoost4J on a machine with JDK ${args.jdk_version}, Spark ${args.spark_version}"
|
|
||||||
} else {
|
|
||||||
echo "Test XGBoost4J on a machine with JDK ${args.jdk_version}"
|
|
||||||
}
|
|
||||||
def container_type = "jvm_cross"
|
|
||||||
def docker_binary = "docker"
|
|
||||||
def spark_arg = (args.spark_version != null) ? "--build-arg SPARK_VERSION=${args.spark_version}" : ""
|
|
||||||
def docker_args = "--build-arg JDK_VERSION=${args.jdk_version} ${spark_arg}"
|
|
||||||
// Run integration tests only when spark_version is given
|
|
||||||
def docker_extra_params = (args.spark_version != null) ? "CI_DOCKER_EXTRA_PARAMS_INIT='-e RUN_INTEGRATION_TEST=1'" : ""
|
|
||||||
sh """
|
|
||||||
${docker_extra_params} ${dockerRun} ${container_type} ${docker_binary} ${docker_args} tests/ci_build/test_jvm_cross.sh
|
|
||||||
"""
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def DeployJVMPackages(args) {
|
|
||||||
node('linux && cpu') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
if (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release')) {
|
|
||||||
echo 'Deploying to xgboost-maven-repo S3 repo...'
|
|
||||||
sh """
|
|
||||||
${dockerRun} jvm_gpu_build docker --build-arg CUDA_VERSION_ARG=10.0 tests/ci_build/deploy_jvm_packages.sh ${args.spark_version}
|
|
||||||
"""
|
|
||||||
}
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,143 +0,0 @@
|
|||||||
#!/usr/bin/groovy
|
|
||||||
// -*- mode: groovy -*-
|
|
||||||
|
|
||||||
/* Jenkins pipeline for Windows AMD64 target */
|
|
||||||
|
|
||||||
import groovy.transform.Field
|
|
||||||
|
|
||||||
@Field
|
|
||||||
def commit_id // necessary to pass a variable from one stage to another
|
|
||||||
|
|
||||||
pipeline {
|
|
||||||
agent none
|
|
||||||
|
|
||||||
// Setup common job properties
|
|
||||||
options {
|
|
||||||
timestamps()
|
|
||||||
timeout(time: 240, unit: 'MINUTES')
|
|
||||||
buildDiscarder(logRotator(numToKeepStr: '10'))
|
|
||||||
preserveStashes()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build stages
|
|
||||||
stages {
|
|
||||||
stage('Jenkins Win64: Initialize') {
|
|
||||||
agent { label 'job_initializer' }
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
def buildNumber = env.BUILD_NUMBER as int
|
|
||||||
if (buildNumber > 1) milestone(buildNumber - 1)
|
|
||||||
milestone(buildNumber)
|
|
||||||
checkoutSrcs()
|
|
||||||
commit_id = "${GIT_COMMIT}"
|
|
||||||
}
|
|
||||||
sh 'python3 tests/jenkins_get_approval.py'
|
|
||||||
stash name: 'srcs'
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stage('Jenkins Win64: Build') {
|
|
||||||
agent none
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
parallel ([
|
|
||||||
'build-win64-cuda10.1': { BuildWin64() }
|
|
||||||
])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stage('Jenkins Win64: Test') {
|
|
||||||
agent none
|
|
||||||
steps {
|
|
||||||
script {
|
|
||||||
parallel ([
|
|
||||||
'test-win64-cuda10.1': { TestWin64() },
|
|
||||||
])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// check out source code from git
|
|
||||||
def checkoutSrcs() {
|
|
||||||
retry(5) {
|
|
||||||
try {
|
|
||||||
timeout(time: 2, unit: 'MINUTES') {
|
|
||||||
checkout scm
|
|
||||||
sh 'git submodule update --init'
|
|
||||||
}
|
|
||||||
} catch (exc) {
|
|
||||||
deleteDir()
|
|
||||||
error "Failed to fetch source codes"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def BuildWin64() {
|
|
||||||
node('win64 && cuda10_unified') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
echo "Building XGBoost for Windows AMD64 target..."
|
|
||||||
bat "nvcc --version"
|
|
||||||
def arch_flag = ""
|
|
||||||
if (env.BRANCH_NAME != 'master' && !(env.BRANCH_NAME.startsWith('release'))) {
|
|
||||||
arch_flag = "-DGPU_COMPUTE_VER=75"
|
|
||||||
}
|
|
||||||
bat """
|
|
||||||
mkdir build
|
|
||||||
cd build
|
|
||||||
cmake .. -G"Visual Studio 15 2017 Win64" -DUSE_CUDA=ON -DCMAKE_VERBOSE_MAKEFILE=ON -DGOOGLE_TEST=ON -DUSE_DMLC_GTEST=ON ${arch_flag} -DCMAKE_UNITY_BUILD=ON
|
|
||||||
"""
|
|
||||||
bat """
|
|
||||||
cd build
|
|
||||||
"C:\\Program Files (x86)\\Microsoft Visual Studio\\2017\\Community\\MSBuild\\15.0\\Bin\\MSBuild.exe" xgboost.sln /m /p:Configuration=Release /nodeReuse:false
|
|
||||||
"""
|
|
||||||
bat """
|
|
||||||
cd python-package
|
|
||||||
conda activate && python setup.py bdist_wheel --universal && for /R %%i in (dist\\*.whl) DO python ../tests/ci_build/rename_whl.py "%%i" ${commit_id} win_amd64
|
|
||||||
"""
|
|
||||||
echo "Insert vcomp140.dll (OpenMP runtime) into the wheel..."
|
|
||||||
bat """
|
|
||||||
cd python-package\\dist
|
|
||||||
COPY /B ..\\..\\tests\\ci_build\\insert_vcomp140.py
|
|
||||||
conda activate && python insert_vcomp140.py *.whl
|
|
||||||
"""
|
|
||||||
echo 'Stashing Python wheel...'
|
|
||||||
stash name: 'xgboost_whl', includes: 'python-package/dist/*.whl'
|
|
||||||
if (env.BRANCH_NAME == 'master' || env.BRANCH_NAME.startsWith('release')) {
|
|
||||||
echo 'Uploading Python wheel...'
|
|
||||||
path = ("${BRANCH_NAME}" == 'master') ? '' : "${BRANCH_NAME}/"
|
|
||||||
s3Upload bucket: 'xgboost-nightly-builds', path: path, acl: 'PublicRead', workingDir: 'python-package/dist', includePathPattern:'**/*.whl'
|
|
||||||
}
|
|
||||||
echo 'Stashing C++ test executable (testxgboost)...'
|
|
||||||
stash name: 'xgboost_cpp_tests', includes: 'build/testxgboost.exe'
|
|
||||||
stash name: 'xgboost_cli', includes: 'xgboost.exe'
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
def TestWin64() {
|
|
||||||
node('win64 && cuda10_unified') {
|
|
||||||
unstash name: 'srcs'
|
|
||||||
unstash name: 'xgboost_whl'
|
|
||||||
unstash name: 'xgboost_cli'
|
|
||||||
unstash name: 'xgboost_cpp_tests'
|
|
||||||
echo "Test Win64"
|
|
||||||
bat "nvcc --version"
|
|
||||||
echo "Running C++ tests..."
|
|
||||||
bat "build\\testxgboost.exe"
|
|
||||||
echo "Installing Python dependencies..."
|
|
||||||
def env_name = 'win64_' + UUID.randomUUID().toString().replaceAll('-', '')
|
|
||||||
bat "conda env create -n ${env_name} --file=tests/ci_build/conda_env/win64_test.yml"
|
|
||||||
echo "Installing Python wheel..."
|
|
||||||
bat """
|
|
||||||
conda activate ${env_name} && for /R %%i in (python-package\\dist\\*.whl) DO python -m pip install "%%i"
|
|
||||||
"""
|
|
||||||
echo "Running Python tests..."
|
|
||||||
bat "conda activate ${env_name} && python -m pytest -v -s -rxXs --fulltrace tests\\python"
|
|
||||||
bat """
|
|
||||||
conda activate ${env_name} && python -m pytest -v -s -rxXs --fulltrace -m "(not slow) and (not mgpu)" tests\\python-gpu
|
|
||||||
"""
|
|
||||||
bat "conda env remove --name ${env_name}"
|
|
||||||
deleteDir()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
160
Makefile
160
Makefile
@@ -1,160 +0,0 @@
|
|||||||
ifndef DMLC_CORE
|
|
||||||
DMLC_CORE = dmlc-core
|
|
||||||
endif
|
|
||||||
|
|
||||||
ifndef RABIT
|
|
||||||
RABIT = rabit
|
|
||||||
endif
|
|
||||||
|
|
||||||
ROOTDIR = $(CURDIR)
|
|
||||||
|
|
||||||
# workarounds for some buggy old make & msys2 versions seen in windows
|
|
||||||
ifeq (NA, $(shell test ! -d "$(ROOTDIR)" && echo NA ))
|
|
||||||
$(warning Attempting to fix non-existing ROOTDIR [$(ROOTDIR)])
|
|
||||||
ROOTDIR := $(shell pwd)
|
|
||||||
$(warning New ROOTDIR [$(ROOTDIR)] $(shell test -d "$(ROOTDIR)" && echo " is OK" ))
|
|
||||||
endif
|
|
||||||
MAKE_OK := $(shell "$(MAKE)" -v 2> /dev/null)
|
|
||||||
ifndef MAKE_OK
|
|
||||||
$(warning Attempting to recover non-functional MAKE [$(MAKE)])
|
|
||||||
MAKE := $(shell which make 2> /dev/null)
|
|
||||||
MAKE_OK := $(shell "$(MAKE)" -v 2> /dev/null)
|
|
||||||
endif
|
|
||||||
$(warning MAKE [$(MAKE)] - $(if $(MAKE_OK),checked OK,PROBLEM))
|
|
||||||
|
|
||||||
include $(DMLC_CORE)/make/dmlc.mk
|
|
||||||
|
|
||||||
# set compiler defaults for OSX versus *nix
|
|
||||||
# let people override either
|
|
||||||
OS := $(shell uname)
|
|
||||||
ifeq ($(OS), Darwin)
|
|
||||||
ifndef CC
|
|
||||||
export CC = $(if $(shell which clang), clang, gcc)
|
|
||||||
endif
|
|
||||||
ifndef CXX
|
|
||||||
export CXX = $(if $(shell which clang++), clang++, g++)
|
|
||||||
endif
|
|
||||||
else
|
|
||||||
# linux defaults
|
|
||||||
ifndef CC
|
|
||||||
export CC = gcc
|
|
||||||
endif
|
|
||||||
ifndef CXX
|
|
||||||
export CXX = g++
|
|
||||||
endif
|
|
||||||
endif
|
|
||||||
|
|
||||||
export CFLAGS= -DDMLC_LOG_CUSTOMIZE=1 -std=c++14 -Wall -Wno-unknown-pragmas -Iinclude $(ADD_CFLAGS)
|
|
||||||
CFLAGS += -I$(DMLC_CORE)/include -I$(RABIT)/include -I$(GTEST_PATH)/include
|
|
||||||
|
|
||||||
ifeq ($(TEST_COVER), 1)
|
|
||||||
CFLAGS += -g -O0 -fprofile-arcs -ftest-coverage
|
|
||||||
else
|
|
||||||
CFLAGS += -O3 -funroll-loops
|
|
||||||
endif
|
|
||||||
|
|
||||||
ifndef LINT_LANG
|
|
||||||
LINT_LANG= "all"
|
|
||||||
endif
|
|
||||||
|
|
||||||
# specify tensor path
|
|
||||||
.PHONY: clean all lint clean_all doxygen rcpplint pypack Rpack Rbuild Rcheck
|
|
||||||
|
|
||||||
build/%.o: src/%.cc
|
|
||||||
@mkdir -p $(@D)
|
|
||||||
$(CXX) $(CFLAGS) -MM -MT build/$*.o $< >build/$*.d
|
|
||||||
$(CXX) -c $(CFLAGS) $< -o $@
|
|
||||||
|
|
||||||
# The should be equivalent to $(ALL_OBJ) except for build/cli_main.o
|
|
||||||
amalgamation/xgboost-all0.o: amalgamation/xgboost-all0.cc
|
|
||||||
$(CXX) -c $(CFLAGS) $< -o $@
|
|
||||||
|
|
||||||
rcpplint:
|
|
||||||
python3 dmlc-core/scripts/lint.py xgboost ${LINT_LANG} R-package/src
|
|
||||||
|
|
||||||
lint: rcpplint
|
|
||||||
python3 dmlc-core/scripts/lint.py --exclude_path python-package/xgboost/dmlc-core \
|
|
||||||
python-package/xgboost/include python-package/xgboost/lib \
|
|
||||||
python-package/xgboost/make python-package/xgboost/rabit \
|
|
||||||
python-package/xgboost/src --pylint-rc ${PWD}/python-package/.pylintrc xgboost \
|
|
||||||
${LINT_LANG} include src python-package
|
|
||||||
|
|
||||||
ifeq ($(TEST_COVER), 1)
|
|
||||||
cover: check
|
|
||||||
@- $(foreach COV_OBJ, $(COVER_OBJ), \
|
|
||||||
gcov -pbcul -o $(shell dirname $(COV_OBJ)) $(COV_OBJ) > gcov.log || cat gcov.log; \
|
|
||||||
)
|
|
||||||
endif
|
|
||||||
|
|
||||||
|
|
||||||
# dask is required to pass, others are not
|
|
||||||
# If any of the dask tests failed, contributor won't see the other error.
|
|
||||||
mypy:
|
|
||||||
cd python-package; \
|
|
||||||
mypy ./xgboost/dask.py ../tests/python/test_with_dask.py --follow-imports=silent; \
|
|
||||||
mypy ../tests/python-gpu/test_gpu_with_dask.py --follow-imports=silent; \
|
|
||||||
mypy . || true ;
|
|
||||||
|
|
||||||
clean:
|
|
||||||
$(RM) -rf build lib bin *~ */*~ */*/*~ */*/*/*~ */*.o */*/*.o */*/*/*.o #xgboost
|
|
||||||
$(RM) -rf build_tests *.gcov tests/cpp/xgboost_test
|
|
||||||
if [ -d "R-package/src" ]; then \
|
|
||||||
cd R-package/src; \
|
|
||||||
$(RM) -rf rabit src include dmlc-core amalgamation *.so *.dll; \
|
|
||||||
cd $(ROOTDIR); \
|
|
||||||
fi
|
|
||||||
|
|
||||||
clean_all: clean
|
|
||||||
cd $(DMLC_CORE); "$(MAKE)" clean; cd $(ROOTDIR)
|
|
||||||
cd $(RABIT); "$(MAKE)" clean; cd $(ROOTDIR)
|
|
||||||
|
|
||||||
# create pip source dist (sdist) pack for PyPI
|
|
||||||
pippack: clean_all
|
|
||||||
cd python-package; python setup.py sdist; mv dist/*.tar.gz ..; cd ..
|
|
||||||
|
|
||||||
# Script to make a clean installable R package.
|
|
||||||
Rpack: clean_all
|
|
||||||
rm -rf xgboost xgboost*.tar.gz
|
|
||||||
cp -r R-package xgboost
|
|
||||||
rm -rf xgboost/src/*.o xgboost/src/*.so xgboost/src/*.dll
|
|
||||||
rm -rf xgboost/src/*/*.o
|
|
||||||
rm -rf xgboost/demo/*.model xgboost/demo/*.buffer xgboost/demo/*.txt
|
|
||||||
rm -rf xgboost/demo/runall.R
|
|
||||||
cp -r src xgboost/src/src
|
|
||||||
cp -r include xgboost/src/include
|
|
||||||
cp -r amalgamation xgboost/src/amalgamation
|
|
||||||
mkdir -p xgboost/src/rabit
|
|
||||||
cp -r rabit/include xgboost/src/rabit/include
|
|
||||||
cp -r rabit/src xgboost/src/rabit/src
|
|
||||||
rm -rf xgboost/src/rabit/src/*.o
|
|
||||||
mkdir -p xgboost/src/dmlc-core
|
|
||||||
cp -r dmlc-core/include xgboost/src/dmlc-core/include
|
|
||||||
cp -r dmlc-core/src xgboost/src/dmlc-core/src
|
|
||||||
cp ./LICENSE xgboost
|
|
||||||
# Modify PKGROOT in Makevars.in
|
|
||||||
cat R-package/src/Makevars.in|sed '2s/.*/PKGROOT=./' > xgboost/src/Makevars.in
|
|
||||||
# Configure Makevars.win (Windows-specific Makevars, likely using MinGW)
|
|
||||||
cp xgboost/src/Makevars.in xgboost/src/Makevars.win
|
|
||||||
cat xgboost/src/Makevars.in| sed '3s/.*/ENABLE_STD_THREAD=0/' > xgboost/src/Makevars.win
|
|
||||||
sed -i -e 's/@OPENMP_CXXFLAGS@/$$\(SHLIB_OPENMP_CXXFLAGS\)/g' xgboost/src/Makevars.win
|
|
||||||
sed -i -e 's/-pthread/$$\(SHLIB_PTHREAD_FLAGS\)/g' xgboost/src/Makevars.win
|
|
||||||
sed -i -e 's/@ENDIAN_FLAG@/-DDMLC_CMAKE_LITTLE_ENDIAN=1/g' xgboost/src/Makevars.win
|
|
||||||
sed -i -e 's/@BACKTRACE_LIB@//g' xgboost/src/Makevars.win
|
|
||||||
sed -i -e 's/@OPENMP_LIB@//g' xgboost/src/Makevars.win
|
|
||||||
rm -f xgboost/src/Makevars.win-e # OSX sed create this extra file; remove it
|
|
||||||
bash R-package/remove_warning_suppression_pragma.sh
|
|
||||||
bash xgboost/remove_warning_suppression_pragma.sh
|
|
||||||
rm xgboost/remove_warning_suppression_pragma.sh
|
|
||||||
rm -rfv xgboost/tests/helper_scripts/
|
|
||||||
|
|
||||||
R ?= R
|
|
||||||
|
|
||||||
Rbuild: Rpack
|
|
||||||
$(R) CMD build xgboost
|
|
||||||
rm -rf xgboost
|
|
||||||
|
|
||||||
Rcheck: Rbuild
|
|
||||||
$(R) CMD check --as-cran xgboost*.tar.gz
|
|
||||||
|
|
||||||
-include build/*.d
|
|
||||||
-include build/*/*.d
|
|
||||||
@@ -16,7 +16,6 @@ target_compile_definitions(xgboost-r
|
|||||||
-DDMLC_LOG_BEFORE_THROW=0
|
-DDMLC_LOG_BEFORE_THROW=0
|
||||||
-DDMLC_DISABLE_STDIN=1
|
-DDMLC_DISABLE_STDIN=1
|
||||||
-DDMLC_LOG_CUSTOMIZE=1
|
-DDMLC_LOG_CUSTOMIZE=1
|
||||||
-DRABIT_CUSTOMIZE_MSG_
|
|
||||||
-DRABIT_STRICT_CXX98_)
|
-DRABIT_STRICT_CXX98_)
|
||||||
target_include_directories(xgboost-r
|
target_include_directories(xgboost-r
|
||||||
PRIVATE
|
PRIVATE
|
||||||
@@ -31,7 +30,7 @@ if (USE_OPENMP)
|
|||||||
endif (USE_OPENMP)
|
endif (USE_OPENMP)
|
||||||
set_target_properties(
|
set_target_properties(
|
||||||
xgboost-r PROPERTIES
|
xgboost-r PROPERTIES
|
||||||
CXX_STANDARD 14
|
CXX_STANDARD 17
|
||||||
CXX_STANDARD_REQUIRED ON
|
CXX_STANDARD_REQUIRED ON
|
||||||
POSITION_INDEPENDENT_CODE ON)
|
POSITION_INDEPENDENT_CODE ON)
|
||||||
|
|
||||||
|
|||||||
@@ -1,12 +1,12 @@
|
|||||||
Package: xgboost
|
Package: xgboost
|
||||||
Type: Package
|
Type: Package
|
||||||
Title: Extreme Gradient Boosting
|
Title: Extreme Gradient Boosting
|
||||||
Version: 1.4.0.1
|
Version: 2.0.0.1
|
||||||
Date: 2020-08-28
|
Date: 2023-09-11
|
||||||
Authors@R: c(
|
Authors@R: c(
|
||||||
person("Tianqi", "Chen", role = c("aut"),
|
person("Tianqi", "Chen", role = c("aut"),
|
||||||
email = "tianqi.tchen@gmail.com"),
|
email = "tianqi.tchen@gmail.com"),
|
||||||
person("Tong", "He", role = c("aut", "cre"),
|
person("Tong", "He", role = c("aut"),
|
||||||
email = "hetong007@gmail.com"),
|
email = "hetong007@gmail.com"),
|
||||||
person("Michael", "Benesty", role = c("aut"),
|
person("Michael", "Benesty", role = c("aut"),
|
||||||
email = "michael@benesty.fr"),
|
email = "michael@benesty.fr"),
|
||||||
@@ -26,9 +26,12 @@ Authors@R: c(
|
|||||||
person("Min", "Lin", role = c("aut")),
|
person("Min", "Lin", role = c("aut")),
|
||||||
person("Yifeng", "Geng", role = c("aut")),
|
person("Yifeng", "Geng", role = c("aut")),
|
||||||
person("Yutian", "Li", role = c("aut")),
|
person("Yutian", "Li", role = c("aut")),
|
||||||
|
person("Jiaming", "Yuan", role = c("aut", "cre"),
|
||||||
|
email = "jm.yuan@outlook.com"),
|
||||||
person("XGBoost contributors", role = c("cph"),
|
person("XGBoost contributors", role = c("cph"),
|
||||||
comment = "base XGBoost implementation")
|
comment = "base XGBoost implementation")
|
||||||
)
|
)
|
||||||
|
Maintainer: Jiaming Yuan <jm.yuan@outlook.com>
|
||||||
Description: Extreme Gradient Boosting, which is an efficient implementation
|
Description: Extreme Gradient Boosting, which is an efficient implementation
|
||||||
of the gradient boosting framework from Chen & Guestrin (2016) <doi:10.1145/2939672.2939785>.
|
of the gradient boosting framework from Chen & Guestrin (2016) <doi:10.1145/2939672.2939785>.
|
||||||
This package is its R interface. The package includes efficient linear
|
This package is its R interface. The package includes efficient linear
|
||||||
@@ -51,10 +54,8 @@ Suggests:
|
|||||||
Ckmeans.1d.dp (>= 3.3.1),
|
Ckmeans.1d.dp (>= 3.3.1),
|
||||||
vcd (>= 1.3),
|
vcd (>= 1.3),
|
||||||
testthat,
|
testthat,
|
||||||
lintr,
|
|
||||||
igraph (>= 1.0.1),
|
igraph (>= 1.0.1),
|
||||||
float,
|
float,
|
||||||
crayon,
|
|
||||||
titanic
|
titanic
|
||||||
Depends:
|
Depends:
|
||||||
R (>= 3.3.0)
|
R (>= 3.3.0)
|
||||||
@@ -62,7 +63,7 @@ Imports:
|
|||||||
Matrix (>= 1.1-0),
|
Matrix (>= 1.1-0),
|
||||||
methods,
|
methods,
|
||||||
data.table (>= 1.9.6),
|
data.table (>= 1.9.6),
|
||||||
magrittr (>= 1.5),
|
|
||||||
jsonlite (>= 1.0),
|
jsonlite (>= 1.0),
|
||||||
RoxygenNote: 7.1.1
|
RoxygenNote: 7.2.3
|
||||||
SystemRequirements: GNU make, C++14
|
Encoding: UTF-8
|
||||||
|
SystemRequirements: GNU make, C++17
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
Copyright (c) 2014 by Tianqi Chen and Contributors
|
Copyright (c) 2014-2023, Tianqi Chen and XBGoost Contributors
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
you may not use this file except in compliance with the License.
|
you may not use this file except in compliance with the License.
|
||||||
|
|||||||
@@ -82,7 +82,6 @@ importFrom(graphics,points)
|
|||||||
importFrom(graphics,title)
|
importFrom(graphics,title)
|
||||||
importFrom(jsonlite,fromJSON)
|
importFrom(jsonlite,fromJSON)
|
||||||
importFrom(jsonlite,toJSON)
|
importFrom(jsonlite,toJSON)
|
||||||
importFrom(magrittr,"%>%")
|
|
||||||
importFrom(stats,median)
|
importFrom(stats,median)
|
||||||
importFrom(stats,predict)
|
importFrom(stats,predict)
|
||||||
importFrom(utils,head)
|
importFrom(utils,head)
|
||||||
|
|||||||
@@ -114,7 +114,7 @@ cb.evaluation.log <- function() {
|
|||||||
if (is.null(mnames) || any(mnames == ""))
|
if (is.null(mnames) || any(mnames == ""))
|
||||||
stop("bst_evaluation must have non-empty names")
|
stop("bst_evaluation must have non-empty names")
|
||||||
|
|
||||||
mnames <<- gsub('-', '_', names(env$bst_evaluation))
|
mnames <<- gsub('-', '_', names(env$bst_evaluation), fixed = TRUE)
|
||||||
if (!is.null(env$bst_evaluation_err))
|
if (!is.null(env$bst_evaluation_err))
|
||||||
mnames <<- c(paste0(mnames, '_mean'), paste0(mnames, '_std'))
|
mnames <<- c(paste0(mnames, '_mean'), paste0(mnames, '_std'))
|
||||||
}
|
}
|
||||||
@@ -185,10 +185,10 @@ cb.reset.parameters <- function(new_params) {
|
|||||||
|
|
||||||
if (typeof(new_params) != "list")
|
if (typeof(new_params) != "list")
|
||||||
stop("'new_params' must be a list")
|
stop("'new_params' must be a list")
|
||||||
pnames <- gsub("\\.", "_", names(new_params))
|
pnames <- gsub(".", "_", names(new_params), fixed = TRUE)
|
||||||
nrounds <- NULL
|
nrounds <- NULL
|
||||||
|
|
||||||
# run some checks in the begining
|
# run some checks in the beginning
|
||||||
init <- function(env) {
|
init <- function(env) {
|
||||||
nrounds <<- env$end_iteration - env$begin_iteration + 1
|
nrounds <<- env$end_iteration - env$begin_iteration + 1
|
||||||
|
|
||||||
@@ -263,10 +263,7 @@ cb.reset.parameters <- function(new_params) {
|
|||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{best_score} the evaluation score at the best iteration
|
#' \item \code{best_score} the evaluation score at the best iteration
|
||||||
#' \item \code{best_iteration} at which boosting iteration the best score has occurred (1-based index)
|
#' \item \code{best_iteration} at which boosting iteration the best score has occurred (1-based index)
|
||||||
#' \item \code{best_ntreelimit} to use with the \code{ntreelimit} parameter in \code{predict}.
|
|
||||||
#' It differs from \code{best_iteration} in multiclass or random forest settings.
|
|
||||||
#' }
|
#' }
|
||||||
#'
|
|
||||||
#' The Same values are also stored as xgb-attributes:
|
#' The Same values are also stored as xgb-attributes:
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{best_iteration} is stored as a 0-based iteration index (for interoperability of binary models)
|
#' \item \code{best_iteration} is stored as a 0-based iteration index (for interoperability of binary models)
|
||||||
@@ -303,9 +300,9 @@ cb.early.stop <- function(stopping_rounds, maximize = FALSE,
|
|||||||
if (length(env$bst_evaluation) == 0)
|
if (length(env$bst_evaluation) == 0)
|
||||||
stop("For early stopping, watchlist must have at least one element")
|
stop("For early stopping, watchlist must have at least one element")
|
||||||
|
|
||||||
eval_names <- gsub('-', '_', names(env$bst_evaluation))
|
eval_names <- gsub('-', '_', names(env$bst_evaluation), fixed = TRUE)
|
||||||
if (!is.null(metric_name)) {
|
if (!is.null(metric_name)) {
|
||||||
metric_idx <<- which(gsub('-', '_', metric_name) == eval_names)
|
metric_idx <<- which(gsub('-', '_', metric_name, fixed = TRUE) == eval_names)
|
||||||
if (length(metric_idx) == 0)
|
if (length(metric_idx) == 0)
|
||||||
stop("'metric_name' for early stopping is not one of the following:\n",
|
stop("'metric_name' for early stopping is not one of the following:\n",
|
||||||
paste(eval_names, collapse = ' '), '\n')
|
paste(eval_names, collapse = ' '), '\n')
|
||||||
@@ -322,7 +319,7 @@ cb.early.stop <- function(stopping_rounds, maximize = FALSE,
|
|||||||
|
|
||||||
# maximize is usually NULL when not set in xgb.train and built-in metrics
|
# maximize is usually NULL when not set in xgb.train and built-in metrics
|
||||||
if (is.null(maximize))
|
if (is.null(maximize))
|
||||||
maximize <<- grepl('(_auc|_map|_ndcg)', metric_name)
|
maximize <<- grepl('(_auc|_map|_ndcg|_pre)', metric_name)
|
||||||
|
|
||||||
if (verbose && NVL(env$rank, 0) == 0)
|
if (verbose && NVL(env$rank, 0) == 0)
|
||||||
cat("Will train until ", metric_name, " hasn't improved in ",
|
cat("Will train until ", metric_name, " hasn't improved in ",
|
||||||
@@ -498,13 +495,12 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
rep(NA_real_, N)
|
rep(NA_real_, N)
|
||||||
}
|
}
|
||||||
|
|
||||||
ntreelimit <- NVL(env$basket$best_ntreelimit,
|
iterationrange <- c(1, NVL(env$basket$best_iteration, env$end_iteration) + 1)
|
||||||
env$end_iteration * env$num_parallel_tree)
|
|
||||||
if (NVL(env$params[['booster']], '') == 'gblinear') {
|
if (NVL(env$params[['booster']], '') == 'gblinear') {
|
||||||
ntreelimit <- 0 # must be 0 for gblinear
|
iterationrange <- c(1, 1) # must be 0 for gblinear
|
||||||
}
|
}
|
||||||
for (fd in env$bst_folds) {
|
for (fd in env$bst_folds) {
|
||||||
pr <- predict(fd$bst, fd$watchlist[[2]], ntreelimit = ntreelimit, reshape = TRUE)
|
pr <- predict(fd$bst, fd$watchlist[[2]], iterationrange = iterationrange, reshape = TRUE)
|
||||||
if (is.matrix(pred)) {
|
if (is.matrix(pred)) {
|
||||||
pred[fd$index, ] <- pr
|
pred[fd$index, ] <- pr
|
||||||
} else {
|
} else {
|
||||||
@@ -515,7 +511,7 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
if (save_models) {
|
if (save_models) {
|
||||||
env$basket$models <- lapply(env$bst_folds, function(fd) {
|
env$basket$models <- lapply(env$bst_folds, function(fd) {
|
||||||
xgb.attr(fd$bst, 'niter') <- env$end_iteration - 1
|
xgb.attr(fd$bst, 'niter') <- env$end_iteration - 1
|
||||||
xgb.Booster.complete(xgb.handleToBooster(fd$bst), saveraw = TRUE)
|
xgb.Booster.complete(xgb.handleToBooster(handle = fd$bst, raw = NULL), saveraw = TRUE)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -533,7 +529,7 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
#' Callback closure for collecting the model coefficients history of a gblinear booster
|
#' Callback closure for collecting the model coefficients history of a gblinear booster
|
||||||
#' during its training.
|
#' during its training.
|
||||||
#'
|
#'
|
||||||
#' @param sparse when set to FALSE/TURE, a dense/sparse matrix is used to store the result.
|
#' @param sparse when set to FALSE/TRUE, a dense/sparse matrix is used to store the result.
|
||||||
#' Sparse format is useful when one expects only a subset of coefficients to be non-zero,
|
#' Sparse format is useful when one expects only a subset of coefficients to be non-zero,
|
||||||
#' when using the "thrifty" feature selector with fairly small number of top features
|
#' when using the "thrifty" feature selector with fairly small number of top features
|
||||||
#' selected per iteration.
|
#' selected per iteration.
|
||||||
@@ -548,9 +544,11 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
#'
|
#'
|
||||||
#' @return
|
#' @return
|
||||||
#' Results are stored in the \code{coefs} element of the closure.
|
#' Results are stored in the \code{coefs} element of the closure.
|
||||||
#' The \code{\link{xgb.gblinear.history}} convenience function provides an easy way to access it.
|
#' The \code{\link{xgb.gblinear.history}} convenience function provides an easy
|
||||||
|
#' way to access it.
|
||||||
#' With \code{xgb.train}, it is either a dense of a sparse matrix.
|
#' With \code{xgb.train}, it is either a dense of a sparse matrix.
|
||||||
#' While with \code{xgb.cv}, it is a list (an element per each fold) of such matrices.
|
#' While with \code{xgb.cv}, it is a list (an element per each fold) of such
|
||||||
|
#' matrices.
|
||||||
#'
|
#'
|
||||||
#' @seealso
|
#' @seealso
|
||||||
#' \code{\link{callbacks}}, \code{\link{xgb.gblinear.history}}.
|
#' \code{\link{callbacks}}, \code{\link{xgb.gblinear.history}}.
|
||||||
@@ -560,10 +558,9 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
#' #
|
#' #
|
||||||
#' # In the iris dataset, it is hard to linearly separate Versicolor class from the rest
|
#' # In the iris dataset, it is hard to linearly separate Versicolor class from the rest
|
||||||
#' # without considering the 2nd order interactions:
|
#' # without considering the 2nd order interactions:
|
||||||
#' require(magrittr)
|
|
||||||
#' x <- model.matrix(Species ~ .^2, iris)[,-1]
|
#' x <- model.matrix(Species ~ .^2, iris)[,-1]
|
||||||
#' colnames(x)
|
#' colnames(x)
|
||||||
#' dtrain <- xgb.DMatrix(scale(x), label = 1*(iris$Species == "versicolor"))
|
#' dtrain <- xgb.DMatrix(scale(x), label = 1*(iris$Species == "versicolor"), nthread = 2)
|
||||||
#' param <- list(booster = "gblinear", objective = "reg:logistic", eval_metric = "auc",
|
#' param <- list(booster = "gblinear", objective = "reg:logistic", eval_metric = "auc",
|
||||||
#' lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
#' lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
||||||
#' # For 'shotgun', which is a default linear updater, using high eta values may result in
|
#' # For 'shotgun', which is a default linear updater, using high eta values may result in
|
||||||
@@ -581,7 +578,7 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
#' bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 200, eta = 0.8,
|
#' bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 200, eta = 0.8,
|
||||||
#' updater = 'coord_descent', feature_selector = 'thrifty', top_k = 1,
|
#' updater = 'coord_descent', feature_selector = 'thrifty', top_k = 1,
|
||||||
#' callbacks = list(cb.gblinear.history()))
|
#' callbacks = list(cb.gblinear.history()))
|
||||||
#' xgb.gblinear.history(bst) %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst), type = 'l')
|
||||||
#' # Componentwise boosting is known to have similar effect to Lasso regularization.
|
#' # Componentwise boosting is known to have similar effect to Lasso regularization.
|
||||||
#' # Try experimenting with various values of top_k, eta, nrounds,
|
#' # Try experimenting with various values of top_k, eta, nrounds,
|
||||||
#' # as well as different feature_selectors.
|
#' # as well as different feature_selectors.
|
||||||
@@ -590,37 +587,39 @@ cb.cv.predict <- function(save_models = FALSE) {
|
|||||||
#' bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 100, eta = 0.8,
|
#' bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 100, eta = 0.8,
|
||||||
#' callbacks = list(cb.gblinear.history()))
|
#' callbacks = list(cb.gblinear.history()))
|
||||||
#' # coefficients in the CV fold #3
|
#' # coefficients in the CV fold #3
|
||||||
#' xgb.gblinear.history(bst)[[3]] %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst)[[3]], type = 'l')
|
||||||
#'
|
#'
|
||||||
#'
|
#'
|
||||||
#' #### Multiclass classification:
|
#' #### Multiclass classification:
|
||||||
#' #
|
#' #
|
||||||
#' dtrain <- xgb.DMatrix(scale(x), label = as.numeric(iris$Species) - 1)
|
#' dtrain <- xgb.DMatrix(scale(x), label = as.numeric(iris$Species) - 1, nthread = 1)
|
||||||
#' param <- list(booster = "gblinear", objective = "multi:softprob", num_class = 3,
|
#' param <- list(booster = "gblinear", objective = "multi:softprob", num_class = 3,
|
||||||
#' lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
#' lambda = 0.0003, alpha = 0.0003, nthread = 1)
|
||||||
#' # For the default linear updater 'shotgun' it sometimes is helpful
|
#' # For the default linear updater 'shotgun' it sometimes is helpful
|
||||||
#' # to use smaller eta to reduce instability
|
#' # to use smaller eta to reduce instability
|
||||||
#' bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 70, eta = 0.5,
|
#' bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 50, eta = 0.5,
|
||||||
#' callbacks = list(cb.gblinear.history()))
|
#' callbacks = list(cb.gblinear.history()))
|
||||||
#' # Will plot the coefficient paths separately for each class:
|
#' # Will plot the coefficient paths separately for each class:
|
||||||
#' xgb.gblinear.history(bst, class_index = 0) %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst, class_index = 0), type = 'l')
|
||||||
#' xgb.gblinear.history(bst, class_index = 1) %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst, class_index = 1), type = 'l')
|
||||||
#' xgb.gblinear.history(bst, class_index = 2) %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst, class_index = 2), type = 'l')
|
||||||
#'
|
#'
|
||||||
#' # CV:
|
#' # CV:
|
||||||
#' bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 70, eta = 0.5,
|
#' bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 70, eta = 0.5,
|
||||||
#' callbacks = list(cb.gblinear.history(FALSE)))
|
#' callbacks = list(cb.gblinear.history(FALSE)))
|
||||||
#' # 1st forld of 1st class
|
#' # 1st fold of 1st class
|
||||||
#' xgb.gblinear.history(bst, class_index = 0)[[1]] %>% matplot(type = 'l')
|
#' matplot(xgb.gblinear.history(bst, class_index = 0)[[1]], type = 'l')
|
||||||
#'
|
#'
|
||||||
#' @export
|
#' @export
|
||||||
cb.gblinear.history <- function(sparse = FALSE) {
|
cb.gblinear.history <- function(sparse = FALSE) {
|
||||||
coefs <- NULL
|
coefs <- NULL
|
||||||
|
|
||||||
init <- function(env) {
|
init <- function(env) {
|
||||||
if (!is.null(env$bst)) { # xgb.train:
|
# xgb.train(): bst will be present
|
||||||
} else if (!is.null(env$bst_folds)) { # xgb.cv:
|
# xgb.cv(): bst_folds will be present
|
||||||
} else stop("Parent frame has neither 'bst' nor 'bst_folds'")
|
if (is.null(env$bst) && is.null(env$bst_folds)) {
|
||||||
|
stop("Parent frame has neither 'bst' nor 'bst_folds'")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
# convert from list to (sparse) matrix
|
# convert from list to (sparse) matrix
|
||||||
@@ -642,9 +641,14 @@ cb.gblinear.history <- function(sparse=FALSE) {
|
|||||||
if (!is.null(env$bst)) { # # xgb.train:
|
if (!is.null(env$bst)) { # # xgb.train:
|
||||||
coefs <<- list2mat(coefs)
|
coefs <<- list2mat(coefs)
|
||||||
} else { # xgb.cv:
|
} else { # xgb.cv:
|
||||||
# first lapply transposes the list
|
# second lapply transposes the list
|
||||||
coefs <<- lapply(seq_along(coefs[[1]]), function(i) lapply(coefs, "[[", i)) %>%
|
coefs <<- lapply(
|
||||||
lapply(function(x) list2mat(x))
|
X = lapply(
|
||||||
|
X = seq_along(coefs[[1]]),
|
||||||
|
FUN = function(i) lapply(coefs, "[[", i)
|
||||||
|
),
|
||||||
|
FUN = list2mat
|
||||||
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -655,7 +659,7 @@ cb.gblinear.history <- function(sparse=FALSE) {
|
|||||||
} else { # xgb.cv:
|
} else { # xgb.cv:
|
||||||
cf <- vector("list", length(env$bst_folds))
|
cf <- vector("list", length(env$bst_folds))
|
||||||
for (i in seq_along(env$bst_folds)) {
|
for (i in seq_along(env$bst_folds)) {
|
||||||
dmp <- xgb.dump(xgb.handleToBooster(env$bst_folds[[i]]$bst))
|
dmp <- xgb.dump(xgb.handleToBooster(handle = env$bst_folds[[i]]$bst, raw = NULL))
|
||||||
cf[[i]] <- as.numeric(grep('(booster|bias|weigh)', dmp, invert = TRUE, value = TRUE))
|
cf[[i]] <- as.numeric(grep('(booster|bias|weigh)', dmp, invert = TRUE, value = TRUE))
|
||||||
if (sparse) cf[[i]] <- as(cf[[i]], "sparseVector")
|
if (sparse) cf[[i]] <- as(cf[[i]], "sparseVector")
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
#
|
#
|
||||||
# This file is for the low level reuseable utility functions
|
# This file is for the low level reusable utility functions
|
||||||
# that are not supposed to be visibe to a user.
|
# that are not supposed to be visible to a user.
|
||||||
#
|
#
|
||||||
|
|
||||||
#
|
#
|
||||||
@@ -38,11 +38,11 @@ check.booster.params <- function(params, ...) {
|
|||||||
stop("params must be a list")
|
stop("params must be a list")
|
||||||
|
|
||||||
# in R interface, allow for '.' instead of '_' in parameter names
|
# in R interface, allow for '.' instead of '_' in parameter names
|
||||||
names(params) <- gsub("\\.", "_", names(params))
|
names(params) <- gsub(".", "_", names(params), fixed = TRUE)
|
||||||
|
|
||||||
# merge parameters from the params and the dots-expansion
|
# merge parameters from the params and the dots-expansion
|
||||||
dot_params <- list(...)
|
dot_params <- list(...)
|
||||||
names(dot_params) <- gsub("\\.", "_", names(dot_params))
|
names(dot_params) <- gsub(".", "_", names(dot_params), fixed = TRUE)
|
||||||
if (length(intersect(names(params),
|
if (length(intersect(names(params),
|
||||||
names(dot_params))) > 0)
|
names(dot_params))) > 0)
|
||||||
stop("Same parameters in 'params' and in the call are not allowed. Please check your 'params' list.")
|
stop("Same parameters in 'params' and in the call are not allowed. Please check your 'params' list.")
|
||||||
@@ -140,7 +140,7 @@ check.custom.eval <- function(env = parent.frame()) {
|
|||||||
|
|
||||||
|
|
||||||
# Update a booster handle for an iteration with dtrain data
|
# Update a booster handle for an iteration with dtrain data
|
||||||
xgb.iter.update <- function(booster_handle, dtrain, iter, obj = NULL) {
|
xgb.iter.update <- function(booster_handle, dtrain, iter, obj) {
|
||||||
if (!identical(class(booster_handle), "xgb.Booster.handle")) {
|
if (!identical(class(booster_handle), "xgb.Booster.handle")) {
|
||||||
stop("booster_handle must be of xgb.Booster.handle class")
|
stop("booster_handle must be of xgb.Booster.handle class")
|
||||||
}
|
}
|
||||||
@@ -163,7 +163,7 @@ xgb.iter.update <- function(booster_handle, dtrain, iter, obj = NULL) {
|
|||||||
# Evaluate one iteration.
|
# Evaluate one iteration.
|
||||||
# Returns a named vector of evaluation metrics
|
# Returns a named vector of evaluation metrics
|
||||||
# with the names in a 'datasetname-metricname' format.
|
# with the names in a 'datasetname-metricname' format.
|
||||||
xgb.iter.eval <- function(booster_handle, watchlist, iter, feval = NULL) {
|
xgb.iter.eval <- function(booster_handle, watchlist, iter, feval) {
|
||||||
if (!identical(class(booster_handle), "xgb.Booster.handle"))
|
if (!identical(class(booster_handle), "xgb.Booster.handle"))
|
||||||
stop("class of booster_handle must be xgb.Booster.handle")
|
stop("class of booster_handle must be xgb.Booster.handle")
|
||||||
|
|
||||||
@@ -178,7 +178,8 @@ xgb.iter.eval <- function(booster_handle, watchlist, iter, feval = NULL) {
|
|||||||
} else {
|
} else {
|
||||||
res <- sapply(seq_along(watchlist), function(j) {
|
res <- sapply(seq_along(watchlist), function(j) {
|
||||||
w <- watchlist[[j]]
|
w <- watchlist[[j]]
|
||||||
preds <- predict(booster_handle, w, outputmargin = TRUE, ntreelimit = 0) # predict using all trees
|
## predict using all trees
|
||||||
|
preds <- predict(booster_handle, w, outputmargin = TRUE, iterationrange = c(1, 1))
|
||||||
eval_res <- feval(preds, w)
|
eval_res <- feval(preds, w)
|
||||||
out <- eval_res$value
|
out <- eval_res$value
|
||||||
names(out) <- paste0(evnames[j], "-", eval_res$metric)
|
names(out) <- paste0(evnames[j], "-", eval_res$metric)
|
||||||
@@ -233,7 +234,7 @@ generate.cv.folds <- function(nfold, nrows, stratified, label, params) {
|
|||||||
y <- factor(y)
|
y <- factor(y)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
folds <- xgb.createFolds(y, nfold)
|
folds <- xgb.createFolds(y = y, k = nfold)
|
||||||
} else {
|
} else {
|
||||||
# make simple non-stratified folds
|
# make simple non-stratified folds
|
||||||
kstep <- length(rnd_idx) %/% nfold
|
kstep <- length(rnd_idx) %/% nfold
|
||||||
@@ -250,8 +251,7 @@ generate.cv.folds <- function(nfold, nrows, stratified, label, params) {
|
|||||||
# Creates CV folds stratified by the values of y.
|
# Creates CV folds stratified by the values of y.
|
||||||
# It was borrowed from caret::createFolds and simplified
|
# It was borrowed from caret::createFolds and simplified
|
||||||
# by always returning an unnamed list of fold indices.
|
# by always returning an unnamed list of fold indices.
|
||||||
xgb.createFolds <- function(y, k = 10)
|
xgb.createFolds <- function(y, k) {
|
||||||
{
|
|
||||||
if (is.numeric(y)) {
|
if (is.numeric(y)) {
|
||||||
## Group the numeric data based on their magnitudes
|
## Group the numeric data based on their magnitudes
|
||||||
## and sample within those groups.
|
## and sample within those groups.
|
||||||
@@ -284,7 +284,7 @@ xgb.createFolds <- function(y, k = 10)
|
|||||||
for (i in seq_along(numInClass)) {
|
for (i in seq_along(numInClass)) {
|
||||||
## create a vector of integers from 1:k as many times as possible without
|
## create a vector of integers from 1:k as many times as possible without
|
||||||
## going over the number of samples in the class. Note that if the number
|
## going over the number of samples in the class. Note that if the number
|
||||||
## of samples in a class is less than k, nothing is producd here.
|
## of samples in a class is less than k, nothing is produced here.
|
||||||
seqVector <- rep(seq_len(k), numInClass[i] %/% k)
|
seqVector <- rep(seq_len(k), numInClass[i] %/% k)
|
||||||
## add enough random integers to get length(seqVector) == numInClass[i]
|
## add enough random integers to get length(seqVector) == numInClass[i]
|
||||||
if (numInClass[i] %% k > 0) seqVector <- c(seqVector, sample.int(k, numInClass[i] %% k))
|
if (numInClass[i] %% k > 0) seqVector <- c(seqVector, sample.int(k, numInClass[i] %% k))
|
||||||
|
|||||||
@@ -1,7 +1,6 @@
|
|||||||
# Construct an internal xgboost Booster and return a handle to it.
|
# Construct an internal xgboost Booster and return a handle to it.
|
||||||
# internal utility function
|
# internal utility function
|
||||||
xgb.Booster.handle <- function(params = list(), cachelist = list(),
|
xgb.Booster.handle <- function(params, cachelist, modelfile, handle) {
|
||||||
modelfile = NULL) {
|
|
||||||
if (typeof(cachelist) != "list" ||
|
if (typeof(cachelist) != "list" ||
|
||||||
!all(vapply(cachelist, inherits, logical(1), what = 'xgb.DMatrix'))) {
|
!all(vapply(cachelist, inherits, logical(1), what = 'xgb.DMatrix'))) {
|
||||||
stop("cachelist must be a list of xgb.DMatrix objects")
|
stop("cachelist must be a list of xgb.DMatrix objects")
|
||||||
@@ -12,7 +11,7 @@ xgb.Booster.handle <- function(params = list(), cachelist = list(),
|
|||||||
## A filename
|
## A filename
|
||||||
handle <- .Call(XGBoosterCreate_R, cachelist)
|
handle <- .Call(XGBoosterCreate_R, cachelist)
|
||||||
modelfile <- path.expand(modelfile)
|
modelfile <- path.expand(modelfile)
|
||||||
.Call(XGBoosterLoadModel_R, handle, modelfile[1])
|
.Call(XGBoosterLoadModel_R, handle, enc2utf8(modelfile[1]))
|
||||||
class(handle) <- "xgb.Booster.handle"
|
class(handle) <- "xgb.Booster.handle"
|
||||||
if (length(params) > 0) {
|
if (length(params) > 0) {
|
||||||
xgb.parameters(handle) <- params
|
xgb.parameters(handle) <- params
|
||||||
@@ -20,7 +19,7 @@ xgb.Booster.handle <- function(params = list(), cachelist = list(),
|
|||||||
return(handle)
|
return(handle)
|
||||||
} else if (typeof(modelfile) == "raw") {
|
} else if (typeof(modelfile) == "raw") {
|
||||||
## A memory buffer
|
## A memory buffer
|
||||||
bst <- xgb.unserialize(modelfile)
|
bst <- xgb.unserialize(modelfile, handle)
|
||||||
xgb.parameters(bst) <- params
|
xgb.parameters(bst) <- params
|
||||||
return (bst)
|
return (bst)
|
||||||
} else if (inherits(modelfile, "xgb.Booster")) {
|
} else if (inherits(modelfile, "xgb.Booster")) {
|
||||||
@@ -44,7 +43,7 @@ xgb.Booster.handle <- function(params = list(), cachelist = list(),
|
|||||||
|
|
||||||
# Convert xgb.Booster.handle to xgb.Booster
|
# Convert xgb.Booster.handle to xgb.Booster
|
||||||
# internal utility function
|
# internal utility function
|
||||||
xgb.handleToBooster <- function(handle, raw = NULL) {
|
xgb.handleToBooster <- function(handle, raw) {
|
||||||
bst <- list(handle = handle, raw = raw)
|
bst <- list(handle = handle, raw = raw)
|
||||||
class(bst) <- "xgb.Booster"
|
class(bst) <- "xgb.Booster"
|
||||||
return(bst)
|
return(bst)
|
||||||
@@ -129,7 +128,12 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
stop("argument type must be xgb.Booster")
|
stop("argument type must be xgb.Booster")
|
||||||
|
|
||||||
if (is.null.handle(object$handle)) {
|
if (is.null.handle(object$handle)) {
|
||||||
object$handle <- xgb.Booster.handle(modelfile = object$raw)
|
object$handle <- xgb.Booster.handle(
|
||||||
|
params = list(),
|
||||||
|
cachelist = list(),
|
||||||
|
modelfile = object$raw,
|
||||||
|
handle = object$handle
|
||||||
|
)
|
||||||
} else {
|
} else {
|
||||||
if (is.null(object$raw) && saveraw) {
|
if (is.null(object$raw) && saveraw) {
|
||||||
object$raw <- xgb.serialize(object$handle)
|
object$raw <- xgb.serialize(object$handle)
|
||||||
@@ -162,14 +166,17 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' Predicted values based on either xgboost model or model handle object.
|
#' Predicted values based on either xgboost model or model handle object.
|
||||||
#'
|
#'
|
||||||
#' @param object Object of class \code{xgb.Booster} or \code{xgb.Booster.handle}
|
#' @param object Object of class \code{xgb.Booster} or \code{xgb.Booster.handle}
|
||||||
#' @param newdata takes \code{matrix}, \code{dgCMatrix}, local data file or \code{xgb.DMatrix}.
|
#' @param newdata takes \code{matrix}, \code{dgCMatrix}, \code{dgRMatrix}, \code{dsparseVector},
|
||||||
|
#' local data file or \code{xgb.DMatrix}.
|
||||||
|
#'
|
||||||
|
#' For single-row predictions on sparse data, it's recommended to use CSR format. If passing
|
||||||
|
#' a sparse vector, it will take it as a row vector.
|
||||||
#' @param missing Missing is only used when input is dense matrix. Pick a float value that represents
|
#' @param missing Missing is only used when input is dense matrix. Pick a float value that represents
|
||||||
#' missing values in data (e.g., sometimes 0 or some other extreme value is used).
|
#' missing values in data (e.g., sometimes 0 or some other extreme value is used).
|
||||||
#' @param outputmargin whether the prediction should be returned in the for of original untransformed
|
#' @param outputmargin whether the prediction should be returned in the for of original untransformed
|
||||||
#' sum of predictions from boosting iterations' results. E.g., setting \code{outputmargin=TRUE} for
|
#' sum of predictions from boosting iterations' results. E.g., setting \code{outputmargin=TRUE} for
|
||||||
#' logistic regression would result in predictions for log-odds instead of probabilities.
|
#' logistic regression would result in predictions for log-odds instead of probabilities.
|
||||||
#' @param ntreelimit limit the number of model's trees or boosting iterations used in prediction (see Details).
|
#' @param ntreelimit Deprecated, use \code{iterationrange} instead.
|
||||||
#' It will use all the trees by default (\code{NULL} value).
|
|
||||||
#' @param predleaf whether predict leaf index.
|
#' @param predleaf whether predict leaf index.
|
||||||
#' @param predcontrib whether to return feature contributions to individual predictions (see Details).
|
#' @param predcontrib whether to return feature contributions to individual predictions (see Details).
|
||||||
#' @param approxcontrib whether to use a fast approximation for feature contributions (see Details).
|
#' @param approxcontrib whether to use a fast approximation for feature contributions (see Details).
|
||||||
@@ -179,16 +186,19 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' or predinteraction flags is TRUE.
|
#' or predinteraction flags is TRUE.
|
||||||
#' @param training whether is the prediction result used for training. For dart booster,
|
#' @param training whether is the prediction result used for training. For dart booster,
|
||||||
#' training predicting will perform dropout.
|
#' training predicting will perform dropout.
|
||||||
|
#' @param iterationrange Specifies which layer of trees are used in prediction. For
|
||||||
|
#' example, if a random forest is trained with 100 rounds. Specifying
|
||||||
|
#' `iterationrange=(1, 21)`, then only the forests built during [1, 21) (half open set)
|
||||||
|
#' rounds are used in this prediction. It's 1-based index just like R vector. When set
|
||||||
|
#' to \code{c(1, 1)} XGBoost will use all trees.
|
||||||
|
#' @param strict_shape Default is \code{FALSE}. When it's set to \code{TRUE}, output
|
||||||
|
#' type and shape of prediction are invariant to model type.
|
||||||
|
#'
|
||||||
#' @param ... Parameters passed to \code{predict.xgb.Booster}
|
#' @param ... Parameters passed to \code{predict.xgb.Booster}
|
||||||
#'
|
#'
|
||||||
#' @details
|
#' @details
|
||||||
#' Note that \code{ntreelimit} is not necessarily equal to the number of boosting iterations
|
|
||||||
#' and it is not necessarily equal to the number of trees in a model.
|
|
||||||
#' E.g., in a random forest-like model, \code{ntreelimit} would limit the number of trees.
|
|
||||||
#' But for multiclass classification, while there are multiple trees per iteration,
|
|
||||||
#' \code{ntreelimit} limits the number of boosting iterations.
|
|
||||||
#'
|
#'
|
||||||
#' Also note that \code{ntreelimit} would currently do nothing for predictions from gblinear,
|
#' Note that \code{iterationrange} would currently do nothing for predictions from gblinear,
|
||||||
#' since gblinear doesn't keep its boosting history.
|
#' since gblinear doesn't keep its boosting history.
|
||||||
#'
|
#'
|
||||||
#' One possible practical applications of the \code{predleaf} option is to use the model
|
#' One possible practical applications of the \code{predleaf} option is to use the model
|
||||||
@@ -208,8 +218,13 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' Since it quadratically depends on the number of features, it is recommended to perform selection
|
#' Since it quadratically depends on the number of features, it is recommended to perform selection
|
||||||
#' of the most important features first. See below about the format of the returned results.
|
#' of the most important features first. See below about the format of the returned results.
|
||||||
#'
|
#'
|
||||||
|
#' The \code{predict()} method uses as many threads as defined in \code{xgb.Booster} object (all by default).
|
||||||
|
#' If you want to change their number, then assign a new number to \code{nthread} using \code{\link{xgb.parameters<-}}.
|
||||||
|
#' Note also that converting a matrix to \code{\link{xgb.DMatrix}} uses multiple threads too.
|
||||||
|
#'
|
||||||
#' @return
|
#' @return
|
||||||
#' For regression or binary classification, it returns a vector of length \code{nrows(newdata)}.
|
#' The return type is different depending whether \code{strict_shape} is set to \code{TRUE}. By default,
|
||||||
|
#' for regression or binary classification, it returns a vector of length \code{nrows(newdata)}.
|
||||||
#' For multiclass classification, either a \code{num_class * nrows(newdata)} vector or
|
#' For multiclass classification, either a \code{num_class * nrows(newdata)} vector or
|
||||||
#' a \code{(nrows(newdata), num_class)} dimension matrix is returned, depending on
|
#' a \code{(nrows(newdata), num_class)} dimension matrix is returned, depending on
|
||||||
#' the \code{reshape} value.
|
#' the \code{reshape} value.
|
||||||
@@ -231,6 +246,13 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' For a multiclass case, a list of \code{num_class} elements is returned, where each element is
|
#' For a multiclass case, a list of \code{num_class} elements is returned, where each element is
|
||||||
#' such an array.
|
#' such an array.
|
||||||
#'
|
#'
|
||||||
|
#' When \code{strict_shape} is set to \code{TRUE}, the output is always an array. For
|
||||||
|
#' normal prediction, the output is a 2-dimension array \code{(num_class, nrow(newdata))}.
|
||||||
|
#'
|
||||||
|
#' For \code{predcontrib = TRUE}, output is \code{(ncol(newdata) + 1, num_class, nrow(newdata))}
|
||||||
|
#' For \code{predinteraction = TRUE}, output is \code{(ncol(newdata) + 1, ncol(newdata) + 1, num_class, nrow(newdata))}
|
||||||
|
#' For \code{predleaf = TRUE}, output is \code{(n_trees_in_forest, num_class, n_iterations, nrow(newdata))}
|
||||||
|
#'
|
||||||
#' @seealso
|
#' @seealso
|
||||||
#' \code{\link{xgb.train}}.
|
#' \code{\link{xgb.train}}.
|
||||||
#'
|
#'
|
||||||
@@ -253,7 +275,7 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' # use all trees by default
|
#' # use all trees by default
|
||||||
#' pred <- predict(bst, test$data)
|
#' pred <- predict(bst, test$data)
|
||||||
#' # use only the 1st tree
|
#' # use only the 1st tree
|
||||||
#' pred1 <- predict(bst, test$data, ntreelimit = 1)
|
#' pred1 <- predict(bst, test$data, iterationrange = c(1, 2))
|
||||||
#'
|
#'
|
||||||
#' # Predicting tree leafs:
|
#' # Predicting tree leafs:
|
||||||
#' # the result is an nsamples X ntrees matrix
|
#' # the result is an nsamples X ntrees matrix
|
||||||
@@ -305,101 +327,159 @@ xgb.Booster.complete <- function(object, saveraw = TRUE) {
|
|||||||
#' all.equal(pred, pred_labels)
|
#' all.equal(pred, pred_labels)
|
||||||
#' # prediction from using only 5 iterations should result
|
#' # prediction from using only 5 iterations should result
|
||||||
#' # in the same error as seen in iteration 5:
|
#' # in the same error as seen in iteration 5:
|
||||||
#' pred5 <- predict(bst, as.matrix(iris[, -5]), ntreelimit=5)
|
#' pred5 <- predict(bst, as.matrix(iris[, -5]), iterationrange=c(1, 6))
|
||||||
#' sum(pred5 != lb)/length(lb)
|
#' sum(pred5 != lb)/length(lb)
|
||||||
#'
|
#'
|
||||||
#'
|
|
||||||
#' ## random forest-like model of 25 trees for binary classification:
|
|
||||||
#'
|
|
||||||
#' set.seed(11)
|
|
||||||
#' bst <- xgboost(data = train$data, label = train$label, max_depth = 5,
|
|
||||||
#' nthread = 2, nrounds = 1, objective = "binary:logistic",
|
|
||||||
#' num_parallel_tree = 25, subsample = 0.6, colsample_bytree = 0.1)
|
|
||||||
#' # Inspect the prediction error vs number of trees:
|
|
||||||
#' lb <- test$label
|
|
||||||
#' dtest <- xgb.DMatrix(test$data, label=lb)
|
|
||||||
#' err <- sapply(1:25, function(n) {
|
|
||||||
#' pred <- predict(bst, dtest, ntreelimit=n)
|
|
||||||
#' sum((pred > 0.5) != lb)/length(lb)
|
|
||||||
#' })
|
|
||||||
#' plot(err, type='l', ylim=c(0,0.1), xlab='#trees')
|
|
||||||
#'
|
|
||||||
#' @rdname predict.xgb.Booster
|
#' @rdname predict.xgb.Booster
|
||||||
#' @export
|
#' @export
|
||||||
predict.xgb.Booster <- function(object, newdata, missing = NA, outputmargin = FALSE, ntreelimit = NULL,
|
predict.xgb.Booster <- function(object, newdata, missing = NA, outputmargin = FALSE, ntreelimit = NULL,
|
||||||
predleaf = FALSE, predcontrib = FALSE, approxcontrib = FALSE, predinteraction = FALSE,
|
predleaf = FALSE, predcontrib = FALSE, approxcontrib = FALSE, predinteraction = FALSE,
|
||||||
reshape = FALSE, training = FALSE, ...) {
|
reshape = FALSE, training = FALSE, iterationrange = NULL, strict_shape = FALSE, ...) {
|
||||||
|
|
||||||
object <- xgb.Booster.complete(object, saveraw = FALSE)
|
object <- xgb.Booster.complete(object, saveraw = FALSE)
|
||||||
|
|
||||||
if (!inherits(newdata, "xgb.DMatrix"))
|
if (!inherits(newdata, "xgb.DMatrix"))
|
||||||
newdata <- xgb.DMatrix(newdata, missing = missing)
|
newdata <- xgb.DMatrix(newdata, missing = missing, nthread = NVL(object$params[["nthread"]], -1))
|
||||||
if (!is.null(object[["feature_names"]]) &&
|
if (!is.null(object[["feature_names"]]) &&
|
||||||
!is.null(colnames(newdata)) &&
|
!is.null(colnames(newdata)) &&
|
||||||
!identical(object[["feature_names"]], colnames(newdata)))
|
!identical(object[["feature_names"]], colnames(newdata)))
|
||||||
stop("Feature names stored in `object` and `newdata` are different!")
|
stop("Feature names stored in `object` and `newdata` are different!")
|
||||||
if (is.null(ntreelimit))
|
|
||||||
ntreelimit <- NVL(object$best_ntreelimit, 0)
|
if (NVL(object$params[['booster']], '') == 'gblinear' || is.null(ntreelimit))
|
||||||
if (NVL(object$params[['booster']], '') == 'gblinear')
|
|
||||||
ntreelimit <- 0
|
ntreelimit <- 0
|
||||||
if (ntreelimit < 0)
|
|
||||||
stop("ntreelimit cannot be negative")
|
|
||||||
|
|
||||||
option <- 0L + 1L * as.logical(outputmargin) + 2L * as.logical(predleaf) + 4L * as.logical(predcontrib) +
|
if (ntreelimit != 0 && is.null(iterationrange)) {
|
||||||
8L * as.logical(approxcontrib) + 16L * as.logical(predinteraction)
|
## only ntreelimit, initialize iteration range
|
||||||
|
iterationrange <- c(0, 0)
|
||||||
|
} else if (ntreelimit == 0 && !is.null(iterationrange)) {
|
||||||
|
## only iteration range, handle 1-based indexing
|
||||||
|
iterationrange <- c(iterationrange[1] - 1, iterationrange[2] - 1)
|
||||||
|
} else if (ntreelimit != 0 && !is.null(iterationrange)) {
|
||||||
|
## both are specified, let libgxgboost throw an error
|
||||||
|
} else {
|
||||||
|
## no limit is supplied, use best
|
||||||
|
if (is.null(object$best_iteration)) {
|
||||||
|
iterationrange <- c(0, 0)
|
||||||
|
} else {
|
||||||
|
## We don't need to + 1 as R is 1-based index.
|
||||||
|
iterationrange <- c(0, as.integer(object$best_iteration))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
## Handle the 0 length values.
|
||||||
|
box <- function(val) {
|
||||||
|
if (length(val) == 0) {
|
||||||
|
cval <- vector(, 1)
|
||||||
|
cval[0] <- val
|
||||||
|
return(cval)
|
||||||
|
}
|
||||||
|
return (val)
|
||||||
|
}
|
||||||
|
|
||||||
ret <- .Call(XGBoosterPredict_R, object$handle, newdata, option[1],
|
## We set strict_shape to TRUE then drop the dimensions conditionally
|
||||||
as.integer(ntreelimit), as.integer(training))
|
args <- list(
|
||||||
|
training = box(training),
|
||||||
|
strict_shape = box(TRUE),
|
||||||
|
iteration_begin = box(as.integer(iterationrange[1])),
|
||||||
|
iteration_end = box(as.integer(iterationrange[2])),
|
||||||
|
ntree_limit = box(as.integer(ntreelimit)),
|
||||||
|
type = box(as.integer(0))
|
||||||
|
)
|
||||||
|
|
||||||
|
set_type <- function(type) {
|
||||||
|
if (args$type != 0) {
|
||||||
|
stop("One type of prediction at a time.")
|
||||||
|
}
|
||||||
|
return(box(as.integer(type)))
|
||||||
|
}
|
||||||
|
if (outputmargin) {
|
||||||
|
args$type <- set_type(1)
|
||||||
|
}
|
||||||
|
if (predcontrib) {
|
||||||
|
args$type <- set_type(if (approxcontrib) 3 else 2)
|
||||||
|
}
|
||||||
|
if (predinteraction) {
|
||||||
|
args$type <- set_type(if (approxcontrib) 5 else 4)
|
||||||
|
}
|
||||||
|
if (predleaf) {
|
||||||
|
args$type <- set_type(6)
|
||||||
|
}
|
||||||
|
|
||||||
|
predts <- .Call(
|
||||||
|
XGBoosterPredictFromDMatrix_R, object$handle, newdata, jsonlite::toJSON(args, auto_unbox = TRUE)
|
||||||
|
)
|
||||||
|
names(predts) <- c("shape", "results")
|
||||||
|
shape <- predts$shape
|
||||||
|
ret <- predts$results
|
||||||
|
|
||||||
n_ret <- length(ret)
|
n_ret <- length(ret)
|
||||||
n_row <- nrow(newdata)
|
n_row <- nrow(newdata)
|
||||||
npred_per_case <- n_ret / n_row
|
if (n_row != shape[1]) {
|
||||||
|
stop("Incorrect predict shape.")
|
||||||
|
}
|
||||||
|
|
||||||
if (n_ret %% n_row != 0)
|
arr <- array(data = ret, dim = rev(shape))
|
||||||
stop("prediction length ", n_ret, " is not multiple of nrows(newdata) ", n_row)
|
|
||||||
|
cnames <- if (!is.null(colnames(newdata))) c(colnames(newdata), "BIAS") else NULL
|
||||||
|
n_groups <- shape[2]
|
||||||
|
|
||||||
|
## Needed regardless of whether strict shape is being used.
|
||||||
|
if (predcontrib) {
|
||||||
|
dimnames(arr) <- list(cnames, NULL, NULL)
|
||||||
|
} else if (predinteraction) {
|
||||||
|
dimnames(arr) <- list(cnames, cnames, NULL, NULL)
|
||||||
|
}
|
||||||
|
if (strict_shape) {
|
||||||
|
return(arr) # strict shape is calculated by libxgboost uniformly.
|
||||||
|
}
|
||||||
|
|
||||||
if (predleaf) {
|
if (predleaf) {
|
||||||
ret <- if (n_ret == n_row) {
|
## Predict leaf
|
||||||
matrix(ret, ncol = 1)
|
arr <- if (n_ret == n_row) {
|
||||||
|
matrix(arr, ncol = 1)
|
||||||
} else {
|
} else {
|
||||||
matrix(ret, nrow = n_row, byrow = TRUE)
|
matrix(arr, nrow = n_row, byrow = TRUE)
|
||||||
}
|
}
|
||||||
} else if (predcontrib) {
|
} else if (predcontrib) {
|
||||||
n_col1 <- ncol(newdata) + 1
|
## Predict contribution
|
||||||
n_group <- npred_per_case / n_col1
|
arr <- aperm(a = arr, perm = c(2, 3, 1)) # [group, row, col]
|
||||||
cnames <- if (!is.null(colnames(newdata))) c(colnames(newdata), "BIAS") else NULL
|
arr <- if (n_ret == n_row) {
|
||||||
ret <- if (n_ret == n_row) {
|
matrix(arr, ncol = 1, dimnames = list(NULL, cnames))
|
||||||
matrix(ret, ncol = 1, dimnames = list(NULL, cnames))
|
} else if (n_groups != 1) {
|
||||||
} else if (n_group == 1) {
|
## turns array into list of matrices
|
||||||
matrix(ret, nrow = n_row, byrow = TRUE, dimnames = list(NULL, cnames))
|
lapply(seq_len(n_groups), function(g) arr[g, , ])
|
||||||
} else {
|
} else {
|
||||||
arr <- array(ret, c(n_col1, n_group, n_row),
|
## remove the first axis (group)
|
||||||
dimnames = list(cnames, NULL, NULL)) %>% aperm(c(2, 3, 1)) # [group, row, col]
|
dn <- dimnames(arr)
|
||||||
lapply(seq_len(n_group), function(g) arr[g, , ])
|
matrix(arr[1, , ], nrow = dim(arr)[2], ncol = dim(arr)[3], dimnames = c(dn[2], dn[3]))
|
||||||
}
|
}
|
||||||
} else if (predinteraction) {
|
} else if (predinteraction) {
|
||||||
n_col1 <- ncol(newdata) + 1
|
## Predict interaction
|
||||||
n_group <- npred_per_case / n_col1^2
|
arr <- aperm(a = arr, perm = c(3, 4, 1, 2)) # [group, row, col, col]
|
||||||
cnames <- if (!is.null(colnames(newdata))) c(colnames(newdata), "BIAS") else NULL
|
arr <- if (n_ret == n_row) {
|
||||||
ret <- if (n_ret == n_row) {
|
matrix(arr, ncol = 1, dimnames = list(NULL, cnames))
|
||||||
matrix(ret, ncol = 1, dimnames = list(NULL, cnames))
|
} else if (n_groups != 1) {
|
||||||
} else if (n_group == 1) {
|
## turns array into list of matrices
|
||||||
array(ret, c(n_col1, n_col1, n_row), dimnames = list(cnames, cnames, NULL)) %>% aperm(c(3, 1, 2))
|
lapply(seq_len(n_groups), function(g) arr[g, , , ])
|
||||||
} else {
|
} else {
|
||||||
arr <- array(ret, c(n_col1, n_col1, n_group, n_row),
|
## remove the first axis (group)
|
||||||
dimnames = list(cnames, cnames, NULL, NULL)) %>% aperm(c(3, 4, 1, 2)) # [group, row, col1, col2]
|
arr <- arr[1, , , , drop = FALSE]
|
||||||
lapply(seq_len(n_group), function(g) arr[g, , , ])
|
array(arr, dim = dim(arr)[2:4], dimnames(arr)[2:4])
|
||||||
}
|
}
|
||||||
} else if (reshape && npred_per_case > 1) {
|
} else {
|
||||||
ret <- matrix(ret, nrow = n_row, byrow = TRUE)
|
## Normal prediction
|
||||||
|
arr <- if (reshape && n_groups != 1) {
|
||||||
|
matrix(arr, ncol = n_groups, byrow = TRUE)
|
||||||
|
} else {
|
||||||
|
as.vector(ret)
|
||||||
}
|
}
|
||||||
return(ret)
|
}
|
||||||
|
return(arr)
|
||||||
}
|
}
|
||||||
|
|
||||||
#' @rdname predict.xgb.Booster
|
#' @rdname predict.xgb.Booster
|
||||||
#' @export
|
#' @export
|
||||||
predict.xgb.Booster.handle <- function(object, ...) {
|
predict.xgb.Booster.handle <- function(object, ...) {
|
||||||
|
|
||||||
bst <- xgb.handleToBooster(object)
|
bst <- xgb.handleToBooster(handle = object, raw = NULL)
|
||||||
|
|
||||||
ret <- predict(bst, ...)
|
ret <- predict(bst, ...)
|
||||||
return(ret)
|
return(ret)
|
||||||
@@ -558,7 +638,7 @@ xgb.attributes <- function(object) {
|
|||||||
#' @export
|
#' @export
|
||||||
xgb.config <- function(object) {
|
xgb.config <- function(object) {
|
||||||
handle <- xgb.get.handle(object)
|
handle <- xgb.get.handle(object)
|
||||||
.Call(XGBoosterSaveJsonConfig_R, handle);
|
.Call(XGBoosterSaveJsonConfig_R, handle)
|
||||||
}
|
}
|
||||||
|
|
||||||
#' @rdname xgb.config
|
#' @rdname xgb.config
|
||||||
@@ -600,7 +680,7 @@ xgb.config <- function(object) {
|
|||||||
if (is.null(names(p)) || any(nchar(names(p)) == 0)) {
|
if (is.null(names(p)) || any(nchar(names(p)) == 0)) {
|
||||||
stop("parameter names cannot be empty strings")
|
stop("parameter names cannot be empty strings")
|
||||||
}
|
}
|
||||||
names(p) <- gsub("\\.", "_", names(p))
|
names(p) <- gsub(".", "_", names(p), fixed = TRUE)
|
||||||
p <- lapply(p, function(x) as.character(x)[1])
|
p <- lapply(p, function(x) as.character(x)[1])
|
||||||
handle <- xgb.get.handle(object)
|
handle <- xgb.get.handle(object)
|
||||||
for (i in seq_along(p)) {
|
for (i in seq_along(p)) {
|
||||||
|
|||||||
@@ -1,26 +1,29 @@
|
|||||||
#' Construct xgb.DMatrix object
|
#' Construct xgb.DMatrix object
|
||||||
#'
|
#'
|
||||||
#' Construct xgb.DMatrix object from either a dense matrix, a sparse matrix, or a local file.
|
#' Construct xgb.DMatrix object from either a dense matrix, a sparse matrix, or a local file.
|
||||||
#' Supported input file formats are either a libsvm text file or a binary file that was created previously by
|
#' Supported input file formats are either a LIBSVM text file or a binary file that was created previously by
|
||||||
#' \code{\link{xgb.DMatrix.save}}).
|
#' \code{\link{xgb.DMatrix.save}}).
|
||||||
#'
|
#'
|
||||||
#' @param data a \code{matrix} object (either numeric or integer), a \code{dgCMatrix} object, or a character
|
#' @param data a \code{matrix} object (either numeric or integer), a \code{dgCMatrix} object,
|
||||||
#' string representing a filename.
|
#' a \code{dgRMatrix} object (only when making predictions from a fitted model),
|
||||||
|
#' a \code{dsparseVector} object (only when making predictions from a fitted model, will be
|
||||||
|
#' interpreted as a row vector), or a character string representing a filename.
|
||||||
#' @param info a named list of additional information to store in the \code{xgb.DMatrix} object.
|
#' @param info a named list of additional information to store in the \code{xgb.DMatrix} object.
|
||||||
#' See \code{\link{setinfo}} for the specific allowed kinds of
|
#' See \code{\link{setinfo}} for the specific allowed kinds of
|
||||||
#' @param missing a float value to represents missing values in data (used only when input is a dense matrix).
|
#' @param missing a float value to represents missing values in data (used only when input is a dense matrix).
|
||||||
#' It is useful when a 0 or some other extreme value represents missing values in data.
|
#' It is useful when a 0 or some other extreme value represents missing values in data.
|
||||||
#' @param silent whether to suppress printing an informational message after loading from a file.
|
#' @param silent whether to suppress printing an informational message after loading from a file.
|
||||||
|
#' @param nthread Number of threads used for creating DMatrix.
|
||||||
#' @param ... the \code{info} data could be passed directly as parameters, without creating an \code{info} list.
|
#' @param ... the \code{info} data could be passed directly as parameters, without creating an \code{info} list.
|
||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
#' xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
||||||
#' dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
#' dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
||||||
#' if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
#' if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
||||||
#' @export
|
#' @export
|
||||||
xgb.DMatrix <- function(data, info = list(), missing = NA, silent = FALSE, ...) {
|
xgb.DMatrix <- function(data, info = list(), missing = NA, silent = FALSE, nthread = NULL, ...) {
|
||||||
cnames <- NULL
|
cnames <- NULL
|
||||||
if (typeof(data) == "character") {
|
if (typeof(data) == "character") {
|
||||||
if (length(data) > 1)
|
if (length(data) > 1)
|
||||||
@@ -29,16 +32,50 @@ xgb.DMatrix <- function(data, info = list(), missing = NA, silent = FALSE, ...)
|
|||||||
data <- path.expand(data)
|
data <- path.expand(data)
|
||||||
handle <- .Call(XGDMatrixCreateFromFile_R, data, as.integer(silent))
|
handle <- .Call(XGDMatrixCreateFromFile_R, data, as.integer(silent))
|
||||||
} else if (is.matrix(data)) {
|
} else if (is.matrix(data)) {
|
||||||
handle <- .Call(XGDMatrixCreateFromMat_R, data, missing)
|
handle <- .Call(XGDMatrixCreateFromMat_R, data, missing, as.integer(NVL(nthread, -1)))
|
||||||
cnames <- colnames(data)
|
cnames <- colnames(data)
|
||||||
} else if (inherits(data, "dgCMatrix")) {
|
} else if (inherits(data, "dgCMatrix")) {
|
||||||
handle <- .Call(XGDMatrixCreateFromCSC_R, data@p, data@i, data@x, nrow(data))
|
handle <- .Call(
|
||||||
|
XGDMatrixCreateFromCSC_R,
|
||||||
|
data@p,
|
||||||
|
data@i,
|
||||||
|
data@x,
|
||||||
|
nrow(data),
|
||||||
|
missing,
|
||||||
|
as.integer(NVL(nthread, -1))
|
||||||
|
)
|
||||||
cnames <- colnames(data)
|
cnames <- colnames(data)
|
||||||
|
} else if (inherits(data, "dgRMatrix")) {
|
||||||
|
handle <- .Call(
|
||||||
|
XGDMatrixCreateFromCSR_R,
|
||||||
|
data@p,
|
||||||
|
data@j,
|
||||||
|
data@x,
|
||||||
|
ncol(data),
|
||||||
|
missing,
|
||||||
|
as.integer(NVL(nthread, -1))
|
||||||
|
)
|
||||||
|
cnames <- colnames(data)
|
||||||
|
} else if (inherits(data, "dsparseVector")) {
|
||||||
|
indptr <- c(0L, as.integer(length(data@i)))
|
||||||
|
ind <- as.integer(data@i) - 1L
|
||||||
|
handle <- .Call(
|
||||||
|
XGDMatrixCreateFromCSR_R,
|
||||||
|
indptr,
|
||||||
|
ind,
|
||||||
|
data@x,
|
||||||
|
length(data),
|
||||||
|
missing,
|
||||||
|
as.integer(NVL(nthread, -1))
|
||||||
|
)
|
||||||
} else {
|
} else {
|
||||||
stop("xgb.DMatrix does not support construction from ", typeof(data))
|
stop("xgb.DMatrix does not support construction from ", typeof(data))
|
||||||
}
|
}
|
||||||
dmat <- handle
|
dmat <- handle
|
||||||
attributes(dmat) <- list(.Dimnames = list(NULL, cnames), class = "xgb.DMatrix")
|
attributes(dmat) <- list(class = "xgb.DMatrix")
|
||||||
|
if (!is.null(cnames)) {
|
||||||
|
setinfo(dmat, "feature_name", cnames)
|
||||||
|
}
|
||||||
|
|
||||||
info <- append(info, list(...))
|
info <- append(info, list(...))
|
||||||
for (i in seq_along(info)) {
|
for (i in seq_along(info)) {
|
||||||
@@ -51,12 +88,12 @@ xgb.DMatrix <- function(data, info = list(), missing = NA, silent = FALSE, ...)
|
|||||||
|
|
||||||
# get dmatrix from data, label
|
# get dmatrix from data, label
|
||||||
# internal helper method
|
# internal helper method
|
||||||
xgb.get.DMatrix <- function(data, label = NULL, missing = NA, weight = NULL) {
|
xgb.get.DMatrix <- function(data, label, missing, weight, nthread) {
|
||||||
if (inherits(data, "dgCMatrix") || is.matrix(data)) {
|
if (inherits(data, "dgCMatrix") || is.matrix(data)) {
|
||||||
if (is.null(label)) {
|
if (is.null(label)) {
|
||||||
stop("label must be provided when data is a matrix")
|
stop("label must be provided when data is a matrix")
|
||||||
}
|
}
|
||||||
dtrain <- xgb.DMatrix(data, label = label, missing = missing)
|
dtrain <- xgb.DMatrix(data, label = label, missing = missing, nthread = nthread)
|
||||||
if (!is.null(weight)) {
|
if (!is.null(weight)) {
|
||||||
setinfo(dtrain, "weight", weight)
|
setinfo(dtrain, "weight", weight)
|
||||||
}
|
}
|
||||||
@@ -91,7 +128,7 @@ xgb.get.DMatrix <- function(data, label = NULL, missing = NA, weight = NULL) {
|
|||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' train <- agaricus.train
|
#' train <- agaricus.train
|
||||||
#' dtrain <- xgb.DMatrix(train$data, label=train$label)
|
#' dtrain <- xgb.DMatrix(train$data, label=train$label, nthread = 2)
|
||||||
#'
|
#'
|
||||||
#' stopifnot(nrow(dtrain) == nrow(train$data))
|
#' stopifnot(nrow(dtrain) == nrow(train$data))
|
||||||
#' stopifnot(ncol(dtrain) == ncol(train$data))
|
#' stopifnot(ncol(dtrain) == ncol(train$data))
|
||||||
@@ -119,7 +156,7 @@ dim.xgb.DMatrix <- function(x) {
|
|||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' train <- agaricus.train
|
#' train <- agaricus.train
|
||||||
#' dtrain <- xgb.DMatrix(train$data, label=train$label)
|
#' dtrain <- xgb.DMatrix(train$data, label=train$label, nthread = 2)
|
||||||
#' dimnames(dtrain)
|
#' dimnames(dtrain)
|
||||||
#' colnames(dtrain)
|
#' colnames(dtrain)
|
||||||
#' colnames(dtrain) <- make.names(1:ncol(train$data))
|
#' colnames(dtrain) <- make.names(1:ncol(train$data))
|
||||||
@@ -128,7 +165,9 @@ dim.xgb.DMatrix <- function(x) {
|
|||||||
#' @rdname dimnames.xgb.DMatrix
|
#' @rdname dimnames.xgb.DMatrix
|
||||||
#' @export
|
#' @export
|
||||||
dimnames.xgb.DMatrix <- function(x) {
|
dimnames.xgb.DMatrix <- function(x) {
|
||||||
attr(x, '.Dimnames')
|
fn <- getinfo(x, "feature_name")
|
||||||
|
## row names is null.
|
||||||
|
list(NULL, fn)
|
||||||
}
|
}
|
||||||
|
|
||||||
#' @rdname dimnames.xgb.DMatrix
|
#' @rdname dimnames.xgb.DMatrix
|
||||||
@@ -139,13 +178,13 @@ dimnames.xgb.DMatrix <- function(x) {
|
|||||||
if (!is.null(value[[1L]]))
|
if (!is.null(value[[1L]]))
|
||||||
stop("xgb.DMatrix does not have rownames")
|
stop("xgb.DMatrix does not have rownames")
|
||||||
if (is.null(value[[2]])) {
|
if (is.null(value[[2]])) {
|
||||||
attr(x, '.Dimnames') <- NULL
|
setinfo(x, "feature_name", NULL)
|
||||||
return(x)
|
return(x)
|
||||||
}
|
}
|
||||||
if (ncol(x) != length(value[[2]]))
|
if (ncol(x) != length(value[[2]])) {
|
||||||
stop("can't assign ", length(value[[2]]), " colnames to a ",
|
stop("can't assign ", length(value[[2]]), " colnames to a ", ncol(x), " column xgb.DMatrix")
|
||||||
ncol(x), " column xgb.DMatrix")
|
}
|
||||||
attr(x, '.Dimnames') <- value
|
setinfo(x, "feature_name", value[[2]])
|
||||||
x
|
x
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -161,9 +200,9 @@ dimnames.xgb.DMatrix <- function(x) {
|
|||||||
#' The \code{name} field can be one of the following:
|
#' The \code{name} field can be one of the following:
|
||||||
#'
|
#'
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{label}: label Xgboost learn from ;
|
#' \item \code{label}: label XGBoost learn from ;
|
||||||
#' \item \code{weight}: to do a weight rescale ;
|
#' \item \code{weight}: to do a weight rescale ;
|
||||||
#' \item \code{base_margin}: base margin is the base prediction Xgboost will boost from ;
|
#' \item \code{base_margin}: base margin is the base prediction XGBoost will boost from ;
|
||||||
#' \item \code{nrow}: number of rows of the \code{xgb.DMatrix}.
|
#' \item \code{nrow}: number of rows of the \code{xgb.DMatrix}.
|
||||||
#'
|
#'
|
||||||
#' }
|
#' }
|
||||||
@@ -172,7 +211,7 @@ dimnames.xgb.DMatrix <- function(x) {
|
|||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#'
|
#'
|
||||||
#' labels <- getinfo(dtrain, 'label')
|
#' labels <- getinfo(dtrain, 'label')
|
||||||
#' setinfo(dtrain, 'label', 1-labels)
|
#' setinfo(dtrain, 'label', 1-labels)
|
||||||
@@ -189,11 +228,15 @@ getinfo.xgb.DMatrix <- function(object, name, ...) {
|
|||||||
if (typeof(name) != "character" ||
|
if (typeof(name) != "character" ||
|
||||||
length(name) != 1 ||
|
length(name) != 1 ||
|
||||||
!name %in% c('label', 'weight', 'base_margin', 'nrow',
|
!name %in% c('label', 'weight', 'base_margin', 'nrow',
|
||||||
'label_lower_bound', 'label_upper_bound')) {
|
'label_lower_bound', 'label_upper_bound', "feature_type", "feature_name")) {
|
||||||
stop("getinfo: name must be one of the following\n",
|
stop(
|
||||||
" 'label', 'weight', 'base_margin', 'nrow', 'label_lower_bound', 'label_upper_bound'")
|
"getinfo: name must be one of the following\n",
|
||||||
|
" 'label', 'weight', 'base_margin', 'nrow', 'label_lower_bound', 'label_upper_bound', 'feature_type', 'feature_name'"
|
||||||
|
)
|
||||||
}
|
}
|
||||||
if (name != "nrow"){
|
if (name == "feature_name" || name == "feature_type") {
|
||||||
|
ret <- .Call(XGDMatrixGetStrFeatureInfo_R, object, name)
|
||||||
|
} else if (name != "nrow") {
|
||||||
ret <- .Call(XGDMatrixGetInfo_R, object, name)
|
ret <- .Call(XGDMatrixGetInfo_R, object, name)
|
||||||
} else {
|
} else {
|
||||||
ret <- nrow(object)
|
ret <- nrow(object)
|
||||||
@@ -216,15 +259,15 @@ getinfo.xgb.DMatrix <- function(object, name, ...) {
|
|||||||
#' The \code{name} field can be one of the following:
|
#' The \code{name} field can be one of the following:
|
||||||
#'
|
#'
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{label}: label Xgboost learn from ;
|
#' \item \code{label}: label XGBoost learn from ;
|
||||||
#' \item \code{weight}: to do a weight rescale ;
|
#' \item \code{weight}: to do a weight rescale ;
|
||||||
#' \item \code{base_margin}: base margin is the base prediction Xgboost will boost from ;
|
#' \item \code{base_margin}: base margin is the base prediction XGBoost will boost from ;
|
||||||
#' \item \code{group}: number of rows in each group (to use with \code{rank:pairwise} objective).
|
#' \item \code{group}: number of rows in each group (to use with \code{rank:pairwise} objective).
|
||||||
#' }
|
#' }
|
||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#'
|
#'
|
||||||
#' labels <- getinfo(dtrain, 'label')
|
#' labels <- getinfo(dtrain, 'label')
|
||||||
#' setinfo(dtrain, 'label', 1-labels)
|
#' setinfo(dtrain, 'label', 1-labels)
|
||||||
@@ -271,8 +314,38 @@ setinfo.xgb.DMatrix <- function(object, name, info, ...) {
|
|||||||
.Call(XGDMatrixSetInfo_R, object, name, as.integer(info))
|
.Call(XGDMatrixSetInfo_R, object, name, as.integer(info))
|
||||||
return(TRUE)
|
return(TRUE)
|
||||||
}
|
}
|
||||||
|
if (name == "feature_weights") {
|
||||||
|
if (length(info) != ncol(object)) {
|
||||||
|
stop("The number of feature weights must equal to the number of columns in the input data")
|
||||||
|
}
|
||||||
|
.Call(XGDMatrixSetInfo_R, object, name, as.numeric(info))
|
||||||
|
return(TRUE)
|
||||||
|
}
|
||||||
|
|
||||||
|
set_feat_info <- function(name) {
|
||||||
|
msg <- sprintf(
|
||||||
|
"The number of %s must equal to the number of columns in the input data. %s vs. %s",
|
||||||
|
name,
|
||||||
|
length(info),
|
||||||
|
ncol(object)
|
||||||
|
)
|
||||||
|
if (!is.null(info)) {
|
||||||
|
info <- as.list(info)
|
||||||
|
if (length(info) != ncol(object)) {
|
||||||
|
stop(msg)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
.Call(XGDMatrixSetStrFeatureInfo_R, object, name, info)
|
||||||
|
}
|
||||||
|
if (name == "feature_name") {
|
||||||
|
set_feat_info("feature_name")
|
||||||
|
return(TRUE)
|
||||||
|
}
|
||||||
|
if (name == "feature_type") {
|
||||||
|
set_feat_info("feature_type")
|
||||||
|
return(TRUE)
|
||||||
|
}
|
||||||
stop("setinfo: unknown info name ", name)
|
stop("setinfo: unknown info name ", name)
|
||||||
return(FALSE)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@@ -289,7 +362,7 @@ setinfo.xgb.DMatrix <- function(object, name, info, ...) {
|
|||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#'
|
#'
|
||||||
#' dsub <- slice(dtrain, 1:42)
|
#' dsub <- slice(dtrain, 1:42)
|
||||||
#' labels1 <- getinfo(dsub, 'label')
|
#' labels1 <- getinfo(dsub, 'label')
|
||||||
@@ -345,7 +418,7 @@ slice.xgb.DMatrix <- function(object, idxset, ...) {
|
|||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#'
|
#'
|
||||||
#' dtrain
|
#' dtrain
|
||||||
#' print(dtrain, verbose=TRUE)
|
#' print(dtrain, verbose=TRUE)
|
||||||
@@ -362,7 +435,7 @@ print.xgb.DMatrix <- function(x, verbose = FALSE, ...) {
|
|||||||
cat(infos)
|
cat(infos)
|
||||||
cnames <- colnames(x)
|
cnames <- colnames(x)
|
||||||
cat(' colnames:')
|
cat(' colnames:')
|
||||||
if (verbose & !is.null(cnames)) {
|
if (verbose && !is.null(cnames)) {
|
||||||
cat("\n'")
|
cat("\n'")
|
||||||
cat(cnames, sep = "','")
|
cat(cnames, sep = "','")
|
||||||
cat("'")
|
cat("'")
|
||||||
|
|||||||
@@ -7,7 +7,7 @@
|
|||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
#' xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
||||||
#' dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
#' dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
||||||
#' if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
#' if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
||||||
|
|||||||
@@ -18,7 +18,7 @@
|
|||||||
#'
|
#'
|
||||||
#' International Workshop on Data Mining for Online Advertising (ADKDD) - August 24, 2014
|
#' International Workshop on Data Mining for Online Advertising (ADKDD) - August 24, 2014
|
||||||
#'
|
#'
|
||||||
#' \url{https://research.fb.com/publications/practical-lessons-from-predicting-clicks-on-ads-at-facebook/}.
|
#' \url{https://research.facebook.com/publications/practical-lessons-from-predicting-clicks-on-ads-at-facebook/}.
|
||||||
#'
|
#'
|
||||||
#' Extract explaining the method:
|
#' Extract explaining the method:
|
||||||
#'
|
#'
|
||||||
@@ -48,8 +48,8 @@
|
|||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' data(agaricus.test, package='xgboost')
|
#' data(agaricus.test, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' dtest <- with(agaricus.test, xgb.DMatrix(data, label = label))
|
#' dtest <- with(agaricus.test, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#'
|
#'
|
||||||
#' param <- list(max_depth=2, eta=1, silent=1, objective='binary:logistic')
|
#' param <- list(max_depth=2, eta=1, silent=1, objective='binary:logistic')
|
||||||
#' nrounds = 4
|
#' nrounds = 4
|
||||||
@@ -65,8 +65,12 @@
|
|||||||
#' new.features.test <- xgb.create.features(model = bst, agaricus.test$data)
|
#' new.features.test <- xgb.create.features(model = bst, agaricus.test$data)
|
||||||
#'
|
#'
|
||||||
#' # learning with new features
|
#' # learning with new features
|
||||||
#' new.dtrain <- xgb.DMatrix(data = new.features.train, label = agaricus.train$label)
|
#' new.dtrain <- xgb.DMatrix(
|
||||||
#' new.dtest <- xgb.DMatrix(data = new.features.test, label = agaricus.test$label)
|
#' data = new.features.train, label = agaricus.train$label, nthread = 2
|
||||||
|
#' )
|
||||||
|
#' new.dtest <- xgb.DMatrix(
|
||||||
|
#' data = new.features.test, label = agaricus.test$label, nthread = 2
|
||||||
|
#' )
|
||||||
#' watchlist <- list(train = new.dtrain)
|
#' watchlist <- list(train = new.dtrain)
|
||||||
#' bst <- xgb.train(params = param, data = new.dtrain, nrounds = nrounds, nthread = 2)
|
#' bst <- xgb.train(params = param, data = new.dtrain, nrounds = nrounds, nthread = 2)
|
||||||
#'
|
#'
|
||||||
|
|||||||
@@ -75,9 +75,11 @@
|
|||||||
#' @details
|
#' @details
|
||||||
#' The original sample is randomly partitioned into \code{nfold} equal size subsamples.
|
#' The original sample is randomly partitioned into \code{nfold} equal size subsamples.
|
||||||
#'
|
#'
|
||||||
#' Of the \code{nfold} subsamples, a single subsample is retained as the validation data for testing the model, and the remaining \code{nfold - 1} subsamples are used as training data.
|
#' Of the \code{nfold} subsamples, a single subsample is retained as the validation data for testing the model,
|
||||||
|
#' and the remaining \code{nfold - 1} subsamples are used as training data.
|
||||||
#'
|
#'
|
||||||
#' The cross-validation process is then repeated \code{nrounds} times, with each of the \code{nfold} subsamples used exactly once as the validation data.
|
#' The cross-validation process is then repeated \code{nrounds} times, with each of the
|
||||||
|
#' \code{nfold} subsamples used exactly once as the validation data.
|
||||||
#'
|
#'
|
||||||
#' All observations are used for both training and validation.
|
#' All observations are used for both training and validation.
|
||||||
#'
|
#'
|
||||||
@@ -101,9 +103,7 @@
|
|||||||
#' parameter or randomly generated.
|
#' parameter or randomly generated.
|
||||||
#' \item \code{best_iteration} iteration number with the best evaluation metric value
|
#' \item \code{best_iteration} iteration number with the best evaluation metric value
|
||||||
#' (only available with early stopping).
|
#' (only available with early stopping).
|
||||||
#' \item \code{best_ntreelimit} the \code{ntreelimit} value corresponding to the best iteration,
|
#' \item \code{best_ntreelimit} and the \code{ntreelimit} Deprecated attributes, use \code{best_iteration} instead.
|
||||||
#' which could further be used in \code{predict} method
|
|
||||||
#' (only available with early stopping).
|
|
||||||
#' \item \code{pred} CV prediction values available when \code{prediction} is set.
|
#' \item \code{pred} CV prediction values available when \code{prediction} is set.
|
||||||
#' It is either vector or matrix (see \code{\link{cb.cv.predict}}).
|
#' It is either vector or matrix (see \code{\link{cb.cv.predict}}).
|
||||||
#' \item \code{models} a list of the CV folds' models. It is only available with the explicit
|
#' \item \code{models} a list of the CV folds' models. It is only available with the explicit
|
||||||
@@ -112,7 +112,7 @@
|
|||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' cv <- xgb.cv(data = dtrain, nrounds = 3, nthread = 2, nfold = 5, metrics = list("rmse","auc"),
|
#' cv <- xgb.cv(data = dtrain, nrounds = 3, nthread = 2, nfold = 5, metrics = list("rmse","auc"),
|
||||||
#' max_depth = 3, eta = 1, objective = "binary:logistic")
|
#' max_depth = 3, eta = 1, objective = "binary:logistic")
|
||||||
#' print(cv)
|
#' print(cv)
|
||||||
@@ -135,9 +135,6 @@ xgb.cv <- function(params=list(), data, nrounds, nfold, label = NULL, missing =
|
|||||||
check.custom.obj()
|
check.custom.obj()
|
||||||
check.custom.eval()
|
check.custom.eval()
|
||||||
|
|
||||||
#if (is.null(params[['eval_metric']]) && is.null(feval))
|
|
||||||
# stop("Either 'eval_metric' or 'feval' must be provided for CV")
|
|
||||||
|
|
||||||
# Check the labels
|
# Check the labels
|
||||||
if ((inherits(data, 'xgb.DMatrix') && is.null(getinfo(data, 'label'))) ||
|
if ((inherits(data, 'xgb.DMatrix') && is.null(getinfo(data, 'label'))) ||
|
||||||
(!inherits(data, 'xgb.DMatrix') && is.null(label))) {
|
(!inherits(data, 'xgb.DMatrix') && is.null(label))) {
|
||||||
@@ -161,10 +158,6 @@ xgb.cv <- function(params=list(), data, nrounds, nfold, label = NULL, missing =
|
|||||||
folds <- generate.cv.folds(nfold, nrow(data), stratified, cv_label, params)
|
folds <- generate.cv.folds(nfold, nrow(data), stratified, cv_label, params)
|
||||||
}
|
}
|
||||||
|
|
||||||
# Potential TODO: sequential CV
|
|
||||||
#if (strategy == 'sequential')
|
|
||||||
# stop('Sequential CV strategy is not yet implemented')
|
|
||||||
|
|
||||||
# verbosity & evaluation printing callback:
|
# verbosity & evaluation printing callback:
|
||||||
params <- c(params, list(silent = 1))
|
params <- c(params, list(silent = 1))
|
||||||
print_every_n <- max(as.integer(print_every_n), 1L)
|
print_every_n <- max(as.integer(print_every_n), 1L)
|
||||||
@@ -194,7 +187,13 @@ xgb.cv <- function(params=list(), data, nrounds, nfold, label = NULL, missing =
|
|||||||
|
|
||||||
# create the booster-folds
|
# create the booster-folds
|
||||||
# train_folds
|
# train_folds
|
||||||
dall <- xgb.get.DMatrix(data, label, missing)
|
dall <- xgb.get.DMatrix(
|
||||||
|
data = data,
|
||||||
|
label = label,
|
||||||
|
missing = missing,
|
||||||
|
weight = NULL,
|
||||||
|
nthread = params$nthread
|
||||||
|
)
|
||||||
bst_folds <- lapply(seq_along(folds), function(k) {
|
bst_folds <- lapply(seq_along(folds), function(k) {
|
||||||
dtest <- slice(dall, folds[[k]])
|
dtest <- slice(dall, folds[[k]])
|
||||||
# code originally contributed by @RolandASc on stackoverflow
|
# code originally contributed by @RolandASc on stackoverflow
|
||||||
@@ -202,7 +201,12 @@ xgb.cv <- function(params=list(), data, nrounds, nfold, label = NULL, missing =
|
|||||||
dtrain <- slice(dall, unlist(folds[-k]))
|
dtrain <- slice(dall, unlist(folds[-k]))
|
||||||
else
|
else
|
||||||
dtrain <- slice(dall, train_folds[[k]])
|
dtrain <- slice(dall, train_folds[[k]])
|
||||||
handle <- xgb.Booster.handle(params, list(dtrain, dtest))
|
handle <- xgb.Booster.handle(
|
||||||
|
params = params,
|
||||||
|
cachelist = list(dtrain, dtest),
|
||||||
|
modelfile = NULL,
|
||||||
|
handle = NULL
|
||||||
|
)
|
||||||
list(dtrain = dtrain, bst = handle, watchlist = list(train = dtrain, test = dtest), index = folds[[k]])
|
list(dtrain = dtrain, bst = handle, watchlist = list(train = dtrain, test = dtest), index = folds[[k]])
|
||||||
})
|
})
|
||||||
rm(dall)
|
rm(dall)
|
||||||
@@ -223,8 +227,18 @@ xgb.cv <- function(params=list(), data, nrounds, nfold, label = NULL, missing =
|
|||||||
for (f in cb$pre_iter) f()
|
for (f in cb$pre_iter) f()
|
||||||
|
|
||||||
msg <- lapply(bst_folds, function(fd) {
|
msg <- lapply(bst_folds, function(fd) {
|
||||||
xgb.iter.update(fd$bst, fd$dtrain, iteration - 1, obj)
|
xgb.iter.update(
|
||||||
xgb.iter.eval(fd$bst, fd$watchlist, iteration - 1, feval)
|
booster_handle = fd$bst,
|
||||||
|
dtrain = fd$dtrain,
|
||||||
|
iter = iteration - 1,
|
||||||
|
obj = obj
|
||||||
|
)
|
||||||
|
xgb.iter.eval(
|
||||||
|
booster_handle = fd$bst,
|
||||||
|
watchlist = fd$watchlist,
|
||||||
|
iter = iteration - 1,
|
||||||
|
feval = feval
|
||||||
|
)
|
||||||
})
|
})
|
||||||
msg <- simplify2array(msg)
|
msg <- simplify2array(msg)
|
||||||
bst_evaluation <- rowMeans(msg)
|
bst_evaluation <- rowMeans(msg)
|
||||||
|
|||||||
@@ -6,8 +6,6 @@
|
|||||||
#' @param fname the name of the text file where to save the model text dump.
|
#' @param fname the name of the text file where to save the model text dump.
|
||||||
#' If not provided or set to \code{NULL}, the model is returned as a \code{character} vector.
|
#' If not provided or set to \code{NULL}, the model is returned as a \code{character} vector.
|
||||||
#' @param fmap feature map file representing feature types.
|
#' @param fmap feature map file representing feature types.
|
||||||
#' Detailed description could be found at
|
|
||||||
#' \url{https://github.com/dmlc/xgboost/wiki/Binary-Classification#dump-model}.
|
|
||||||
#' See demo/ for walkthrough example in R, and
|
#' See demo/ for walkthrough example in R, and
|
||||||
#' \url{https://github.com/dmlc/xgboost/blob/master/demo/data/featmap.txt}
|
#' \url{https://github.com/dmlc/xgboost/blob/master/demo/data/featmap.txt}
|
||||||
#' for example Format.
|
#' for example Format.
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
#' @rdname xgb.plot.importance
|
#' @rdname xgb.plot.importance
|
||||||
#' @export
|
#' @export
|
||||||
xgb.ggplot.importance <- function(importance_matrix = NULL, top_n = NULL, measure = NULL,
|
xgb.ggplot.importance <- function(importance_matrix = NULL, top_n = NULL, measure = NULL,
|
||||||
rel_to_first = FALSE, n_clusters = c(1:10), ...) {
|
rel_to_first = FALSE, n_clusters = seq_len(10), ...) {
|
||||||
|
|
||||||
importance_matrix <- xgb.plot.importance(importance_matrix, top_n = top_n, measure = measure,
|
importance_matrix <- xgb.plot.importance(importance_matrix, top_n = top_n, measure = measure,
|
||||||
rel_to_first = rel_to_first, plot = FALSE, ...)
|
rel_to_first = rel_to_first, plot = FALSE, ...)
|
||||||
@@ -142,6 +142,7 @@ xgb.ggplot.shap.summary <- function(data, shap_contrib = NULL, features = NULL,
|
|||||||
#'
|
#'
|
||||||
#' @return A data.table containing the observation ID, the feature name, the
|
#' @return A data.table containing the observation ID, the feature name, the
|
||||||
#' feature value (normalized if specified), and the SHAP contribution value.
|
#' feature value (normalized if specified), and the SHAP contribution value.
|
||||||
|
#' @noRd
|
||||||
prepare.ggplot.shap.data <- function(data_list, normalize = FALSE) {
|
prepare.ggplot.shap.data <- function(data_list, normalize = FALSE) {
|
||||||
data <- data_list[["data"]]
|
data <- data_list[["data"]]
|
||||||
shap_contrib <- data_list[["shap_contrib"]]
|
shap_contrib <- data_list[["shap_contrib"]]
|
||||||
@@ -170,6 +171,7 @@ prepare.ggplot.shap.data <- function(data_list, normalize = FALSE) {
|
|||||||
#' @param x Numeric vector
|
#' @param x Numeric vector
|
||||||
#'
|
#'
|
||||||
#' @return Numeric vector with mean 0 and sd 1.
|
#' @return Numeric vector with mean 0 and sd 1.
|
||||||
|
#' @noRd
|
||||||
normalize <- function(x) {
|
normalize <- function(x) {
|
||||||
loc <- mean(x, na.rm = TRUE)
|
loc <- mean(x, na.rm = TRUE)
|
||||||
scale <- stats::sd(x, na.rm = TRUE)
|
scale <- stats::sd(x, na.rm = TRUE)
|
||||||
@@ -181,7 +183,7 @@ normalize <- function(x) {
|
|||||||
# ... the plots
|
# ... the plots
|
||||||
# cols number of columns
|
# cols number of columns
|
||||||
# internal utility function
|
# internal utility function
|
||||||
multiplot <- function(..., cols = 1) {
|
multiplot <- function(..., cols) {
|
||||||
plots <- list(...)
|
plots <- list(...)
|
||||||
num_plots <- length(plots)
|
num_plots <- length(plots)
|
||||||
|
|
||||||
|
|||||||
@@ -96,40 +96,48 @@ xgb.importance <- function(feature_names = NULL, model = NULL, trees = NULL,
|
|||||||
if (!(is.null(feature_names) || is.character(feature_names)))
|
if (!(is.null(feature_names) || is.character(feature_names)))
|
||||||
stop("feature_names: Has to be a character vector")
|
stop("feature_names: Has to be a character vector")
|
||||||
|
|
||||||
model_text_dump <- xgb.dump(model = model, with_stats = TRUE)
|
model <- xgb.Booster.complete(model)
|
||||||
|
config <- jsonlite::fromJSON(xgb.config(model))
|
||||||
# linear model
|
if (config$learner$gradient_booster$name == "gblinear") {
|
||||||
if (model_text_dump[2] == "bias:"){
|
args <- list(importance_type = "weight", feature_names = feature_names)
|
||||||
weights <- which(model_text_dump == "weight:") %>%
|
results <- .Call(
|
||||||
{model_text_dump[(. + 1):length(model_text_dump)]} %>%
|
XGBoosterFeatureScore_R, model$handle, jsonlite::toJSON(args, auto_unbox = TRUE, null = "null")
|
||||||
as.numeric
|
)
|
||||||
|
names(results) <- c("features", "shape", "weight")
|
||||||
num_class <- NVL(model$params$num_class, 1)
|
if (length(results$shape) == 2) {
|
||||||
if (is.null(feature_names))
|
n_classes <- results$shape[2]
|
||||||
feature_names <- seq(to = length(weights) / num_class) - 1
|
|
||||||
if (length(feature_names) * num_class != length(weights))
|
|
||||||
stop("feature_names length does not match the number of features used in the model")
|
|
||||||
|
|
||||||
result <- if (num_class == 1) {
|
|
||||||
data.table(Feature = feature_names, Weight = weights)[order(-abs(Weight))]
|
|
||||||
} else {
|
} else {
|
||||||
data.table(Feature = rep(feature_names, each = num_class),
|
n_classes <- 0
|
||||||
Weight = weights,
|
|
||||||
Class = seq_len(num_class) - 1)[order(Class, -abs(Weight))]
|
|
||||||
}
|
}
|
||||||
} else { # tree model
|
importance <- if (n_classes == 0) {
|
||||||
result <- xgb.model.dt.tree(feature_names = feature_names,
|
data.table(Feature = results$features, Weight = results$weight)[order(-abs(Weight))]
|
||||||
text = model_text_dump,
|
} else {
|
||||||
trees = trees)[
|
data.table(
|
||||||
Feature != "Leaf", .(Gain = sum(Quality),
|
Feature = rep(results$features, each = n_classes), Weight = results$weight, Class = seq_len(n_classes) - 1
|
||||||
Cover = sum(Cover),
|
)[order(Class, -abs(Weight))]
|
||||||
Frequency = .N), by = Feature][
|
|
||||||
, `:=`(Gain = Gain / sum(Gain),
|
|
||||||
Cover = Cover / sum(Cover),
|
|
||||||
Frequency = Frequency / sum(Frequency))][
|
|
||||||
order(Gain, decreasing = TRUE)]
|
|
||||||
}
|
}
|
||||||
result
|
} else {
|
||||||
|
concatenated <- list()
|
||||||
|
output_names <- vector()
|
||||||
|
for (importance_type in c("weight", "total_gain", "total_cover")) {
|
||||||
|
args <- list(importance_type = importance_type, feature_names = feature_names, tree_idx = trees)
|
||||||
|
results <- .Call(
|
||||||
|
XGBoosterFeatureScore_R, model$handle, jsonlite::toJSON(args, auto_unbox = TRUE, null = "null")
|
||||||
|
)
|
||||||
|
names(results) <- c("features", "shape", importance_type)
|
||||||
|
concatenated[
|
||||||
|
switch(importance_type, "weight" = "Frequency", "total_gain" = "Gain", "total_cover" = "Cover")
|
||||||
|
] <- results[importance_type]
|
||||||
|
output_names <- results$features
|
||||||
|
}
|
||||||
|
importance <- data.table(
|
||||||
|
Feature = output_names,
|
||||||
|
Gain = concatenated$Gain / sum(concatenated$Gain),
|
||||||
|
Cover = concatenated$Cover / sum(concatenated$Cover),
|
||||||
|
Frequency = concatenated$Frequency / sum(concatenated$Frequency)
|
||||||
|
)[order(Gain, decreasing = TRUE)]
|
||||||
|
}
|
||||||
|
importance
|
||||||
}
|
}
|
||||||
|
|
||||||
# Avoid error messages during CRAN check.
|
# Avoid error messages during CRAN check.
|
||||||
|
|||||||
@@ -5,7 +5,7 @@
|
|||||||
#' @param modelfile the name of the binary input file.
|
#' @param modelfile the name of the binary input file.
|
||||||
#'
|
#'
|
||||||
#' @details
|
#' @details
|
||||||
#' The input file is expected to contain a model saved in an xgboost-internal binary format
|
#' The input file is expected to contain a model saved in an xgboost model format
|
||||||
#' using either \code{\link{xgb.save}} or \code{\link{cb.save.model}} in R, or using some
|
#' using either \code{\link{xgb.save}} or \code{\link{cb.save.model}} in R, or using some
|
||||||
#' appropriate methods from other xgboost interfaces. E.g., a model trained in Python and
|
#' appropriate methods from other xgboost interfaces. E.g., a model trained in Python and
|
||||||
#' saved from there in xgboost format, could be loaded from R.
|
#' saved from there in xgboost format, could be loaded from R.
|
||||||
@@ -35,12 +35,24 @@ xgb.load <- function(modelfile) {
|
|||||||
if (is.null(modelfile))
|
if (is.null(modelfile))
|
||||||
stop("xgb.load: modelfile cannot be NULL")
|
stop("xgb.load: modelfile cannot be NULL")
|
||||||
|
|
||||||
handle <- xgb.Booster.handle(modelfile = modelfile)
|
handle <- xgb.Booster.handle(
|
||||||
|
params = list(),
|
||||||
|
cachelist = list(),
|
||||||
|
modelfile = modelfile,
|
||||||
|
handle = NULL
|
||||||
|
)
|
||||||
# re-use modelfile if it is raw so we do not need to serialize
|
# re-use modelfile if it is raw so we do not need to serialize
|
||||||
if (typeof(modelfile) == "raw") {
|
if (typeof(modelfile) == "raw") {
|
||||||
bst <- xgb.handleToBooster(handle, modelfile)
|
warning(
|
||||||
|
paste(
|
||||||
|
"The support for loading raw booster with `xgb.load` will be ",
|
||||||
|
"discontinued in upcoming release. Use `xgb.load.raw` or",
|
||||||
|
" `xgb.unserialize` instead. "
|
||||||
|
)
|
||||||
|
)
|
||||||
|
bst <- xgb.handleToBooster(handle = handle, raw = modelfile)
|
||||||
} else {
|
} else {
|
||||||
bst <- xgb.handleToBooster(handle, NULL)
|
bst <- xgb.handleToBooster(handle = handle, raw = NULL)
|
||||||
}
|
}
|
||||||
bst <- xgb.Booster.complete(bst, saveraw = TRUE)
|
bst <- xgb.Booster.complete(bst, saveraw = TRUE)
|
||||||
return(bst)
|
return(bst)
|
||||||
|
|||||||
@@ -3,12 +3,21 @@
|
|||||||
#' User can generate raw memory buffer by calling xgb.save.raw
|
#' User can generate raw memory buffer by calling xgb.save.raw
|
||||||
#'
|
#'
|
||||||
#' @param buffer the buffer returned by xgb.save.raw
|
#' @param buffer the buffer returned by xgb.save.raw
|
||||||
|
#' @param as_booster Return the loaded model as xgb.Booster instead of xgb.Booster.handle.
|
||||||
#'
|
#'
|
||||||
#' @export
|
#' @export
|
||||||
xgb.load.raw <- function(buffer) {
|
xgb.load.raw <- function(buffer, as_booster = FALSE) {
|
||||||
cachelist <- list()
|
cachelist <- list()
|
||||||
handle <- .Call(XGBoosterCreate_R, cachelist)
|
handle <- .Call(XGBoosterCreate_R, cachelist)
|
||||||
.Call(XGBoosterLoadModelFromRaw_R, handle, buffer)
|
.Call(XGBoosterLoadModelFromRaw_R, handle, buffer)
|
||||||
class(handle) <- "xgb.Booster.handle"
|
class(handle) <- "xgb.Booster.handle"
|
||||||
|
|
||||||
|
if (as_booster) {
|
||||||
|
booster <- list(handle = handle, raw = NULL)
|
||||||
|
class(booster) <- "xgb.Booster"
|
||||||
|
booster <- xgb.Booster.complete(booster, saveraw = TRUE)
|
||||||
|
return(booster)
|
||||||
|
} else {
|
||||||
return (handle)
|
return (handle)
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|||||||
@@ -86,8 +86,7 @@ xgb.model.dt.tree <- function(feature_names = NULL, model = NULL, text = NULL,
|
|||||||
text <- xgb.dump(model = model, with_stats = TRUE)
|
text <- xgb.dump(model = model, with_stats = TRUE)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (length(text) < 2 ||
|
if (length(text) < 2 || !any(grepl('leaf=(\\d+)', text))) {
|
||||||
sum(grepl('yes=(\\d+),no=(\\d+)', text)) < 1) {
|
|
||||||
stop("Non-tree model detected! This function can only be used with tree models.")
|
stop("Non-tree model detected! This function can only be used with tree models.")
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -116,16 +115,28 @@ xgb.model.dt.tree <- function(feature_names = NULL, model = NULL, text = NULL,
|
|||||||
branch_rx <- paste0("f(\\d+)<(", anynumber_regex, ")\\] yes=(\\d+),no=(\\d+),missing=(\\d+),",
|
branch_rx <- paste0("f(\\d+)<(", anynumber_regex, ")\\] yes=(\\d+),no=(\\d+),missing=(\\d+),",
|
||||||
"gain=(", anynumber_regex, "),cover=(", anynumber_regex, ")")
|
"gain=(", anynumber_regex, "),cover=(", anynumber_regex, ")")
|
||||||
branch_cols <- c("Feature", "Split", "Yes", "No", "Missing", "Quality", "Cover")
|
branch_cols <- c("Feature", "Split", "Yes", "No", "Missing", "Quality", "Cover")
|
||||||
td[isLeaf == FALSE,
|
td[
|
||||||
|
isLeaf == FALSE,
|
||||||
(branch_cols) := {
|
(branch_cols) := {
|
||||||
matches <- regmatches(t, regexec(branch_rx, t))
|
matches <- regmatches(t, regexec(branch_rx, t))
|
||||||
# skip some indices with spurious capture groups from anynumber_regex
|
# skip some indices with spurious capture groups from anynumber_regex
|
||||||
xtr <- do.call(rbind, matches)[, c(2, 3, 5, 6, 7, 8, 10), drop = FALSE]
|
xtr <- do.call(rbind, matches)[, c(2, 3, 5, 6, 7, 8, 10), drop = FALSE]
|
||||||
xtr[, 3:5] <- add.tree.id(xtr[, 3:5], Tree)
|
xtr[, 3:5] <- add.tree.id(xtr[, 3:5], Tree)
|
||||||
|
if (length(xtr) == 0) {
|
||||||
|
as.data.table(
|
||||||
|
list(Feature = "NA", Split = "NA", Yes = "NA", No = "NA", Missing = "NA", Quality = "NA", Cover = "NA")
|
||||||
|
)
|
||||||
|
} else {
|
||||||
as.data.table(xtr)
|
as.data.table(xtr)
|
||||||
}]
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
# assign feature_names when available
|
# assign feature_names when available
|
||||||
if (!is.null(feature_names)) {
|
is_stump <- function() {
|
||||||
|
return(length(td$Feature) == 1 && is.na(td$Feature))
|
||||||
|
}
|
||||||
|
if (!is.null(feature_names) && !is_stump()) {
|
||||||
if (length(feature_names) <= max(as.numeric(td$Feature), na.rm = TRUE))
|
if (length(feature_names) <= max(as.numeric(td$Feature), na.rm = TRUE))
|
||||||
stop("feature_names has less elements than there are features used in the model")
|
stop("feature_names has less elements than there are features used in the model")
|
||||||
td[isLeaf == FALSE, Feature := feature_names[as.numeric(Feature) + 1]]
|
td[isLeaf == FALSE, Feature := feature_names[as.numeric(Feature) + 1]]
|
||||||
@@ -134,12 +145,18 @@ xgb.model.dt.tree <- function(feature_names = NULL, model = NULL, text = NULL,
|
|||||||
# parse leaf lines
|
# parse leaf lines
|
||||||
leaf_rx <- paste0("leaf=(", anynumber_regex, "),cover=(", anynumber_regex, ")")
|
leaf_rx <- paste0("leaf=(", anynumber_regex, "),cover=(", anynumber_regex, ")")
|
||||||
leaf_cols <- c("Feature", "Quality", "Cover")
|
leaf_cols <- c("Feature", "Quality", "Cover")
|
||||||
td[isLeaf == TRUE,
|
td[
|
||||||
|
isLeaf == TRUE,
|
||||||
(leaf_cols) := {
|
(leaf_cols) := {
|
||||||
matches <- regmatches(t, regexec(leaf_rx, t))
|
matches <- regmatches(t, regexec(leaf_rx, t))
|
||||||
xtr <- do.call(rbind, matches)[, c(2, 4)]
|
xtr <- do.call(rbind, matches)[, c(2, 4)]
|
||||||
|
if (length(xtr) == 2) {
|
||||||
|
c("Leaf", as.data.table(xtr[1]), as.data.table(xtr[2]))
|
||||||
|
} else {
|
||||||
c("Leaf", as.data.table(xtr))
|
c("Leaf", as.data.table(xtr))
|
||||||
}]
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
# convert some columns to numeric
|
# convert some columns to numeric
|
||||||
numeric_cols <- c("Split", "Quality", "Cover")
|
numeric_cols <- c("Split", "Quality", "Cover")
|
||||||
|
|||||||
@@ -136,7 +136,7 @@ get.leaf.depth <- function(dt_tree) {
|
|||||||
# list of paths to each leaf in a tree
|
# list of paths to each leaf in a tree
|
||||||
paths <- lapply(paths_tmp$vpath, names)
|
paths <- lapply(paths_tmp$vpath, names)
|
||||||
# combine into a resulting path lengths table for a tree
|
# combine into a resulting path lengths table for a tree
|
||||||
data.table(Depth = sapply(paths, length), ID = To[Leaf == TRUE])
|
data.table(Depth = lengths(paths), ID = To[Leaf == TRUE])
|
||||||
}, by = Tree]
|
}, by = Tree]
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -102,7 +102,9 @@ xgb.plot.importance <- function(importance_matrix = NULL, top_n = NULL, measure
|
|||||||
original_mar <- par()$mar
|
original_mar <- par()$mar
|
||||||
|
|
||||||
# reset margins so this function doesn't have side effects
|
# reset margins so this function doesn't have side effects
|
||||||
on.exit({par(mar = original_mar)})
|
on.exit({
|
||||||
|
par(mar = original_mar)
|
||||||
|
})
|
||||||
|
|
||||||
mar <- original_mar
|
mar <- original_mar
|
||||||
if (!is.null(left_margin))
|
if (!is.null(left_margin))
|
||||||
|
|||||||
@@ -62,6 +62,9 @@
|
|||||||
#' @export
|
#' @export
|
||||||
xgb.plot.multi.trees <- function(model, feature_names = NULL, features_keep = 5, plot_width = NULL, plot_height = NULL,
|
xgb.plot.multi.trees <- function(model, feature_names = NULL, features_keep = 5, plot_width = NULL, plot_height = NULL,
|
||||||
render = TRUE, ...) {
|
render = TRUE, ...) {
|
||||||
|
if (!requireNamespace("DiagrammeR", quietly = TRUE)) {
|
||||||
|
stop("DiagrammeR is required for xgb.plot.multi.trees")
|
||||||
|
}
|
||||||
check.deprecation(...)
|
check.deprecation(...)
|
||||||
tree.matrix <- xgb.model.dt.tree(feature_names = feature_names, model = model)
|
tree.matrix <- xgb.model.dt.tree(feature_names = feature_names, model = model)
|
||||||
|
|
||||||
@@ -75,8 +78,8 @@ xgb.plot.multi.trees <- function(model, feature_names = NULL, features_keep = 5,
|
|||||||
while (tree.matrix[, sum(is.na(abs.node.position))] > 0) {
|
while (tree.matrix[, sum(is.na(abs.node.position))] > 0) {
|
||||||
yes.row.nodes <- tree.matrix[abs.node.position %in% precedent.nodes & !is.na(Yes)]
|
yes.row.nodes <- tree.matrix[abs.node.position %in% precedent.nodes & !is.na(Yes)]
|
||||||
no.row.nodes <- tree.matrix[abs.node.position %in% precedent.nodes & !is.na(No)]
|
no.row.nodes <- tree.matrix[abs.node.position %in% precedent.nodes & !is.na(No)]
|
||||||
yes.nodes.abs.pos <- yes.row.nodes[, abs.node.position] %>% paste0("_0")
|
yes.nodes.abs.pos <- paste0(yes.row.nodes[, abs.node.position], "_0")
|
||||||
no.nodes.abs.pos <- no.row.nodes[, abs.node.position] %>% paste0("_1")
|
no.nodes.abs.pos <- paste0(no.row.nodes[, abs.node.position], "_1")
|
||||||
|
|
||||||
tree.matrix[ID %in% yes.row.nodes[, Yes], abs.node.position := yes.nodes.abs.pos]
|
tree.matrix[ID %in% yes.row.nodes[, Yes], abs.node.position := yes.nodes.abs.pos]
|
||||||
tree.matrix[ID %in% no.row.nodes[, No], abs.node.position := no.nodes.abs.pos]
|
tree.matrix[ID %in% no.row.nodes[, No], abs.node.position := no.nodes.abs.pos]
|
||||||
@@ -92,19 +95,28 @@ xgb.plot.multi.trees <- function(model, feature_names = NULL, features_keep = 5,
|
|||||||
nodes.dt <- tree.matrix[
|
nodes.dt <- tree.matrix[
|
||||||
, .(Quality = sum(Quality))
|
, .(Quality = sum(Quality))
|
||||||
, by = .(abs.node.position, Feature)
|
, by = .(abs.node.position, Feature)
|
||||||
][, .(Text = paste0(Feature[1:min(length(Feature), features_keep)],
|
][, .(Text = paste0(
|
||||||
|
paste0(
|
||||||
|
Feature[seq_len(min(length(Feature), features_keep))],
|
||||||
" (",
|
" (",
|
||||||
format(Quality[1:min(length(Quality), features_keep)], digits = 5),
|
format(Quality[seq_len(min(length(Quality), features_keep))], digits = 5),
|
||||||
")") %>%
|
")"
|
||||||
paste0(collapse = "\n"))
|
),
|
||||||
, by = abs.node.position]
|
collapse = "\n"
|
||||||
|
)
|
||||||
|
)
|
||||||
|
, by = abs.node.position
|
||||||
|
]
|
||||||
|
|
||||||
edges.dt <- tree.matrix[Feature != "Leaf", .(abs.node.position, Yes)] %>%
|
edges.dt <- data.table::rbindlist(
|
||||||
list(tree.matrix[Feature != "Leaf", .(abs.node.position, No)]) %>%
|
l = list(
|
||||||
rbindlist() %>%
|
tree.matrix[Feature != "Leaf", .(abs.node.position, Yes)],
|
||||||
setnames(c("From", "To")) %>%
|
tree.matrix[Feature != "Leaf", .(abs.node.position, No)]
|
||||||
.[, .N, .(From, To)] %>%
|
)
|
||||||
.[, N := NULL]
|
)
|
||||||
|
data.table::setnames(edges.dt, c("From", "To"))
|
||||||
|
edges.dt <- edges.dt[, .N, .(From, To)]
|
||||||
|
edges.dt[, N := NULL]
|
||||||
|
|
||||||
nodes <- DiagrammeR::create_node_df(
|
nodes <- DiagrammeR::create_node_df(
|
||||||
n = nrow(nodes.dt),
|
n = nrow(nodes.dt),
|
||||||
@@ -120,21 +132,25 @@ xgb.plot.multi.trees <- function(model, feature_names = NULL, features_keep = 5,
|
|||||||
nodes_df = nodes,
|
nodes_df = nodes,
|
||||||
edges_df = edges,
|
edges_df = edges,
|
||||||
attr_theme = NULL
|
attr_theme = NULL
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "graph",
|
attr_type = "graph",
|
||||||
attr = c("layout", "rankdir"),
|
attr = c("layout", "rankdir"),
|
||||||
value = c("dot", "LR")
|
value = c("dot", "LR")
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "node",
|
attr_type = "node",
|
||||||
attr = c("color", "fillcolor", "style", "shape", "fontname"),
|
attr = c("color", "fillcolor", "style", "shape", "fontname"),
|
||||||
value = c("DimGray", "beige", "filled", "rectangle", "Helvetica")
|
value = c("DimGray", "beige", "filled", "rectangle", "Helvetica")
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "edge",
|
attr_type = "edge",
|
||||||
attr = c("color", "arrowsize", "arrowhead", "fontname"),
|
attr = c("color", "arrowsize", "arrowhead", "fontname"),
|
||||||
value = c("DimGray", "1.5", "vee", "Helvetica"))
|
value = c("DimGray", "1.5", "vee", "Helvetica")
|
||||||
|
)
|
||||||
|
|
||||||
if (!render) return(invisible(graph))
|
if (!render) return(invisible(graph))
|
||||||
|
|
||||||
|
|||||||
@@ -33,7 +33,7 @@
|
|||||||
#' @param col_loess a color to use for the loess curves.
|
#' @param col_loess a color to use for the loess curves.
|
||||||
#' @param span_loess the \code{span} parameter in \code{\link[stats]{loess}}'s call.
|
#' @param span_loess the \code{span} parameter in \code{\link[stats]{loess}}'s call.
|
||||||
#' @param which whether to do univariate or bivariate plotting. NOTE: only 1D is implemented so far.
|
#' @param which whether to do univariate or bivariate plotting. NOTE: only 1D is implemented so far.
|
||||||
#' @param plot whether a plot should be drawn. If FALSE, only a lits of matrices is returned.
|
#' @param plot whether a plot should be drawn. If FALSE, only a list of matrices is returned.
|
||||||
#' @param ... other parameters passed to \code{plot}.
|
#' @param ... other parameters passed to \code{plot}.
|
||||||
#'
|
#'
|
||||||
#' @details
|
#' @details
|
||||||
@@ -143,7 +143,7 @@ xgb.plot.shap <- function(data, shap_contrib = NULL, features = NULL, top_n = 1,
|
|||||||
y <- shap_contrib[, f][ord]
|
y <- shap_contrib[, f][ord]
|
||||||
x_lim <- range(x, na.rm = TRUE)
|
x_lim <- range(x, na.rm = TRUE)
|
||||||
y_lim <- range(y, na.rm = TRUE)
|
y_lim <- range(y, na.rm = TRUE)
|
||||||
do_na <- plot_NA && any(is.na(x))
|
do_na <- plot_NA && anyNA(x)
|
||||||
if (do_na) {
|
if (do_na) {
|
||||||
x_range <- diff(x_lim)
|
x_range <- diff(x_lim)
|
||||||
loc_na <- min(x, na.rm = TRUE) + x_range * pos_NA
|
loc_na <- min(x, na.rm = TRUE) + x_range * pos_NA
|
||||||
@@ -157,7 +157,7 @@ xgb.plot.shap <- function(data, shap_contrib = NULL, features = NULL, top_n = 1,
|
|||||||
plot(x2plot, y, pch = pch, xlab = f, col = col, xlim = x_lim, ylim = y_lim, ylab = ylab, ...)
|
plot(x2plot, y, pch = pch, xlab = f, col = col, xlim = x_lim, ylim = y_lim, ylab = ylab, ...)
|
||||||
grid()
|
grid()
|
||||||
if (plot_loess) {
|
if (plot_loess) {
|
||||||
# compress x to 3 digits, and mean-aggredate y
|
# compress x to 3 digits, and mean-aggregate y
|
||||||
zz <- data.table(x = signif(x, 3), y)[, .(.N, y = mean(y)), x]
|
zz <- data.table(x = signif(x, 3), y)[, .(.N, y = mean(y)), x]
|
||||||
if (nrow(zz) <= 5) {
|
if (nrow(zz) <= 5) {
|
||||||
lines(zz$x, zz$y, col = col_loess)
|
lines(zz$x, zz$y, col = col_loess)
|
||||||
@@ -193,7 +193,7 @@ xgb.plot.shap <- function(data, shap_contrib = NULL, features = NULL, top_n = 1,
|
|||||||
#' hence allows us to see which features have a negative / positive contribution
|
#' hence allows us to see which features have a negative / positive contribution
|
||||||
#' on the model prediction, and whether the contribution is different for larger
|
#' on the model prediction, and whether the contribution is different for larger
|
||||||
#' or smaller values of the feature. We effectively try to replicate the
|
#' or smaller values of the feature. We effectively try to replicate the
|
||||||
#' \code{summary_plot} function from https://github.com/slundberg/shap.
|
#' \code{summary_plot} function from https://github.com/shap/shap.
|
||||||
#'
|
#'
|
||||||
#' @inheritParams xgb.plot.shap
|
#' @inheritParams xgb.plot.shap
|
||||||
#'
|
#'
|
||||||
@@ -202,7 +202,7 @@ xgb.plot.shap <- function(data, shap_contrib = NULL, features = NULL, top_n = 1,
|
|||||||
#'
|
#'
|
||||||
#' @examples # See \code{\link{xgb.plot.shap}}.
|
#' @examples # See \code{\link{xgb.plot.shap}}.
|
||||||
#' @seealso \code{\link{xgb.plot.shap}}, \code{\link{xgb.ggplot.shap.summary}},
|
#' @seealso \code{\link{xgb.plot.shap}}, \code{\link{xgb.ggplot.shap.summary}},
|
||||||
#' \url{https://github.com/slundberg/shap}
|
#' \url{https://github.com/shap/shap}
|
||||||
xgb.plot.shap.summary <- function(data, shap_contrib = NULL, features = NULL, top_n = 10, model = NULL,
|
xgb.plot.shap.summary <- function(data, shap_contrib = NULL, features = NULL, top_n = 10, model = NULL,
|
||||||
trees = NULL, target_class = NULL, approxcontrib = FALSE, subsample = NULL) {
|
trees = NULL, target_class = NULL, approxcontrib = FALSE, subsample = NULL) {
|
||||||
# Only ggplot implementation is available.
|
# Only ggplot implementation is available.
|
||||||
@@ -272,8 +272,8 @@ xgb.shap.data <- function(data, shap_contrib = NULL, features = NULL, top_n = 1,
|
|||||||
imp <- xgb.importance(model = model, trees = trees, feature_names = colnames(data))
|
imp <- xgb.importance(model = model, trees = trees, feature_names = colnames(data))
|
||||||
}
|
}
|
||||||
top_n <- top_n[1]
|
top_n <- top_n[1]
|
||||||
if (top_n < 1 | top_n > 100) stop("top_n: must be an integer within [1, 100]")
|
if (top_n < 1 || top_n > 100) stop("top_n: must be an integer within [1, 100]")
|
||||||
features <- imp$Feature[1:min(top_n, NROW(imp))]
|
features <- imp$Feature[seq_len(min(top_n, NROW(imp)))]
|
||||||
}
|
}
|
||||||
if (is.character(features)) {
|
if (is.character(features)) {
|
||||||
features <- match(features, colnames(data))
|
features <- match(features, colnames(data))
|
||||||
|
|||||||
@@ -34,7 +34,7 @@
|
|||||||
#' The branches that also used for missing values are marked as bold
|
#' The branches that also used for missing values are marked as bold
|
||||||
#' (as in "carrying extra capacity").
|
#' (as in "carrying extra capacity").
|
||||||
#'
|
#'
|
||||||
#' This function uses \href{http://www.graphviz.org/}{GraphViz} as a backend of DiagrammeR.
|
#' This function uses \href{https://www.graphviz.org/}{GraphViz} as a backend of DiagrammeR.
|
||||||
#'
|
#'
|
||||||
#' @return
|
#' @return
|
||||||
#'
|
#'
|
||||||
@@ -98,34 +98,46 @@ xgb.plot.tree <- function(feature_names = NULL, model = NULL, trees = NULL, plot
|
|||||||
data = dt$Feature,
|
data = dt$Feature,
|
||||||
fontcolor = "black")
|
fontcolor = "black")
|
||||||
|
|
||||||
|
if (nrow(dt[Feature != "Leaf"]) != 0) {
|
||||||
edges <- DiagrammeR::create_edge_df(
|
edges <- DiagrammeR::create_edge_df(
|
||||||
from = match(dt[Feature != "Leaf", c(ID)] %>% rep(2), dt$ID),
|
from = match(rep(dt[Feature != "Leaf", c(ID)], 2), dt$ID),
|
||||||
to = match(dt[Feature != "Leaf", c(Yes, No)], dt$ID),
|
to = match(dt[Feature != "Leaf", c(Yes, No)], dt$ID),
|
||||||
label = dt[Feature != "Leaf", paste("<", Split)] %>%
|
label = c(
|
||||||
c(rep("", nrow(dt[Feature != "Leaf"]))),
|
dt[Feature != "Leaf", paste("<", Split)],
|
||||||
style = dt[Feature != "Leaf", ifelse(Missing == Yes, "bold", "solid")] %>%
|
rep("", nrow(dt[Feature != "Leaf"]))
|
||||||
c(dt[Feature != "Leaf", ifelse(Missing == No, "bold", "solid")]),
|
),
|
||||||
|
style = c(
|
||||||
|
dt[Feature != "Leaf", ifelse(Missing == Yes, "bold", "solid")],
|
||||||
|
dt[Feature != "Leaf", ifelse(Missing == No, "bold", "solid")]
|
||||||
|
),
|
||||||
rel = "leading_to")
|
rel = "leading_to")
|
||||||
|
} else {
|
||||||
|
edges <- NULL
|
||||||
|
}
|
||||||
|
|
||||||
graph <- DiagrammeR::create_graph(
|
graph <- DiagrammeR::create_graph(
|
||||||
nodes_df = nodes,
|
nodes_df = nodes,
|
||||||
edges_df = edges,
|
edges_df = edges,
|
||||||
attr_theme = NULL
|
attr_theme = NULL
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "graph",
|
attr_type = "graph",
|
||||||
attr = c("layout", "rankdir"),
|
attr = c("layout", "rankdir"),
|
||||||
value = c("dot", "LR")
|
value = c("dot", "LR")
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "node",
|
attr_type = "node",
|
||||||
attr = c("color", "style", "fontname"),
|
attr = c("color", "style", "fontname"),
|
||||||
value = c("DimGray", "filled", "Helvetica")
|
value = c("DimGray", "filled", "Helvetica")
|
||||||
) %>%
|
)
|
||||||
DiagrammeR::add_global_graph_attrs(
|
graph <- DiagrammeR::add_global_graph_attrs(
|
||||||
|
graph = graph,
|
||||||
attr_type = "edge",
|
attr_type = "edge",
|
||||||
attr = c("color", "arrowsize", "arrowhead", "fontname"),
|
attr = c("color", "arrowsize", "arrowhead", "fontname"),
|
||||||
value = c("DimGray", "1.5", "vee", "Helvetica"))
|
value = c("DimGray", "1.5", "vee", "Helvetica")
|
||||||
|
)
|
||||||
|
|
||||||
if (!render) return(invisible(graph))
|
if (!render) return(invisible(graph))
|
||||||
|
|
||||||
|
|||||||
@@ -43,6 +43,6 @@ xgb.save <- function(model, fname) {
|
|||||||
}
|
}
|
||||||
model <- xgb.Booster.complete(model, saveraw = FALSE)
|
model <- xgb.Booster.complete(model, saveraw = FALSE)
|
||||||
fname <- path.expand(fname)
|
fname <- path.expand(fname)
|
||||||
.Call(XGBoosterSaveModel_R, model$handle, fname[1])
|
.Call(XGBoosterSaveModel_R, model$handle, enc2utf8(fname[1]))
|
||||||
return(TRUE)
|
return(TRUE)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,6 +4,14 @@
|
|||||||
#' Save xgboost model from xgboost or xgb.train
|
#' Save xgboost model from xgboost or xgb.train
|
||||||
#'
|
#'
|
||||||
#' @param model the model object.
|
#' @param model the model object.
|
||||||
|
#' @param raw_format The format for encoding the booster. Available options are
|
||||||
|
#' \itemize{
|
||||||
|
#' \item \code{json}: Encode the booster into JSON text document.
|
||||||
|
#' \item \code{ubj}: Encode the booster into Universal Binary JSON.
|
||||||
|
#' \item \code{deprecated}: Encode the booster into old customized binary format.
|
||||||
|
#' }
|
||||||
|
#'
|
||||||
|
#' Right now the default is \code{deprecated} but will be changed to \code{ubj} in upcoming release.
|
||||||
#'
|
#'
|
||||||
#' @examples
|
#' @examples
|
||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
@@ -17,7 +25,8 @@
|
|||||||
#' pred <- predict(bst, test$data)
|
#' pred <- predict(bst, test$data)
|
||||||
#'
|
#'
|
||||||
#' @export
|
#' @export
|
||||||
xgb.save.raw <- function(model) {
|
xgb.save.raw <- function(model, raw_format = "deprecated") {
|
||||||
handle <- xgb.get.handle(model)
|
handle <- xgb.get.handle(model)
|
||||||
.Call(XGBoosterModelToRaw_R, handle)
|
args <- list(format = raw_format)
|
||||||
|
.Call(XGBoosterSaveModelToRaw_R, handle, jsonlite::toJSON(args, auto_unbox = TRUE))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -18,17 +18,37 @@
|
|||||||
#' 2.1. Parameters for Tree Booster
|
#' 2.1. Parameters for Tree Booster
|
||||||
#'
|
#'
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{eta} control the learning rate: scale the contribution of each tree by a factor of \code{0 < eta < 1} when it is added to the current approximation. Used to prevent overfitting by making the boosting process more conservative. Lower value for \code{eta} implies larger value for \code{nrounds}: low \code{eta} value means model more robust to overfitting but slower to compute. Default: 0.3
|
#' \item{ \code{eta} control the learning rate: scale the contribution of each tree by a factor of \code{0 < eta < 1}
|
||||||
#' \item \code{gamma} minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.
|
#' when it is added to the current approximation.
|
||||||
|
#' Used to prevent overfitting by making the boosting process more conservative.
|
||||||
|
#' Lower value for \code{eta} implies larger value for \code{nrounds}: low \code{eta} value means model
|
||||||
|
#' more robust to overfitting but slower to compute. Default: 0.3}
|
||||||
|
#' \item{ \code{gamma} minimum loss reduction required to make a further partition on a leaf node of the tree.
|
||||||
|
#' the larger, the more conservative the algorithm will be.}
|
||||||
#' \item \code{max_depth} maximum depth of a tree. Default: 6
|
#' \item \code{max_depth} maximum depth of a tree. Default: 6
|
||||||
#' \item \code{min_child_weight} minimum sum of instance weight (hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be. Default: 1
|
#' \item{\code{min_child_weight} minimum sum of instance weight (hessian) needed in a child.
|
||||||
#' \item \code{subsample} subsample ratio of the training instance. Setting it to 0.5 means that xgboost randomly collected half of the data instances to grow trees and this will prevent overfitting. It makes computation shorter (because less data to analyse). It is advised to use this parameter with \code{eta} and increase \code{nrounds}. Default: 1
|
#' If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight,
|
||||||
|
#' then the building process will give up further partitioning.
|
||||||
|
#' In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node.
|
||||||
|
#' The larger, the more conservative the algorithm will be. Default: 1}
|
||||||
|
#' \item{ \code{subsample} subsample ratio of the training instance.
|
||||||
|
#' Setting it to 0.5 means that xgboost randomly collected half of the data instances to grow trees
|
||||||
|
#' and this will prevent overfitting. It makes computation shorter (because less data to analyse).
|
||||||
|
#' It is advised to use this parameter with \code{eta} and increase \code{nrounds}. Default: 1}
|
||||||
#' \item \code{colsample_bytree} subsample ratio of columns when constructing each tree. Default: 1
|
#' \item \code{colsample_bytree} subsample ratio of columns when constructing each tree. Default: 1
|
||||||
#' \item \code{lambda} L2 regularization term on weights. Default: 1
|
#' \item \code{lambda} L2 regularization term on weights. Default: 1
|
||||||
#' \item \code{alpha} L1 regularization term on weights. (there is no L1 reg on bias because it is not important). Default: 0
|
#' \item \code{alpha} L1 regularization term on weights. (there is no L1 reg on bias because it is not important). Default: 0
|
||||||
#' \item \code{num_parallel_tree} Experimental parameter. number of trees to grow per round. Useful to test Random Forest through Xgboost (set \code{colsample_bytree < 1}, \code{subsample < 1} and \code{round = 1}) accordingly. Default: 1
|
#' \item{ \code{num_parallel_tree} Experimental parameter. number of trees to grow per round.
|
||||||
#' \item \code{monotone_constraints} A numerical vector consists of \code{1}, \code{0} and \code{-1} with its length equals to the number of features in the training data. \code{1} is increasing, \code{-1} is decreasing and \code{0} is no constraint.
|
#' Useful to test Random Forest through XGBoost
|
||||||
#' \item \code{interaction_constraints} A list of vectors specifying feature indices of permitted interactions. Each item of the list represents one permitted interaction where specified features are allowed to interact with each other. Feature index values should start from \code{0} (\code{0} references the first column). Leave argument unspecified for no interaction constraints.
|
#' (set \code{colsample_bytree < 1}, \code{subsample < 1} and \code{round = 1}) accordingly.
|
||||||
|
#' Default: 1}
|
||||||
|
#' \item{ \code{monotone_constraints} A numerical vector consists of \code{1}, \code{0} and \code{-1} with its length
|
||||||
|
#' equals to the number of features in the training data.
|
||||||
|
#' \code{1} is increasing, \code{-1} is decreasing and \code{0} is no constraint.}
|
||||||
|
#' \item{ \code{interaction_constraints} A list of vectors specifying feature indices of permitted interactions.
|
||||||
|
#' Each item of the list represents one permitted interaction where specified features are allowed to interact with each other.
|
||||||
|
#' Feature index values should start from \code{0} (\code{0} references the first column).
|
||||||
|
#' Leave argument unspecified for no interaction constraints.}
|
||||||
#' }
|
#' }
|
||||||
#'
|
#'
|
||||||
#' 2.2. Parameters for Linear Booster
|
#' 2.2. Parameters for Linear Booster
|
||||||
@@ -42,29 +62,53 @@
|
|||||||
#' 3. Task Parameters
|
#' 3. Task Parameters
|
||||||
#'
|
#'
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{objective} specify the learning task and the corresponding learning objective, users can pass a self-defined function to it. The default objective options are below:
|
#' \item{ \code{objective} specify the learning task and the corresponding learning objective, users can pass a self-defined function to it.
|
||||||
|
#' The default objective options are below:
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{reg:squarederror} Regression with squared loss (Default).
|
#' \item \code{reg:squarederror} Regression with squared loss (Default).
|
||||||
#' \item \code{reg:squaredlogerror}: regression with squared log loss \eqn{1/2 * (log(pred + 1) - log(label + 1))^2}. All inputs are required to be greater than -1. Also, see metric rmsle for possible issue with this objective.
|
#' \item{ \code{reg:squaredlogerror}: regression with squared log loss \eqn{1/2 * (log(pred + 1) - log(label + 1))^2}.
|
||||||
|
#' All inputs are required to be greater than -1.
|
||||||
|
#' Also, see metric rmsle for possible issue with this objective.}
|
||||||
#' \item \code{reg:logistic} logistic regression.
|
#' \item \code{reg:logistic} logistic regression.
|
||||||
#' \item \code{reg:pseudohubererror}: regression with Pseudo Huber loss, a twice differentiable alternative to absolute loss.
|
#' \item \code{reg:pseudohubererror}: regression with Pseudo Huber loss, a twice differentiable alternative to absolute loss.
|
||||||
#' \item \code{binary:logistic} logistic regression for binary classification. Output probability.
|
#' \item \code{binary:logistic} logistic regression for binary classification. Output probability.
|
||||||
#' \item \code{binary:logitraw} logistic regression for binary classification, output score before logistic transformation.
|
#' \item \code{binary:logitraw} logistic regression for binary classification, output score before logistic transformation.
|
||||||
#' \item \code{binary:hinge}: hinge loss for binary classification. This makes predictions of 0 or 1, rather than producing probabilities.
|
#' \item \code{binary:hinge}: hinge loss for binary classification. This makes predictions of 0 or 1, rather than producing probabilities.
|
||||||
#' \item \code{count:poisson}: poisson regression for count data, output mean of poisson distribution. \code{max_delta_step} is set to 0.7 by default in poisson regression (used to safeguard optimization).
|
#' \item{ \code{count:poisson}: Poisson regression for count data, output mean of Poisson distribution.
|
||||||
#' \item \code{survival:cox}: Cox regression for right censored survival time data (negative values are considered right censored). Note that predictions are returned on the hazard ratio scale (i.e., as HR = exp(marginal_prediction) in the proportional hazard function \code{h(t) = h0(t) * HR)}.
|
#' \code{max_delta_step} is set to 0.7 by default in poisson regression (used to safeguard optimization).}
|
||||||
#' \item \code{survival:aft}: Accelerated failure time model for censored survival time data. See \href{https://xgboost.readthedocs.io/en/latest/tutorials/aft_survival_analysis.html}{Survival Analysis with Accelerated Failure Time} for details.
|
#' \item{ \code{survival:cox}: Cox regression for right censored survival time data (negative values are considered right censored).
|
||||||
#' \item \code{aft_loss_distribution}: Probabilty Density Function used by \code{survival:aft} and \code{aft-nloglik} metric.
|
#' Note that predictions are returned on the hazard ratio scale (i.e., as HR = exp(marginal_prediction) in the proportional
|
||||||
#' \item \code{multi:softmax} set xgboost to do multiclass classification using the softmax objective. Class is represented by a number and should be from 0 to \code{num_class - 1}.
|
#' hazard function \code{h(t) = h0(t) * HR)}.}
|
||||||
#' \item \code{multi:softprob} same as softmax, but prediction outputs a vector of ndata * nclass elements, which can be further reshaped to ndata, nclass matrix. The result contains predicted probabilities of each data point belonging to each class.
|
#' \item{ \code{survival:aft}: Accelerated failure time model for censored survival time data. See
|
||||||
|
#' \href{https://xgboost.readthedocs.io/en/latest/tutorials/aft_survival_analysis.html}{Survival Analysis with Accelerated Failure Time}
|
||||||
|
#' for details.}
|
||||||
|
#' \item \code{aft_loss_distribution}: Probability Density Function used by \code{survival:aft} and \code{aft-nloglik} metric.
|
||||||
|
#' \item{ \code{multi:softmax} set xgboost to do multiclass classification using the softmax objective.
|
||||||
|
#' Class is represented by a number and should be from 0 to \code{num_class - 1}.}
|
||||||
|
#' \item{ \code{multi:softprob} same as softmax, but prediction outputs a vector of ndata * nclass elements, which can be
|
||||||
|
#' further reshaped to ndata, nclass matrix. The result contains predicted probabilities of each data point belonging
|
||||||
|
#' to each class.}
|
||||||
#' \item \code{rank:pairwise} set xgboost to do ranking task by minimizing the pairwise loss.
|
#' \item \code{rank:pairwise} set xgboost to do ranking task by minimizing the pairwise loss.
|
||||||
#' \item \code{rank:ndcg}: Use LambdaMART to perform list-wise ranking where \href{https://en.wikipedia.org/wiki/Discounted_cumulative_gain}{Normalized Discounted Cumulative Gain (NDCG)} is maximized.
|
#' \item{ \code{rank:ndcg}: Use LambdaMART to perform list-wise ranking where
|
||||||
#' \item \code{rank:map}: Use LambdaMART to perform list-wise ranking where \href{https://en.wikipedia.org/wiki/Evaluation_measures_(information_retrieval)#Mean_average_precision}{Mean Average Precision (MAP)} is maximized.
|
#' \href{https://en.wikipedia.org/wiki/Discounted_cumulative_gain}{Normalized Discounted Cumulative Gain (NDCG)} is maximized.}
|
||||||
#' \item \code{reg:gamma}: gamma regression with log-link. Output is a mean of gamma distribution. It might be useful, e.g., for modeling insurance claims severity, or for any outcome that might be \href{https://en.wikipedia.org/wiki/Gamma_distribution#Applications}{gamma-distributed}.
|
#' \item{ \code{rank:map}: Use LambdaMART to perform list-wise ranking where
|
||||||
#' \item \code{reg:tweedie}: Tweedie regression with log-link. It might be useful, e.g., for modeling total loss in insurance, or for any outcome that might be \href{https://en.wikipedia.org/wiki/Tweedie_distribution#Applications}{Tweedie-distributed}.
|
#' \href{https://en.wikipedia.org/wiki/Evaluation_measures_(information_retrieval)#Mean_average_precision}{Mean Average Precision (MAP)}
|
||||||
|
#' is maximized.}
|
||||||
|
#' \item{ \code{reg:gamma}: gamma regression with log-link.
|
||||||
|
#' Output is a mean of gamma distribution.
|
||||||
|
#' It might be useful, e.g., for modeling insurance claims severity, or for any outcome that might be
|
||||||
|
#' \href{https://en.wikipedia.org/wiki/Gamma_distribution#Applications}{gamma-distributed}.}
|
||||||
|
#' \item{ \code{reg:tweedie}: Tweedie regression with log-link.
|
||||||
|
#' It might be useful, e.g., for modeling total loss in insurance, or for any outcome that might be
|
||||||
|
#' \href{https://en.wikipedia.org/wiki/Tweedie_distribution#Applications}{Tweedie-distributed}.}
|
||||||
|
#' }
|
||||||
#' }
|
#' }
|
||||||
#' \item \code{base_score} the initial prediction score of all instances, global bias. Default: 0.5
|
#' \item \code{base_score} the initial prediction score of all instances, global bias. Default: 0.5
|
||||||
#' \item \code{eval_metric} evaluation metrics for validation data. Users can pass a self-defined function to it. Default: metric will be assigned according to objective(rmse for regression, and error for classification, mean average precision for ranking). List is provided in detail section.
|
#' \item{ \code{eval_metric} evaluation metrics for validation data.
|
||||||
|
#' Users can pass a self-defined function to it.
|
||||||
|
#' Default: metric will be assigned according to objective
|
||||||
|
#' (rmse for regression, and error for classification, mean average precision for ranking).
|
||||||
|
#' List is provided in detail section.}
|
||||||
#' }
|
#' }
|
||||||
#'
|
#'
|
||||||
#' @param data training dataset. \code{xgb.train} accepts only an \code{xgb.DMatrix} as the input.
|
#' @param data training dataset. \code{xgb.train} accepts only an \code{xgb.DMatrix} as the input.
|
||||||
@@ -126,11 +170,11 @@
|
|||||||
#' Parallelization is automatically enabled if \code{OpenMP} is present.
|
#' Parallelization is automatically enabled if \code{OpenMP} is present.
|
||||||
#' Number of threads can also be manually specified via \code{nthread} parameter.
|
#' Number of threads can also be manually specified via \code{nthread} parameter.
|
||||||
#'
|
#'
|
||||||
#' The evaluation metric is chosen automatically by Xgboost (according to the objective)
|
#' The evaluation metric is chosen automatically by XGBoost (according to the objective)
|
||||||
#' when the \code{eval_metric} parameter is not provided.
|
#' when the \code{eval_metric} parameter is not provided.
|
||||||
#' User may set one or several \code{eval_metric} parameters.
|
#' User may set one or several \code{eval_metric} parameters.
|
||||||
#' Note that when using a customized metric, only this single metric can be used.
|
#' Note that when using a customized metric, only this single metric can be used.
|
||||||
#' The following is the list of built-in metrics for which Xgboost provides optimized implementation:
|
#' The following is the list of built-in metrics for which XGBoost provides optimized implementation:
|
||||||
#' \itemize{
|
#' \itemize{
|
||||||
#' \item \code{rmse} root mean square error. \url{https://en.wikipedia.org/wiki/Root_mean_square_error}
|
#' \item \code{rmse} root mean square error. \url{https://en.wikipedia.org/wiki/Root_mean_square_error}
|
||||||
#' \item \code{logloss} negative log-likelihood. \url{https://en.wikipedia.org/wiki/Log-likelihood}
|
#' \item \code{logloss} negative log-likelihood. \url{https://en.wikipedia.org/wiki/Log-likelihood}
|
||||||
@@ -141,7 +185,8 @@
|
|||||||
#' \item \code{merror} Multiclass classification error rate. It is calculated as \code{(# wrong cases) / (# all cases)}.
|
#' \item \code{merror} Multiclass classification error rate. It is calculated as \code{(# wrong cases) / (# all cases)}.
|
||||||
#' \item \code{mae} Mean absolute error
|
#' \item \code{mae} Mean absolute error
|
||||||
#' \item \code{mape} Mean absolute percentage error
|
#' \item \code{mape} Mean absolute percentage error
|
||||||
#' \item \code{auc} Area under the curve. \url{https://en.wikipedia.org/wiki/Receiver_operating_characteristic#'Area_under_curve} for ranking evaluation.
|
#' \item{ \code{auc} Area under the curve.
|
||||||
|
#' \url{https://en.wikipedia.org/wiki/Receiver_operating_characteristic#'Area_under_curve} for ranking evaluation.}
|
||||||
#' \item \code{aucpr} Area under the PR curve. \url{https://en.wikipedia.org/wiki/Precision_and_recall} for ranking evaluation.
|
#' \item \code{aucpr} Area under the PR curve. \url{https://en.wikipedia.org/wiki/Precision_and_recall} for ranking evaluation.
|
||||||
#' \item \code{ndcg} Normalized Discounted Cumulative Gain (for ranking task). \url{https://en.wikipedia.org/wiki/NDCG}
|
#' \item \code{ndcg} Normalized Discounted Cumulative Gain (for ranking task). \url{https://en.wikipedia.org/wiki/NDCG}
|
||||||
#' }
|
#' }
|
||||||
@@ -171,9 +216,6 @@
|
|||||||
#' explicitly passed.
|
#' explicitly passed.
|
||||||
#' \item \code{best_iteration} iteration number with the best evaluation metric value
|
#' \item \code{best_iteration} iteration number with the best evaluation metric value
|
||||||
#' (only available with early stopping).
|
#' (only available with early stopping).
|
||||||
#' \item \code{best_ntreelimit} the \code{ntreelimit} value corresponding to the best iteration,
|
|
||||||
#' which could further be used in \code{predict} method
|
|
||||||
#' (only available with early stopping).
|
|
||||||
#' \item \code{best_score} the best evaluation metric value during early stopping.
|
#' \item \code{best_score} the best evaluation metric value during early stopping.
|
||||||
#' (only available with early stopping).
|
#' (only available with early stopping).
|
||||||
#' \item \code{feature_names} names of the training dataset features
|
#' \item \code{feature_names} names of the training dataset features
|
||||||
@@ -195,8 +237,8 @@
|
|||||||
#' data(agaricus.train, package='xgboost')
|
#' data(agaricus.train, package='xgboost')
|
||||||
#' data(agaricus.test, package='xgboost')
|
#' data(agaricus.test, package='xgboost')
|
||||||
#'
|
#'
|
||||||
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label))
|
#' dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' dtest <- with(agaricus.test, xgb.DMatrix(data, label = label))
|
#' dtest <- with(agaricus.test, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
#' watchlist <- list(train = dtrain, eval = dtest)
|
#' watchlist <- list(train = dtrain, eval = dtest)
|
||||||
#'
|
#'
|
||||||
#' ## A simple xgb.train example:
|
#' ## A simple xgb.train example:
|
||||||
@@ -279,6 +321,10 @@ xgb.train <- function(params = list(), data, nrounds, watchlist = list(),
|
|||||||
if (is.null(evnames) || any(evnames == ""))
|
if (is.null(evnames) || any(evnames == ""))
|
||||||
stop("each element of the watchlist must have a name tag")
|
stop("each element of the watchlist must have a name tag")
|
||||||
}
|
}
|
||||||
|
# Handle multiple evaluation metrics given as a list
|
||||||
|
for (m in params$eval_metric) {
|
||||||
|
params <- c(params, list(eval_metric = m))
|
||||||
|
}
|
||||||
|
|
||||||
# evaluation printing callback
|
# evaluation printing callback
|
||||||
params <- c(params)
|
params <- c(params)
|
||||||
@@ -317,8 +363,13 @@ xgb.train <- function(params = list(), data, nrounds, watchlist = list(),
|
|||||||
is_update <- NVL(params[['process_type']], '.') == 'update'
|
is_update <- NVL(params[['process_type']], '.') == 'update'
|
||||||
|
|
||||||
# Construct a booster (either a new one or load from xgb_model)
|
# Construct a booster (either a new one or load from xgb_model)
|
||||||
handle <- xgb.Booster.handle(params, append(watchlist, dtrain), xgb_model)
|
handle <- xgb.Booster.handle(
|
||||||
bst <- xgb.handleToBooster(handle)
|
params = params,
|
||||||
|
cachelist = append(watchlist, dtrain),
|
||||||
|
modelfile = xgb_model,
|
||||||
|
handle = NULL
|
||||||
|
)
|
||||||
|
bst <- xgb.handleToBooster(handle = handle, raw = NULL)
|
||||||
|
|
||||||
# extract parameters that can affect the relationship b/w #trees and #iterations
|
# extract parameters that can affect the relationship b/w #trees and #iterations
|
||||||
num_class <- max(as.numeric(NVL(params[['num_class']], 1)), 1)
|
num_class <- max(as.numeric(NVL(params[['num_class']], 1)), 1)
|
||||||
@@ -344,10 +395,21 @@ xgb.train <- function(params = list(), data, nrounds, watchlist = list(),
|
|||||||
|
|
||||||
for (f in cb$pre_iter) f()
|
for (f in cb$pre_iter) f()
|
||||||
|
|
||||||
xgb.iter.update(bst$handle, dtrain, iteration - 1, obj)
|
xgb.iter.update(
|
||||||
|
booster_handle = bst$handle,
|
||||||
|
dtrain = dtrain,
|
||||||
|
iter = iteration - 1,
|
||||||
|
obj = obj
|
||||||
|
)
|
||||||
|
|
||||||
if (length(watchlist) > 0)
|
if (length(watchlist) > 0) {
|
||||||
bst_evaluation <- xgb.iter.eval(bst$handle, watchlist, iteration - 1, feval)
|
bst_evaluation <- xgb.iter.eval( # nolint: object_usage_linter
|
||||||
|
booster_handle = bst$handle,
|
||||||
|
watchlist = watchlist,
|
||||||
|
iter = iteration - 1,
|
||||||
|
feval = feval
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
xgb.attr(bst$handle, 'niter') <- iteration - 1
|
xgb.attr(bst$handle, 'niter') <- iteration - 1
|
||||||
|
|
||||||
|
|||||||
@@ -1,11 +1,21 @@
|
|||||||
#' Load the instance back from \code{\link{xgb.serialize}}
|
#' Load the instance back from \code{\link{xgb.serialize}}
|
||||||
#'
|
#'
|
||||||
#' @param buffer the buffer containing booster instance saved by \code{\link{xgb.serialize}}
|
#' @param buffer the buffer containing booster instance saved by \code{\link{xgb.serialize}}
|
||||||
|
#' @param handle An \code{xgb.Booster.handle} object which will be overwritten with
|
||||||
|
#' the new deserialized object. Must be a null handle (e.g. when loading the model through
|
||||||
|
#' `readRDS`). If not provided, a new handle will be created.
|
||||||
|
#' @return An \code{xgb.Booster.handle} object.
|
||||||
#'
|
#'
|
||||||
#' @export
|
#' @export
|
||||||
xgb.unserialize <- function(buffer) {
|
xgb.unserialize <- function(buffer, handle = NULL) {
|
||||||
cachelist <- list()
|
cachelist <- list()
|
||||||
|
if (is.null(handle)) {
|
||||||
handle <- .Call(XGBoosterCreate_R, cachelist)
|
handle <- .Call(XGBoosterCreate_R, cachelist)
|
||||||
|
} else {
|
||||||
|
if (!is.null.handle(handle))
|
||||||
|
stop("'handle' is not null/empty. Cannot overwrite existing handle.")
|
||||||
|
.Call(XGBoosterCreateInEmptyObj_R, cachelist, handle)
|
||||||
|
}
|
||||||
tryCatch(
|
tryCatch(
|
||||||
.Call(XGBoosterUnserializeFromBuffer_R, handle, buffer),
|
.Call(XGBoosterUnserializeFromBuffer_R, handle, buffer),
|
||||||
error = function(e) {
|
error = function(e) {
|
||||||
|
|||||||
@@ -9,8 +9,14 @@ xgboost <- function(data = NULL, label = NULL, missing = NA, weight = NULL,
|
|||||||
early_stopping_rounds = NULL, maximize = NULL,
|
early_stopping_rounds = NULL, maximize = NULL,
|
||||||
save_period = NULL, save_name = "xgboost.model",
|
save_period = NULL, save_name = "xgboost.model",
|
||||||
xgb_model = NULL, callbacks = list(), ...) {
|
xgb_model = NULL, callbacks = list(), ...) {
|
||||||
|
merged <- check.booster.params(params, ...)
|
||||||
dtrain <- xgb.get.DMatrix(data, label, missing, weight)
|
dtrain <- xgb.get.DMatrix(
|
||||||
|
data = data,
|
||||||
|
label = label,
|
||||||
|
missing = missing,
|
||||||
|
weight = weight,
|
||||||
|
nthread = merged$nthread
|
||||||
|
)
|
||||||
|
|
||||||
watchlist <- list(train = dtrain)
|
watchlist <- list(train = dtrain)
|
||||||
|
|
||||||
@@ -90,7 +96,6 @@ NULL
|
|||||||
#' @importFrom data.table setkey
|
#' @importFrom data.table setkey
|
||||||
#' @importFrom data.table setkeyv
|
#' @importFrom data.table setkeyv
|
||||||
#' @importFrom data.table setnames
|
#' @importFrom data.table setnames
|
||||||
#' @importFrom magrittr %>%
|
|
||||||
#' @importFrom jsonlite fromJSON
|
#' @importFrom jsonlite fromJSON
|
||||||
#' @importFrom jsonlite toJSON
|
#' @importFrom jsonlite toJSON
|
||||||
#' @importFrom utils object.size str tail
|
#' @importFrom utils object.size str tail
|
||||||
|
|||||||
@@ -30,4 +30,4 @@ Examples
|
|||||||
Development
|
Development
|
||||||
-----------
|
-----------
|
||||||
|
|
||||||
* See the [R Package section](https://xgboost.readthedocs.io/en/latest/contribute.html#r-package) of the contributors guide.
|
* See the [R Package section](https://xgboost.readthedocs.io/en/latest/contrib/coding_guide.html#r-coding-guideline) of the contributors guide.
|
||||||
|
|||||||
@@ -1,4 +1,3 @@
|
|||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
|
|
||||||
rm -f src/Makevars
|
rm -f src/Makevars
|
||||||
rm -f CMakeLists.txt
|
|
||||||
|
|||||||
1851
R-package/configure
vendored
1851
R-package/configure
vendored
File diff suppressed because it is too large
Load Diff
@@ -2,10 +2,25 @@
|
|||||||
|
|
||||||
AC_PREREQ(2.69)
|
AC_PREREQ(2.69)
|
||||||
|
|
||||||
AC_INIT([xgboost],[0.6-3],[],[xgboost],[])
|
AC_INIT([xgboost],[2.0.0],[],[xgboost],[])
|
||||||
|
|
||||||
# Use this line to set CC variable to a C compiler
|
: ${R_HOME=`R RHOME`}
|
||||||
AC_PROG_CC
|
if test -z "${R_HOME}"; then
|
||||||
|
echo "could not determine R_HOME"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
CXX17=`"${R_HOME}/bin/R" CMD config CXX17`
|
||||||
|
CXX17STD=`"${R_HOME}/bin/R" CMD config CXX17STD`
|
||||||
|
CXX="${CXX17} ${CXX17STD}"
|
||||||
|
CXXFLAGS=`"${R_HOME}/bin/R" CMD config CXXFLAGS`
|
||||||
|
|
||||||
|
CC=`"${R_HOME}/bin/R" CMD config CC`
|
||||||
|
CFLAGS=`"${R_HOME}/bin/R" CMD config CFLAGS`
|
||||||
|
CPPFLAGS=`"${R_HOME}/bin/R" CMD config CPPFLAGS`
|
||||||
|
|
||||||
|
LDFLAGS=`"${R_HOME}/bin/R" CMD config LDFLAGS`
|
||||||
|
AC_LANG(C++)
|
||||||
|
|
||||||
### Check whether backtrace() is part of libc or the external lib libexecinfo
|
### Check whether backtrace() is part of libc or the external lib libexecinfo
|
||||||
AC_MSG_CHECKING([Backtrace lib])
|
AC_MSG_CHECKING([Backtrace lib])
|
||||||
@@ -28,12 +43,19 @@ fi
|
|||||||
|
|
||||||
if test `uname -s` = "Darwin"
|
if test `uname -s` = "Darwin"
|
||||||
then
|
then
|
||||||
OPENMP_CXXFLAGS='-Xclang -fopenmp'
|
if command -v brew &> /dev/null
|
||||||
OPENMP_LIB='-lomp'
|
then
|
||||||
|
HOMEBREW_LIBOMP_PREFIX=`brew --prefix libomp`
|
||||||
|
else
|
||||||
|
# Homebrew not found
|
||||||
|
HOMEBREW_LIBOMP_PREFIX=''
|
||||||
|
fi
|
||||||
|
OPENMP_CXXFLAGS="-Xpreprocessor -fopenmp -I${HOMEBREW_LIBOMP_PREFIX}/include"
|
||||||
|
OPENMP_LIB="-lomp -L${HOMEBREW_LIBOMP_PREFIX}/lib"
|
||||||
ac_pkg_openmp=no
|
ac_pkg_openmp=no
|
||||||
AC_MSG_CHECKING([whether OpenMP will work in a package])
|
AC_MSG_CHECKING([whether OpenMP will work in a package])
|
||||||
AC_LANG_CONFTEST([AC_LANG_PROGRAM([[#include <omp.h>]], [[ return (omp_get_max_threads() <= 1); ]])])
|
AC_LANG_CONFTEST([AC_LANG_PROGRAM([[#include <omp.h>]], [[ return (omp_get_max_threads() <= 1); ]])])
|
||||||
${CC} -o conftest conftest.c ${OPENMP_LIB} ${OPENMP_CXXFLAGS} 2>/dev/null && ./conftest && ac_pkg_openmp=yes
|
${CXX} -o conftest conftest.cpp ${CPPFLAGS} ${LDFLAGS} ${OPENMP_LIB} ${OPENMP_CXXFLAGS} 2>/dev/null && ./conftest && ac_pkg_openmp=yes
|
||||||
AC_MSG_RESULT([${ac_pkg_openmp}])
|
AC_MSG_RESULT([${ac_pkg_openmp}])
|
||||||
if test "${ac_pkg_openmp}" = no; then
|
if test "${ac_pkg_openmp}" = no; then
|
||||||
OPENMP_CXXFLAGS=''
|
OPENMP_CXXFLAGS=''
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
basic_walkthrough Basic feature walkthrough
|
basic_walkthrough Basic feature walkthrough
|
||||||
caret_wrapper Use xgboost to train in caret library
|
caret_wrapper Use xgboost to train in caret library
|
||||||
custom_objective Cutomize loss function, and evaluation metric
|
custom_objective Customize loss function, and evaluation metric
|
||||||
boost_from_prediction Boosting from existing prediction
|
boost_from_prediction Boosting from existing prediction
|
||||||
predict_first_ntree Predicting using first n trees
|
predict_first_ntree Predicting using first n trees
|
||||||
generalized_linear_model Generalized Linear Model
|
generalized_linear_model Generalized Linear Model
|
||||||
@@ -8,8 +8,8 @@ cross_validation Cross validation
|
|||||||
create_sparse_matrix Create Sparse Matrix
|
create_sparse_matrix Create Sparse Matrix
|
||||||
predict_leaf_indices Predicting the corresponding leaves
|
predict_leaf_indices Predicting the corresponding leaves
|
||||||
early_stopping Early Stop in training
|
early_stopping Early Stop in training
|
||||||
poisson_regression Poisson Regression on count data
|
poisson_regression Poisson regression on count data
|
||||||
tweedie_regression Tweddie Regression
|
tweedie_regression Tweedie regression
|
||||||
gpu_accelerated GPU-accelerated tree building algorithms
|
gpu_accelerated GPU-accelerated tree building algorithms
|
||||||
interaction_constraints Interaction constraints among features
|
interaction_constraints Interaction constraints among features
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ XGBoost R Feature Walkthrough
|
|||||||
====
|
====
|
||||||
* [Basic walkthrough of wrappers](basic_walkthrough.R)
|
* [Basic walkthrough of wrappers](basic_walkthrough.R)
|
||||||
* [Train a xgboost model from caret library](caret_wrapper.R)
|
* [Train a xgboost model from caret library](caret_wrapper.R)
|
||||||
* [Cutomize loss function, and evaluation metric](custom_objective.R)
|
* [Customize loss function, and evaluation metric](custom_objective.R)
|
||||||
* [Boosting from existing prediction](boost_from_prediction.R)
|
* [Boosting from existing prediction](boost_from_prediction.R)
|
||||||
* [Predicting using first n trees](predict_first_ntree.R)
|
* [Predicting using first n trees](predict_first_ntree.R)
|
||||||
* [Generalized Linear Model](generalized_linear_model.R)
|
* [Generalized Linear Model](generalized_linear_model.R)
|
||||||
|
|||||||
@@ -40,7 +40,7 @@ print("Train xgboost with verbose 2, also print information about tree")
|
|||||||
bst <- xgboost(data = dtrain, max_depth = 2, eta = 1, nrounds = 2,
|
bst <- xgboost(data = dtrain, max_depth = 2, eta = 1, nrounds = 2,
|
||||||
nthread = 2, objective = "binary:logistic", verbose = 2)
|
nthread = 2, objective = "binary:logistic", verbose = 2)
|
||||||
|
|
||||||
# you can also specify data as file path to a LibSVM format input
|
# you can also specify data as file path to a LIBSVM format input
|
||||||
# since we do not have this file with us, the following line is just for illustration
|
# since we do not have this file with us, the following line is just for illustration
|
||||||
# bst <- xgboost(data = 'agaricus.train.svm', max_depth = 2, eta = 1, nrounds = 2,objective = "binary:logistic")
|
# bst <- xgboost(data = 'agaricus.train.svm', max_depth = 2, eta = 1, nrounds = 2,objective = "binary:logistic")
|
||||||
|
|
||||||
@@ -63,7 +63,7 @@ print(paste("sum(abs(pred2-pred))=", sum(abs(pred2 - pred))))
|
|||||||
# save model to R's raw vector
|
# save model to R's raw vector
|
||||||
raw <- xgb.save.raw(bst)
|
raw <- xgb.save.raw(bst)
|
||||||
# load binary model to R
|
# load binary model to R
|
||||||
bst3 <- xgb.load(raw)
|
bst3 <- xgb.load.raw(raw)
|
||||||
pred3 <- predict(bst3, test$data)
|
pred3 <- predict(bst3, test$data)
|
||||||
# pred3 should be identical to pred
|
# pred3 should be identical to pred
|
||||||
print(paste("sum(abs(pred3-pred))=", sum(abs(pred3 - pred))))
|
print(paste("sum(abs(pred3-pred))=", sum(abs(pred3 - pred))))
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
# install development version of caret library that contains xgboost models
|
# install development version of caret library that contains xgboost models
|
||||||
devtools::install_github("topepo/caret/pkg/caret")
|
|
||||||
require(caret)
|
require(caret)
|
||||||
require(xgboost)
|
require(xgboost)
|
||||||
require(data.table)
|
require(data.table)
|
||||||
@@ -8,14 +7,23 @@ require(e1071)
|
|||||||
|
|
||||||
# Load Arthritis dataset in memory.
|
# Load Arthritis dataset in memory.
|
||||||
data(Arthritis)
|
data(Arthritis)
|
||||||
# Create a copy of the dataset with data.table package (data.table is 100% compliant with R dataframe but its syntax is a lot more consistent and its performance are really good).
|
# Create a copy of the dataset with data.table package
|
||||||
|
# (data.table is 100% compliant with R dataframe but its syntax is a lot more consistent
|
||||||
|
# and its performance are really good).
|
||||||
df <- data.table(Arthritis, keep.rownames = FALSE)
|
df <- data.table(Arthritis, keep.rownames = FALSE)
|
||||||
|
|
||||||
# Let's add some new categorical features to see if it helps. Of course these feature are highly correlated to the Age feature. Usually it's not a good thing in ML, but Tree algorithms (including boosted trees) are able to select the best features, even in case of highly correlated features.
|
# Let's add some new categorical features to see if it helps.
|
||||||
# For the first feature we create groups of age by rounding the real age. Note that we transform it to factor (categorical data) so the algorithm treat them as independant values.
|
# Of course these feature are highly correlated to the Age feature.
|
||||||
|
# Usually it's not a good thing in ML, but Tree algorithms (including boosted trees) are able to select the best features,
|
||||||
|
# even in case of highly correlated features.
|
||||||
|
# For the first feature we create groups of age by rounding the real age.
|
||||||
|
# Note that we transform it to factor (categorical data) so the algorithm treat them as independant values.
|
||||||
df[, AgeDiscret := as.factor(round(Age / 10, 0))]
|
df[, AgeDiscret := as.factor(round(Age / 10, 0))]
|
||||||
|
|
||||||
# Here is an even stronger simplification of the real age with an arbitrary split at 30 years old. I choose this value based on nothing. We will see later if simplifying the information based on arbitrary values is a good strategy (I am sure you already have an idea of how well it will work!).
|
# Here is an even stronger simplification of the real age with an arbitrary split at 30 years old.
|
||||||
|
# I choose this value based on nothing.
|
||||||
|
# We will see later if simplifying the information based on arbitrary values is a good strategy
|
||||||
|
# (I am sure you already have an idea of how well it will work!).
|
||||||
df[, AgeCat := as.factor(ifelse(Age > 30, "Old", "Young"))]
|
df[, AgeCat := as.factor(ifelse(Age > 30, "Old", "Young"))]
|
||||||
|
|
||||||
# We remove ID as there is nothing to learn from this feature (it will just add some noise as the dataset is small).
|
# We remove ID as there is nothing to learn from this feature (it will just add some noise as the dataset is small).
|
||||||
@@ -28,7 +36,8 @@ fitControl <- trainControl(method = "repeatedcv", number = 10, repeats = 2, sear
|
|||||||
# train a xgbTree model using caret::train
|
# train a xgbTree model using caret::train
|
||||||
model <- train(factor(Improved) ~ ., data = df, method = "xgbTree", trControl = fitControl)
|
model <- train(factor(Improved) ~ ., data = df, method = "xgbTree", trControl = fitControl)
|
||||||
|
|
||||||
# Instead of tree for our boosters, you can also fit a linear regression or logistic regression model using xgbLinear
|
# Instead of tree for our boosters, you can also fit a linear regression or logistic regression model
|
||||||
|
# using xgbLinear
|
||||||
# model <- train(factor(Improved)~., data = df, method = "xgbLinear", trControl = fitControl)
|
# model <- train(factor(Improved)~., data = df, method = "xgbLinear", trControl = fitControl)
|
||||||
|
|
||||||
# See model results
|
# See model results
|
||||||
|
|||||||
@@ -2,39 +2,52 @@ require(xgboost)
|
|||||||
require(Matrix)
|
require(Matrix)
|
||||||
require(data.table)
|
require(data.table)
|
||||||
if (!require(vcd)) {
|
if (!require(vcd)) {
|
||||||
install.packages('vcd') #Available in Cran. Used for its dataset with categorical values.
|
install.packages('vcd') #Available in CRAN. Used for its dataset with categorical values.
|
||||||
require(vcd)
|
require(vcd)
|
||||||
}
|
}
|
||||||
# According to its documentation, Xgboost works only on numbers.
|
# According to its documentation, XGBoost works only on numbers.
|
||||||
# Sometimes the dataset we have to work on have categorical data.
|
# Sometimes the dataset we have to work on have categorical data.
|
||||||
# A categorical variable is one which have a fixed number of values. By example, if for each observation a variable called "Colour" can have only "red", "blue" or "green" as value, it is a categorical variable.
|
# A categorical variable is one which have a fixed number of values.
|
||||||
|
# By example, if for each observation a variable called "Colour" can have only
|
||||||
|
# "red", "blue" or "green" as value, it is a categorical variable.
|
||||||
#
|
#
|
||||||
# In R, categorical variable is called Factor.
|
# In R, categorical variable is called Factor.
|
||||||
# Type ?factor in console for more information.
|
# Type ?factor in console for more information.
|
||||||
#
|
#
|
||||||
# In this demo we will see how to transform a dense dataframe with categorical variables to a sparse matrix before analyzing it in Xgboost.
|
# In this demo we will see how to transform a dense dataframe with categorical variables to a sparse matrix
|
||||||
|
# before analyzing it in XGBoost.
|
||||||
# The method we are going to see is usually called "one hot encoding".
|
# The method we are going to see is usually called "one hot encoding".
|
||||||
|
|
||||||
#load Arthritis dataset in memory.
|
#load Arthritis dataset in memory.
|
||||||
data(Arthritis)
|
data(Arthritis)
|
||||||
|
|
||||||
# create a copy of the dataset with data.table package (data.table is 100% compliant with R dataframe but its syntax is a lot more consistent and its performance are really good).
|
# create a copy of the dataset with data.table package
|
||||||
|
# (data.table is 100% compliant with R dataframe but its syntax is a lot more consistent
|
||||||
|
# and its performance are really good).
|
||||||
df <- data.table(Arthritis, keep.rownames = FALSE)
|
df <- data.table(Arthritis, keep.rownames = FALSE)
|
||||||
|
|
||||||
# Let's have a look to the data.table
|
# Let's have a look to the data.table
|
||||||
cat("Print the dataset\n")
|
cat("Print the dataset\n")
|
||||||
print(df)
|
print(df)
|
||||||
|
|
||||||
# 2 columns have factor type, one has ordinal type (ordinal variable is a categorical variable with values wich can be ordered, here: None > Some > Marked).
|
# 2 columns have factor type, one has ordinal type
|
||||||
|
# (ordinal variable is a categorical variable with values which can be ordered, here: None > Some > Marked).
|
||||||
cat("Structure of the dataset\n")
|
cat("Structure of the dataset\n")
|
||||||
str(df)
|
str(df)
|
||||||
|
|
||||||
# Let's add some new categorical features to see if it helps. Of course these feature are highly correlated to the Age feature. Usually it's not a good thing in ML, but Tree algorithms (including boosted trees) are able to select the best features, even in case of highly correlated features.
|
# Let's add some new categorical features to see if it helps.
|
||||||
|
# Of course these feature are highly correlated to the Age feature.
|
||||||
|
# Usually it's not a good thing in ML, but Tree algorithms (including boosted trees) are able to select the best features,
|
||||||
|
# even in case of highly correlated features.
|
||||||
|
|
||||||
# For the first feature we create groups of age by rounding the real age. Note that we transform it to factor (categorical data) so the algorithm treat them as independant values.
|
# For the first feature we create groups of age by rounding the real age.
|
||||||
|
# Note that we transform it to factor (categorical data) so the algorithm treat them as independent values.
|
||||||
df[, AgeDiscret := as.factor(round(Age / 10, 0))]
|
df[, AgeDiscret := as.factor(round(Age / 10, 0))]
|
||||||
|
|
||||||
# Here is an even stronger simplification of the real age with an arbitrary split at 30 years old. I choose this value based on nothing. We will see later if simplifying the information based on arbitrary values is a good strategy (I am sure you already have an idea of how well it will work!).
|
# Here is an even stronger simplification of the real age with an arbitrary split at 30 years old.
|
||||||
|
# I choose this value based on nothing.
|
||||||
|
# We will see later if simplifying the information based on arbitrary values is a good strategy
|
||||||
|
# (I am sure you already have an idea of how well it will work!).
|
||||||
df[, AgeCat := as.factor(ifelse(Age > 30, "Old", "Young"))]
|
df[, AgeCat := as.factor(ifelse(Age > 30, "Old", "Young"))]
|
||||||
|
|
||||||
# We remove ID as there is nothing to learn from this feature (it will just add some noise as the dataset is small).
|
# We remove ID as there is nothing to learn from this feature (it will just add some noise as the dataset is small).
|
||||||
@@ -48,7 +61,10 @@ print(levels(df[, Treatment]))
|
|||||||
# This method is also called one hot encoding.
|
# This method is also called one hot encoding.
|
||||||
# The purpose is to transform each value of each categorical feature in one binary feature.
|
# The purpose is to transform each value of each categorical feature in one binary feature.
|
||||||
#
|
#
|
||||||
# Let's take, the column Treatment will be replaced by two columns, Placebo, and Treated. Each of them will be binary. For example an observation which had the value Placebo in column Treatment before the transformation will have, after the transformation, the value 1 in the new column Placebo and the value 0 in the new column Treated.
|
# Let's take, the column Treatment will be replaced by two columns, Placebo, and Treated.
|
||||||
|
# Each of them will be binary.
|
||||||
|
# For example an observation which had the value Placebo in column Treatment before the transformation will have, after the transformation,
|
||||||
|
# the value 1 in the new column Placebo and the value 0 in the new column Treated.
|
||||||
#
|
#
|
||||||
# Formulae Improved~.-1 used below means transform all categorical features but column Improved to binary values.
|
# Formulae Improved~.-1 used below means transform all categorical features but column Improved to binary values.
|
||||||
# Column Improved is excluded because it will be our output column, the one we want to predict.
|
# Column Improved is excluded because it will be our output column, the one we want to predict.
|
||||||
@@ -70,7 +86,10 @@ bst <- xgboost(data = sparse_matrix, label = output_vector, max_depth = 9,
|
|||||||
|
|
||||||
importance <- xgb.importance(feature_names = colnames(sparse_matrix), model = bst)
|
importance <- xgb.importance(feature_names = colnames(sparse_matrix), model = bst)
|
||||||
print(importance)
|
print(importance)
|
||||||
# According to the matrix below, the most important feature in this dataset to predict if the treatment will work is the Age. The second most important feature is having received a placebo or not. The sex is third. Then we see our generated features (AgeDiscret). We can see that their contribution is very low (Gain column).
|
# According to the matrix below, the most important feature in this dataset to predict if the treatment will work is the Age.
|
||||||
|
# The second most important feature is having received a placebo or not.
|
||||||
|
# The sex is third.
|
||||||
|
# Then we see our generated features (AgeDiscret). We can see that their contribution is very low (Gain column).
|
||||||
|
|
||||||
# Does these result make sense?
|
# Does these result make sense?
|
||||||
# Let's check some Chi2 between each of these features and the outcome.
|
# Let's check some Chi2 between each of these features and the outcome.
|
||||||
@@ -82,8 +101,17 @@ print(chisq.test(df$AgeDiscret, df$Y))
|
|||||||
# Our first simplification of Age gives a Pearson correlation of 8.
|
# Our first simplification of Age gives a Pearson correlation of 8.
|
||||||
|
|
||||||
print(chisq.test(df$AgeCat, df$Y))
|
print(chisq.test(df$AgeCat, df$Y))
|
||||||
# The perfectly random split I did between young and old at 30 years old have a low correlation of 2. It's a result we may expect as may be in my mind > 30 years is being old (I am 32 and starting feeling old, this may explain that), but for the illness we are studying, the age to be vulnerable is not the same. Don't let your "gut" lower the quality of your model. In "data science", there is science :-)
|
# The perfectly random split I did between young and old at 30 years old have a low correlation of 2.
|
||||||
|
# It's a result we may expect as may be in my mind > 30 years is being old (I am 32 and starting feeling old, this may explain that),
|
||||||
|
# but for the illness we are studying, the age to be vulnerable is not the same.
|
||||||
|
# Don't let your "gut" lower the quality of your model. In "data science", there is science :-)
|
||||||
|
|
||||||
# As you can see, in general destroying information by simplifying it won't improve your model. Chi2 just demonstrates that. But in more complex cases, creating a new feature based on existing one which makes link with the outcome more obvious may help the algorithm and improve the model. The case studied here is not enough complex to show that. Check Kaggle forum for some challenging datasets.
|
# As you can see, in general destroying information by simplifying it won't improve your model.
|
||||||
|
# Chi2 just demonstrates that.
|
||||||
|
# But in more complex cases, creating a new feature based on existing one which makes link with the outcome
|
||||||
|
# more obvious may help the algorithm and improve the model.
|
||||||
|
# The case studied here is not enough complex to show that. Check Kaggle forum for some challenging datasets.
|
||||||
# However it's almost always worse when you add some arbitrary rules.
|
# However it's almost always worse when you add some arbitrary rules.
|
||||||
# Moreover, you can notice that even if we have added some not useful new features highly correlated with other features, the boosting tree algorithm have been able to choose the best one, which in this case is the Age. Linear model may not be that strong in these scenario.
|
# Moreover, you can notice that even if we have added some not useful new features highly correlated with
|
||||||
|
# other features, the boosting tree algorithm have been able to choose the best one, which in this case is the Age.
|
||||||
|
# Linear model may not be that strong in these scenario.
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ cat('running cross validation\n')
|
|||||||
# do cross validation, this will print result out as
|
# do cross validation, this will print result out as
|
||||||
# [iteration] metric_name:mean_value+std_value
|
# [iteration] metric_name:mean_value+std_value
|
||||||
# std_value is standard deviation of the metric
|
# std_value is standard deviation of the metric
|
||||||
xgb.cv(param, dtrain, nrounds, nfold = 5, metrics = {'error'})
|
xgb.cv(param, dtrain, nrounds, nfold = 5, metrics = 'error')
|
||||||
|
|
||||||
cat('running cross validation, disable standard deviation display\n')
|
cat('running cross validation, disable standard deviation display\n')
|
||||||
# do cross validation, this will print result out as
|
# do cross validation, this will print result out as
|
||||||
@@ -22,10 +22,10 @@ xgb.cv(param, dtrain, nrounds, nfold = 5,
|
|||||||
metrics = 'error', showsd = FALSE)
|
metrics = 'error', showsd = FALSE)
|
||||||
|
|
||||||
###
|
###
|
||||||
# you can also do cross validation with cutomized loss function
|
# you can also do cross validation with customized loss function
|
||||||
# See custom_objective.R
|
# See custom_objective.R
|
||||||
##
|
##
|
||||||
print ('running cross validation, with cutomsized loss function')
|
print ('running cross validation, with customized loss function')
|
||||||
|
|
||||||
logregobj <- function(preds, dtrain) {
|
logregobj <- function(preds, dtrain) {
|
||||||
labels <- getinfo(dtrain, "label")
|
labels <- getinfo(dtrain, "label")
|
||||||
|
|||||||
@@ -23,9 +23,9 @@ logregobj <- function(preds, dtrain) {
|
|||||||
|
|
||||||
# user defined evaluation function, return a pair metric_name, result
|
# user defined evaluation function, return a pair metric_name, result
|
||||||
# NOTE: when you do customized loss function, the default prediction value is margin
|
# NOTE: when you do customized loss function, the default prediction value is margin
|
||||||
# this may make buildin evalution metric not function properly
|
# this may make builtin evaluation metric not function properly
|
||||||
# for example, we are doing logistic loss, the prediction is score before logistic transformation
|
# for example, we are doing logistic loss, the prediction is score before logistic transformation
|
||||||
# the buildin evaluation error assumes input is after logistic transformation
|
# the builtin evaluation error assumes input is after logistic transformation
|
||||||
# Take this in mind when you use the customization, and maybe you need write customized evaluation function
|
# Take this in mind when you use the customization, and maybe you need write customized evaluation function
|
||||||
evalerror <- function(preds, dtrain) {
|
evalerror <- function(preds, dtrain) {
|
||||||
labels <- getinfo(dtrain, "label")
|
labels <- getinfo(dtrain, "label")
|
||||||
|
|||||||
@@ -21,9 +21,9 @@ logregobj <- function(preds, dtrain) {
|
|||||||
}
|
}
|
||||||
# user defined evaluation function, return a pair metric_name, result
|
# user defined evaluation function, return a pair metric_name, result
|
||||||
# NOTE: when you do customized loss function, the default prediction value is margin
|
# NOTE: when you do customized loss function, the default prediction value is margin
|
||||||
# this may make buildin evalution metric not function properly
|
# this may make builtin evaluation metric not function properly
|
||||||
# for example, we are doing logistic loss, the prediction is score before logistic transformation
|
# for example, we are doing logistic loss, the prediction is score before logistic transformation
|
||||||
# the buildin evaluation error assumes input is after logistic transformation
|
# the builtin evaluation error assumes input is after logistic transformation
|
||||||
# Take this in mind when you use the customization, and maybe you need write customized evaluation function
|
# Take this in mind when you use the customization, and maybe you need write customized evaluation function
|
||||||
evalerror <- function(preds, dtrain) {
|
evalerror <- function(preds, dtrain) {
|
||||||
labels <- getinfo(dtrain, "label")
|
labels <- getinfo(dtrain, "label")
|
||||||
|
|||||||
@@ -33,7 +33,7 @@ treeInteractions <- function(input_tree, input_max_depth) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
# Extract nodes with interactions
|
# Extract nodes with interactions
|
||||||
interaction_trees <- trees[!is.na(Split) & !is.na(parent_1),
|
interaction_trees <- trees[!is.na(Split) & !is.na(parent_1), # nolint: object_usage_linter
|
||||||
c('Feature', paste0('parent_feat_', 1:(input_max_depth - 1))),
|
c('Feature', paste0('parent_feat_', 1:(input_max_depth - 1))),
|
||||||
with = FALSE]
|
with = FALSE]
|
||||||
interaction_trees_split <- split(interaction_trees, seq_len(nrow(interaction_trees)))
|
interaction_trees_split <- split(interaction_trees, seq_len(nrow(interaction_trees)))
|
||||||
@@ -44,7 +44,7 @@ treeInteractions <- function(input_tree, input_max_depth) {
|
|||||||
|
|
||||||
# Remove non-interactions (same variable)
|
# Remove non-interactions (same variable)
|
||||||
interaction_list <- lapply(interaction_list, unique) # remove same variables
|
interaction_list <- lapply(interaction_list, unique) # remove same variables
|
||||||
interaction_length <- sapply(interaction_list, length)
|
interaction_length <- lengths(interaction_list)
|
||||||
interaction_list <- interaction_list[interaction_length > 1]
|
interaction_list <- interaction_list[interaction_length > 1]
|
||||||
interaction_list <- unique(lapply(interaction_list, sort))
|
interaction_list <- unique(lapply(interaction_list, sort))
|
||||||
return(interaction_list)
|
return(interaction_list)
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
# running all scripts in demo folder
|
# running all scripts in demo folder, removed during packaging.
|
||||||
demo(basic_walkthrough, package = 'xgboost')
|
demo(basic_walkthrough, package = 'xgboost')
|
||||||
demo(custom_objective, package = 'xgboost')
|
demo(custom_objective, package = 'xgboost')
|
||||||
demo(boost_from_prediction, package = 'xgboost')
|
demo(boost_from_prediction, package = 'xgboost')
|
||||||
|
|||||||
@@ -79,9 +79,9 @@ end_of_table <- empty_lines[empty_lines > start_index][1L]
|
|||||||
|
|
||||||
# Read the contents of the table
|
# Read the contents of the table
|
||||||
exported_symbols <- objdump_results[(start_index + 1L):end_of_table]
|
exported_symbols <- objdump_results[(start_index + 1L):end_of_table]
|
||||||
exported_symbols <- gsub("\t", "", exported_symbols)
|
exported_symbols <- gsub("\t", "", exported_symbols, fixed = TRUE)
|
||||||
exported_symbols <- gsub(".*\\] ", "", exported_symbols)
|
exported_symbols <- gsub(".*\\] ", "", exported_symbols)
|
||||||
exported_symbols <- gsub(" ", "", exported_symbols)
|
exported_symbols <- gsub(" ", "", exported_symbols, fixed = TRUE)
|
||||||
|
|
||||||
# Write R.def file
|
# Write R.def file
|
||||||
writeLines(
|
writeLines(
|
||||||
|
|||||||
@@ -38,10 +38,7 @@ The following additional fields are assigned to the model's R object:
|
|||||||
\itemize{
|
\itemize{
|
||||||
\item \code{best_score} the evaluation score at the best iteration
|
\item \code{best_score} the evaluation score at the best iteration
|
||||||
\item \code{best_iteration} at which boosting iteration the best score has occurred (1-based index)
|
\item \code{best_iteration} at which boosting iteration the best score has occurred (1-based index)
|
||||||
\item \code{best_ntreelimit} to use with the \code{ntreelimit} parameter in \code{predict}.
|
|
||||||
It differs from \code{best_iteration} in multiclass or random forest settings.
|
|
||||||
}
|
}
|
||||||
|
|
||||||
The Same values are also stored as xgb-attributes:
|
The Same values are also stored as xgb-attributes:
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{best_iteration} is stored as a 0-based iteration index (for interoperability of binary models)
|
\item \code{best_iteration} is stored as a 0-based iteration index (for interoperability of binary models)
|
||||||
|
|||||||
@@ -8,16 +8,18 @@ during its training.}
|
|||||||
cb.gblinear.history(sparse = FALSE)
|
cb.gblinear.history(sparse = FALSE)
|
||||||
}
|
}
|
||||||
\arguments{
|
\arguments{
|
||||||
\item{sparse}{when set to FALSE/TURE, a dense/sparse matrix is used to store the result.
|
\item{sparse}{when set to FALSE/TRUE, a dense/sparse matrix is used to store the result.
|
||||||
Sparse format is useful when one expects only a subset of coefficients to be non-zero,
|
Sparse format is useful when one expects only a subset of coefficients to be non-zero,
|
||||||
when using the "thrifty" feature selector with fairly small number of top features
|
when using the "thrifty" feature selector with fairly small number of top features
|
||||||
selected per iteration.}
|
selected per iteration.}
|
||||||
}
|
}
|
||||||
\value{
|
\value{
|
||||||
Results are stored in the \code{coefs} element of the closure.
|
Results are stored in the \code{coefs} element of the closure.
|
||||||
The \code{\link{xgb.gblinear.history}} convenience function provides an easy way to access it.
|
The \code{\link{xgb.gblinear.history}} convenience function provides an easy
|
||||||
|
way to access it.
|
||||||
With \code{xgb.train}, it is either a dense of a sparse matrix.
|
With \code{xgb.train}, it is either a dense of a sparse matrix.
|
||||||
While with \code{xgb.cv}, it is a list (an element per each fold) of such matrices.
|
While with \code{xgb.cv}, it is a list (an element per each fold) of such
|
||||||
|
matrices.
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
Callback closure for collecting the model coefficients history of a gblinear booster
|
Callback closure for collecting the model coefficients history of a gblinear booster
|
||||||
@@ -36,10 +38,9 @@ Callback function expects the following values to be set in its calling frame:
|
|||||||
#
|
#
|
||||||
# In the iris dataset, it is hard to linearly separate Versicolor class from the rest
|
# In the iris dataset, it is hard to linearly separate Versicolor class from the rest
|
||||||
# without considering the 2nd order interactions:
|
# without considering the 2nd order interactions:
|
||||||
require(magrittr)
|
|
||||||
x <- model.matrix(Species ~ .^2, iris)[,-1]
|
x <- model.matrix(Species ~ .^2, iris)[,-1]
|
||||||
colnames(x)
|
colnames(x)
|
||||||
dtrain <- xgb.DMatrix(scale(x), label = 1*(iris$Species == "versicolor"))
|
dtrain <- xgb.DMatrix(scale(x), label = 1*(iris$Species == "versicolor"), nthread = 2)
|
||||||
param <- list(booster = "gblinear", objective = "reg:logistic", eval_metric = "auc",
|
param <- list(booster = "gblinear", objective = "reg:logistic", eval_metric = "auc",
|
||||||
lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
||||||
# For 'shotgun', which is a default linear updater, using high eta values may result in
|
# For 'shotgun', which is a default linear updater, using high eta values may result in
|
||||||
@@ -57,7 +58,7 @@ matplot(coef_path, type = 'l')
|
|||||||
bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 200, eta = 0.8,
|
bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 200, eta = 0.8,
|
||||||
updater = 'coord_descent', feature_selector = 'thrifty', top_k = 1,
|
updater = 'coord_descent', feature_selector = 'thrifty', top_k = 1,
|
||||||
callbacks = list(cb.gblinear.history()))
|
callbacks = list(cb.gblinear.history()))
|
||||||
xgb.gblinear.history(bst) \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst), type = 'l')
|
||||||
# Componentwise boosting is known to have similar effect to Lasso regularization.
|
# Componentwise boosting is known to have similar effect to Lasso regularization.
|
||||||
# Try experimenting with various values of top_k, eta, nrounds,
|
# Try experimenting with various values of top_k, eta, nrounds,
|
||||||
# as well as different feature_selectors.
|
# as well as different feature_selectors.
|
||||||
@@ -66,28 +67,28 @@ xgb.gblinear.history(bst) \%>\% matplot(type = 'l')
|
|||||||
bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 100, eta = 0.8,
|
bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 100, eta = 0.8,
|
||||||
callbacks = list(cb.gblinear.history()))
|
callbacks = list(cb.gblinear.history()))
|
||||||
# coefficients in the CV fold #3
|
# coefficients in the CV fold #3
|
||||||
xgb.gblinear.history(bst)[[3]] \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst)[[3]], type = 'l')
|
||||||
|
|
||||||
|
|
||||||
#### Multiclass classification:
|
#### Multiclass classification:
|
||||||
#
|
#
|
||||||
dtrain <- xgb.DMatrix(scale(x), label = as.numeric(iris$Species) - 1)
|
dtrain <- xgb.DMatrix(scale(x), label = as.numeric(iris$Species) - 1, nthread = 1)
|
||||||
param <- list(booster = "gblinear", objective = "multi:softprob", num_class = 3,
|
param <- list(booster = "gblinear", objective = "multi:softprob", num_class = 3,
|
||||||
lambda = 0.0003, alpha = 0.0003, nthread = 2)
|
lambda = 0.0003, alpha = 0.0003, nthread = 1)
|
||||||
# For the default linear updater 'shotgun' it sometimes is helpful
|
# For the default linear updater 'shotgun' it sometimes is helpful
|
||||||
# to use smaller eta to reduce instability
|
# to use smaller eta to reduce instability
|
||||||
bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 70, eta = 0.5,
|
bst <- xgb.train(param, dtrain, list(tr=dtrain), nrounds = 50, eta = 0.5,
|
||||||
callbacks = list(cb.gblinear.history()))
|
callbacks = list(cb.gblinear.history()))
|
||||||
# Will plot the coefficient paths separately for each class:
|
# Will plot the coefficient paths separately for each class:
|
||||||
xgb.gblinear.history(bst, class_index = 0) \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst, class_index = 0), type = 'l')
|
||||||
xgb.gblinear.history(bst, class_index = 1) \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst, class_index = 1), type = 'l')
|
||||||
xgb.gblinear.history(bst, class_index = 2) \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst, class_index = 2), type = 'l')
|
||||||
|
|
||||||
# CV:
|
# CV:
|
||||||
bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 70, eta = 0.5,
|
bst <- xgb.cv(param, dtrain, nfold = 5, nrounds = 70, eta = 0.5,
|
||||||
callbacks = list(cb.gblinear.history(FALSE)))
|
callbacks = list(cb.gblinear.history(FALSE)))
|
||||||
# 1st forld of 1st class
|
# 1st fold of 1st class
|
||||||
xgb.gblinear.history(bst, class_index = 0)[[1]] \%>\% matplot(type = 'l')
|
matplot(xgb.gblinear.history(bst, class_index = 0)[[1]], type = 'l')
|
||||||
|
|
||||||
}
|
}
|
||||||
\seealso{
|
\seealso{
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ be directly used with an \code{xgb.DMatrix} object.
|
|||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
train <- agaricus.train
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
dtrain <- xgb.DMatrix(train$data, label=train$label, nthread = 2)
|
||||||
|
|
||||||
stopifnot(nrow(dtrain) == nrow(train$data))
|
stopifnot(nrow(dtrain) == nrow(train$data))
|
||||||
stopifnot(ncol(dtrain) == ncol(train$data))
|
stopifnot(ncol(dtrain) == ncol(train$data))
|
||||||
|
|||||||
@@ -26,7 +26,7 @@ Since row names are irrelevant, it is recommended to use \code{colnames} directl
|
|||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
train <- agaricus.train
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
dtrain <- xgb.DMatrix(train$data, label=train$label, nthread = 2)
|
||||||
dimnames(dtrain)
|
dimnames(dtrain)
|
||||||
colnames(dtrain)
|
colnames(dtrain)
|
||||||
colnames(dtrain) <- make.names(1:ncol(train$data))
|
colnames(dtrain) <- make.names(1:ncol(train$data))
|
||||||
|
|||||||
@@ -23,9 +23,9 @@ Get information of an xgb.DMatrix object
|
|||||||
The \code{name} field can be one of the following:
|
The \code{name} field can be one of the following:
|
||||||
|
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{label}: label Xgboost learn from ;
|
\item \code{label}: label XGBoost learn from ;
|
||||||
\item \code{weight}: to do a weight rescale ;
|
\item \code{weight}: to do a weight rescale ;
|
||||||
\item \code{base_margin}: base margin is the base prediction Xgboost will boost from ;
|
\item \code{base_margin}: base margin is the base prediction XGBoost will boost from ;
|
||||||
\item \code{nrow}: number of rows of the \code{xgb.DMatrix}.
|
\item \code{nrow}: number of rows of the \code{xgb.DMatrix}.
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -34,8 +34,7 @@ The \code{name} field can be one of the following:
|
|||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
|
|
||||||
labels <- getinfo(dtrain, 'label')
|
labels <- getinfo(dtrain, 'label')
|
||||||
setinfo(dtrain, 'label', 1-labels)
|
setinfo(dtrain, 'label', 1-labels)
|
||||||
|
|||||||
@@ -1,18 +0,0 @@
|
|||||||
% Generated by roxygen2: do not edit by hand
|
|
||||||
% Please edit documentation in R/xgb.ggplot.R
|
|
||||||
\name{normalize}
|
|
||||||
\alias{normalize}
|
|
||||||
\title{Scale feature value to have mean 0, standard deviation 1}
|
|
||||||
\usage{
|
|
||||||
normalize(x)
|
|
||||||
}
|
|
||||||
\arguments{
|
|
||||||
\item{x}{Numeric vector}
|
|
||||||
}
|
|
||||||
\value{
|
|
||||||
Numeric vector with mean 0 and sd 1.
|
|
||||||
}
|
|
||||||
\description{
|
|
||||||
This is used to compare multiple features on the same plot.
|
|
||||||
Internal utility function
|
|
||||||
}
|
|
||||||
@@ -17,6 +17,8 @@
|
|||||||
predinteraction = FALSE,
|
predinteraction = FALSE,
|
||||||
reshape = FALSE,
|
reshape = FALSE,
|
||||||
training = FALSE,
|
training = FALSE,
|
||||||
|
iterationrange = NULL,
|
||||||
|
strict_shape = FALSE,
|
||||||
...
|
...
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -25,7 +27,11 @@
|
|||||||
\arguments{
|
\arguments{
|
||||||
\item{object}{Object of class \code{xgb.Booster} or \code{xgb.Booster.handle}}
|
\item{object}{Object of class \code{xgb.Booster} or \code{xgb.Booster.handle}}
|
||||||
|
|
||||||
\item{newdata}{takes \code{matrix}, \code{dgCMatrix}, local data file or \code{xgb.DMatrix}.}
|
\item{newdata}{takes \code{matrix}, \code{dgCMatrix}, \code{dgRMatrix}, \code{dsparseVector},
|
||||||
|
local data file or \code{xgb.DMatrix}.
|
||||||
|
|
||||||
|
For single-row predictions on sparse data, it's recommended to use CSR format. If passing
|
||||||
|
a sparse vector, it will take it as a row vector.}
|
||||||
|
|
||||||
\item{missing}{Missing is only used when input is dense matrix. Pick a float value that represents
|
\item{missing}{Missing is only used when input is dense matrix. Pick a float value that represents
|
||||||
missing values in data (e.g., sometimes 0 or some other extreme value is used).}
|
missing values in data (e.g., sometimes 0 or some other extreme value is used).}
|
||||||
@@ -34,8 +40,7 @@ missing values in data (e.g., sometimes 0 or some other extreme value is used).}
|
|||||||
sum of predictions from boosting iterations' results. E.g., setting \code{outputmargin=TRUE} for
|
sum of predictions from boosting iterations' results. E.g., setting \code{outputmargin=TRUE} for
|
||||||
logistic regression would result in predictions for log-odds instead of probabilities.}
|
logistic regression would result in predictions for log-odds instead of probabilities.}
|
||||||
|
|
||||||
\item{ntreelimit}{limit the number of model's trees or boosting iterations used in prediction (see Details).
|
\item{ntreelimit}{Deprecated, use \code{iterationrange} instead.}
|
||||||
It will use all the trees by default (\code{NULL} value).}
|
|
||||||
|
|
||||||
\item{predleaf}{whether predict leaf index.}
|
\item{predleaf}{whether predict leaf index.}
|
||||||
|
|
||||||
@@ -52,10 +57,20 @@ or predinteraction flags is TRUE.}
|
|||||||
\item{training}{whether is the prediction result used for training. For dart booster,
|
\item{training}{whether is the prediction result used for training. For dart booster,
|
||||||
training predicting will perform dropout.}
|
training predicting will perform dropout.}
|
||||||
|
|
||||||
|
\item{iterationrange}{Specifies which layer of trees are used in prediction. For
|
||||||
|
example, if a random forest is trained with 100 rounds. Specifying
|
||||||
|
`iterationrange=(1, 21)`, then only the forests built during [1, 21) (half open set)
|
||||||
|
rounds are used in this prediction. It's 1-based index just like R vector. When set
|
||||||
|
to \code{c(1, 1)} XGBoost will use all trees.}
|
||||||
|
|
||||||
|
\item{strict_shape}{Default is \code{FALSE}. When it's set to \code{TRUE}, output
|
||||||
|
type and shape of prediction are invariant to model type.}
|
||||||
|
|
||||||
\item{...}{Parameters passed to \code{predict.xgb.Booster}}
|
\item{...}{Parameters passed to \code{predict.xgb.Booster}}
|
||||||
}
|
}
|
||||||
\value{
|
\value{
|
||||||
For regression or binary classification, it returns a vector of length \code{nrows(newdata)}.
|
The return type is different depending whether \code{strict_shape} is set to \code{TRUE}. By default,
|
||||||
|
for regression or binary classification, it returns a vector of length \code{nrows(newdata)}.
|
||||||
For multiclass classification, either a \code{num_class * nrows(newdata)} vector or
|
For multiclass classification, either a \code{num_class * nrows(newdata)} vector or
|
||||||
a \code{(nrows(newdata), num_class)} dimension matrix is returned, depending on
|
a \code{(nrows(newdata), num_class)} dimension matrix is returned, depending on
|
||||||
the \code{reshape} value.
|
the \code{reshape} value.
|
||||||
@@ -76,18 +91,19 @@ two dimensions. The "+ 1" columns corresponds to bias. Summing this array along
|
|||||||
produce practically the same result as predict with \code{predcontrib = TRUE}.
|
produce practically the same result as predict with \code{predcontrib = TRUE}.
|
||||||
For a multiclass case, a list of \code{num_class} elements is returned, where each element is
|
For a multiclass case, a list of \code{num_class} elements is returned, where each element is
|
||||||
such an array.
|
such an array.
|
||||||
|
|
||||||
|
When \code{strict_shape} is set to \code{TRUE}, the output is always an array. For
|
||||||
|
normal prediction, the output is a 2-dimension array \code{(num_class, nrow(newdata))}.
|
||||||
|
|
||||||
|
For \code{predcontrib = TRUE}, output is \code{(ncol(newdata) + 1, num_class, nrow(newdata))}
|
||||||
|
For \code{predinteraction = TRUE}, output is \code{(ncol(newdata) + 1, ncol(newdata) + 1, num_class, nrow(newdata))}
|
||||||
|
For \code{predleaf = TRUE}, output is \code{(n_trees_in_forest, num_class, n_iterations, nrow(newdata))}
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
Predicted values based on either xgboost model or model handle object.
|
Predicted values based on either xgboost model or model handle object.
|
||||||
}
|
}
|
||||||
\details{
|
\details{
|
||||||
Note that \code{ntreelimit} is not necessarily equal to the number of boosting iterations
|
Note that \code{iterationrange} would currently do nothing for predictions from gblinear,
|
||||||
and it is not necessarily equal to the number of trees in a model.
|
|
||||||
E.g., in a random forest-like model, \code{ntreelimit} would limit the number of trees.
|
|
||||||
But for multiclass classification, while there are multiple trees per iteration,
|
|
||||||
\code{ntreelimit} limits the number of boosting iterations.
|
|
||||||
|
|
||||||
Also note that \code{ntreelimit} would currently do nothing for predictions from gblinear,
|
|
||||||
since gblinear doesn't keep its boosting history.
|
since gblinear doesn't keep its boosting history.
|
||||||
|
|
||||||
One possible practical applications of the \code{predleaf} option is to use the model
|
One possible practical applications of the \code{predleaf} option is to use the model
|
||||||
@@ -106,6 +122,10 @@ With \code{predinteraction = TRUE}, SHAP values of contributions of interaction
|
|||||||
are computed. Note that this operation might be rather expensive in terms of compute and memory.
|
are computed. Note that this operation might be rather expensive in terms of compute and memory.
|
||||||
Since it quadratically depends on the number of features, it is recommended to perform selection
|
Since it quadratically depends on the number of features, it is recommended to perform selection
|
||||||
of the most important features first. See below about the format of the returned results.
|
of the most important features first. See below about the format of the returned results.
|
||||||
|
|
||||||
|
The \code{predict()} method uses as many threads as defined in \code{xgb.Booster} object (all by default).
|
||||||
|
If you want to change their number, then assign a new number to \code{nthread} using \code{\link{xgb.parameters<-}}.
|
||||||
|
Note also that converting a matrix to \code{\link{xgb.DMatrix}} uses multiple threads too.
|
||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
## binary classification:
|
## binary classification:
|
||||||
@@ -120,7 +140,7 @@ bst <- xgboost(data = train$data, label = train$label, max_depth = 2,
|
|||||||
# use all trees by default
|
# use all trees by default
|
||||||
pred <- predict(bst, test$data)
|
pred <- predict(bst, test$data)
|
||||||
# use only the 1st tree
|
# use only the 1st tree
|
||||||
pred1 <- predict(bst, test$data, ntreelimit = 1)
|
pred1 <- predict(bst, test$data, iterationrange = c(1, 2))
|
||||||
|
|
||||||
# Predicting tree leafs:
|
# Predicting tree leafs:
|
||||||
# the result is an nsamples X ntrees matrix
|
# the result is an nsamples X ntrees matrix
|
||||||
@@ -172,25 +192,9 @@ str(pred)
|
|||||||
all.equal(pred, pred_labels)
|
all.equal(pred, pred_labels)
|
||||||
# prediction from using only 5 iterations should result
|
# prediction from using only 5 iterations should result
|
||||||
# in the same error as seen in iteration 5:
|
# in the same error as seen in iteration 5:
|
||||||
pred5 <- predict(bst, as.matrix(iris[, -5]), ntreelimit=5)
|
pred5 <- predict(bst, as.matrix(iris[, -5]), iterationrange=c(1, 6))
|
||||||
sum(pred5 != lb)/length(lb)
|
sum(pred5 != lb)/length(lb)
|
||||||
|
|
||||||
|
|
||||||
## random forest-like model of 25 trees for binary classification:
|
|
||||||
|
|
||||||
set.seed(11)
|
|
||||||
bst <- xgboost(data = train$data, label = train$label, max_depth = 5,
|
|
||||||
nthread = 2, nrounds = 1, objective = "binary:logistic",
|
|
||||||
num_parallel_tree = 25, subsample = 0.6, colsample_bytree = 0.1)
|
|
||||||
# Inspect the prediction error vs number of trees:
|
|
||||||
lb <- test$label
|
|
||||||
dtest <- xgb.DMatrix(test$data, label=lb)
|
|
||||||
err <- sapply(1:25, function(n) {
|
|
||||||
pred <- predict(bst, dtest, ntreelimit=n)
|
|
||||||
sum((pred > 0.5) != lb)/length(lb)
|
|
||||||
})
|
|
||||||
plot(err, type='l', ylim=c(0,0.1), xlab='#trees')
|
|
||||||
|
|
||||||
}
|
}
|
||||||
\references{
|
\references{
|
||||||
Scott M. Lundberg, Su-In Lee, "A Unified Approach to Interpreting Model Predictions", NIPS Proceedings 2017, \url{https://arxiv.org/abs/1705.07874}
|
Scott M. Lundberg, Su-In Lee, "A Unified Approach to Interpreting Model Predictions", NIPS Proceedings 2017, \url{https://arxiv.org/abs/1705.07874}
|
||||||
|
|||||||
@@ -1,27 +0,0 @@
|
|||||||
% Generated by roxygen2: do not edit by hand
|
|
||||||
% Please edit documentation in R/xgb.ggplot.R
|
|
||||||
\name{prepare.ggplot.shap.data}
|
|
||||||
\alias{prepare.ggplot.shap.data}
|
|
||||||
\title{Combine and melt feature values and SHAP contributions for sample
|
|
||||||
observations.}
|
|
||||||
\usage{
|
|
||||||
prepare.ggplot.shap.data(data_list, normalize = FALSE)
|
|
||||||
}
|
|
||||||
\arguments{
|
|
||||||
\item{data_list}{List containing 'data' and 'shap_contrib' returned by
|
|
||||||
\code{xgb.shap.data()}.}
|
|
||||||
|
|
||||||
\item{normalize}{Whether to standardize feature values to have mean 0 and
|
|
||||||
standard deviation 1 (useful for comparing multiple features on the same
|
|
||||||
plot). Default \code{FALSE}.}
|
|
||||||
}
|
|
||||||
\value{
|
|
||||||
A data.table containing the observation ID, the feature name, the
|
|
||||||
feature value (normalized if specified), and the SHAP contribution value.
|
|
||||||
}
|
|
||||||
\description{
|
|
||||||
Conforms to data format required for ggplot functions.
|
|
||||||
}
|
|
||||||
\details{
|
|
||||||
Internal utility function.
|
|
||||||
}
|
|
||||||
@@ -19,8 +19,7 @@ Currently it displays dimensions and presence of info-fields and colnames.
|
|||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
|
|
||||||
dtrain
|
dtrain
|
||||||
print(dtrain, verbose=TRUE)
|
print(dtrain, verbose=TRUE)
|
||||||
|
|||||||
@@ -25,16 +25,15 @@ Set information of an xgb.DMatrix object
|
|||||||
The \code{name} field can be one of the following:
|
The \code{name} field can be one of the following:
|
||||||
|
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{label}: label Xgboost learn from ;
|
\item \code{label}: label XGBoost learn from ;
|
||||||
\item \code{weight}: to do a weight rescale ;
|
\item \code{weight}: to do a weight rescale ;
|
||||||
\item \code{base_margin}: base margin is the base prediction Xgboost will boost from ;
|
\item \code{base_margin}: base margin is the base prediction XGBoost will boost from ;
|
||||||
\item \code{group}: number of rows in each group (to use with \code{rank:pairwise} objective).
|
\item \code{group}: number of rows in each group (to use with \code{rank:pairwise} objective).
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
|
|
||||||
labels <- getinfo(dtrain, 'label')
|
labels <- getinfo(dtrain, 'label')
|
||||||
setinfo(dtrain, 'label', 1-labels)
|
setinfo(dtrain, 'label', 1-labels)
|
||||||
|
|||||||
@@ -28,8 +28,7 @@ original xgb.DMatrix object
|
|||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
|
|
||||||
dsub <- slice(dtrain, 1:42)
|
dsub <- slice(dtrain, 1:42)
|
||||||
labels1 <- getinfo(dsub, 'label')
|
labels1 <- getinfo(dsub, 'label')
|
||||||
|
|||||||
@@ -4,11 +4,20 @@
|
|||||||
\alias{xgb.DMatrix}
|
\alias{xgb.DMatrix}
|
||||||
\title{Construct xgb.DMatrix object}
|
\title{Construct xgb.DMatrix object}
|
||||||
\usage{
|
\usage{
|
||||||
xgb.DMatrix(data, info = list(), missing = NA, silent = FALSE, ...)
|
xgb.DMatrix(
|
||||||
|
data,
|
||||||
|
info = list(),
|
||||||
|
missing = NA,
|
||||||
|
silent = FALSE,
|
||||||
|
nthread = NULL,
|
||||||
|
...
|
||||||
|
)
|
||||||
}
|
}
|
||||||
\arguments{
|
\arguments{
|
||||||
\item{data}{a \code{matrix} object (either numeric or integer), a \code{dgCMatrix} object, or a character
|
\item{data}{a \code{matrix} object (either numeric or integer), a \code{dgCMatrix} object,
|
||||||
string representing a filename.}
|
a \code{dgRMatrix} object (only when making predictions from a fitted model),
|
||||||
|
a \code{dsparseVector} object (only when making predictions from a fitted model, will be
|
||||||
|
interpreted as a row vector), or a character string representing a filename.}
|
||||||
|
|
||||||
\item{info}{a named list of additional information to store in the \code{xgb.DMatrix} object.
|
\item{info}{a named list of additional information to store in the \code{xgb.DMatrix} object.
|
||||||
See \code{\link{setinfo}} for the specific allowed kinds of}
|
See \code{\link{setinfo}} for the specific allowed kinds of}
|
||||||
@@ -18,17 +27,18 @@ It is useful when a 0 or some other extreme value represents missing values in d
|
|||||||
|
|
||||||
\item{silent}{whether to suppress printing an informational message after loading from a file.}
|
\item{silent}{whether to suppress printing an informational message after loading from a file.}
|
||||||
|
|
||||||
|
\item{nthread}{Number of threads used for creating DMatrix.}
|
||||||
|
|
||||||
\item{...}{the \code{info} data could be passed directly as parameters, without creating an \code{info} list.}
|
\item{...}{the \code{info} data could be passed directly as parameters, without creating an \code{info} list.}
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
Construct xgb.DMatrix object from either a dense matrix, a sparse matrix, or a local file.
|
Construct xgb.DMatrix object from either a dense matrix, a sparse matrix, or a local file.
|
||||||
Supported input file formats are either a libsvm text file or a binary file that was created previously by
|
Supported input file formats are either a LIBSVM text file or a binary file that was created previously by
|
||||||
\code{\link{xgb.DMatrix.save}}).
|
\code{\link{xgb.DMatrix.save}}).
|
||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
||||||
dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
||||||
if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
||||||
|
|||||||
@@ -16,8 +16,7 @@ Save xgb.DMatrix object to binary file
|
|||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
train <- agaricus.train
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtrain <- xgb.DMatrix(train$data, label=train$label)
|
|
||||||
xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
xgb.DMatrix.save(dtrain, 'xgb.DMatrix.data')
|
||||||
dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
dtrain <- xgb.DMatrix('xgb.DMatrix.data')
|
||||||
if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
if (file.exists('xgb.DMatrix.data')) file.remove('xgb.DMatrix.data')
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ Joaquin Quinonero Candela)}
|
|||||||
|
|
||||||
International Workshop on Data Mining for Online Advertising (ADKDD) - August 24, 2014
|
International Workshop on Data Mining for Online Advertising (ADKDD) - August 24, 2014
|
||||||
|
|
||||||
\url{https://research.fb.com/publications/practical-lessons-from-predicting-clicks-on-ads-at-facebook/}.
|
\url{https://research.facebook.com/publications/practical-lessons-from-predicting-clicks-on-ads-at-facebook/}.
|
||||||
|
|
||||||
Extract explaining the method:
|
Extract explaining the method:
|
||||||
|
|
||||||
@@ -59,8 +59,8 @@ a rule on certain features."
|
|||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
data(agaricus.test, package='xgboost')
|
data(agaricus.test, package='xgboost')
|
||||||
dtrain <- xgb.DMatrix(data = agaricus.train$data, label = agaricus.train$label)
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtest <- xgb.DMatrix(data = agaricus.test$data, label = agaricus.test$label)
|
dtest <- with(agaricus.test, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
|
|
||||||
param <- list(max_depth=2, eta=1, silent=1, objective='binary:logistic')
|
param <- list(max_depth=2, eta=1, silent=1, objective='binary:logistic')
|
||||||
nrounds = 4
|
nrounds = 4
|
||||||
@@ -76,8 +76,12 @@ new.features.train <- xgb.create.features(model = bst, agaricus.train$data)
|
|||||||
new.features.test <- xgb.create.features(model = bst, agaricus.test$data)
|
new.features.test <- xgb.create.features(model = bst, agaricus.test$data)
|
||||||
|
|
||||||
# learning with new features
|
# learning with new features
|
||||||
new.dtrain <- xgb.DMatrix(data = new.features.train, label = agaricus.train$label)
|
new.dtrain <- xgb.DMatrix(
|
||||||
new.dtest <- xgb.DMatrix(data = new.features.test, label = agaricus.test$label)
|
data = new.features.train, label = agaricus.train$label, nthread = 2
|
||||||
|
)
|
||||||
|
new.dtest <- xgb.DMatrix(
|
||||||
|
data = new.features.test, label = agaricus.test$label, nthread = 2
|
||||||
|
)
|
||||||
watchlist <- list(train = new.dtrain)
|
watchlist <- list(train = new.dtrain)
|
||||||
bst <- xgb.train(params = param, data = new.dtrain, nrounds = nrounds, nthread = 2)
|
bst <- xgb.train(params = param, data = new.dtrain, nrounds = nrounds, nthread = 2)
|
||||||
|
|
||||||
|
|||||||
@@ -135,9 +135,7 @@ An object of class \code{xgb.cv.synchronous} with the following elements:
|
|||||||
parameter or randomly generated.
|
parameter or randomly generated.
|
||||||
\item \code{best_iteration} iteration number with the best evaluation metric value
|
\item \code{best_iteration} iteration number with the best evaluation metric value
|
||||||
(only available with early stopping).
|
(only available with early stopping).
|
||||||
\item \code{best_ntreelimit} the \code{ntreelimit} value corresponding to the best iteration,
|
\item \code{best_ntreelimit} and the \code{ntreelimit} Deprecated attributes, use \code{best_iteration} instead.
|
||||||
which could further be used in \code{predict} method
|
|
||||||
(only available with early stopping).
|
|
||||||
\item \code{pred} CV prediction values available when \code{prediction} is set.
|
\item \code{pred} CV prediction values available when \code{prediction} is set.
|
||||||
It is either vector or matrix (see \code{\link{cb.cv.predict}}).
|
It is either vector or matrix (see \code{\link{cb.cv.predict}}).
|
||||||
\item \code{models} a list of the CV folds' models. It is only available with the explicit
|
\item \code{models} a list of the CV folds' models. It is only available with the explicit
|
||||||
@@ -150,9 +148,11 @@ The cross validation function of xgboost
|
|||||||
\details{
|
\details{
|
||||||
The original sample is randomly partitioned into \code{nfold} equal size subsamples.
|
The original sample is randomly partitioned into \code{nfold} equal size subsamples.
|
||||||
|
|
||||||
Of the \code{nfold} subsamples, a single subsample is retained as the validation data for testing the model, and the remaining \code{nfold - 1} subsamples are used as training data.
|
Of the \code{nfold} subsamples, a single subsample is retained as the validation data for testing the model,
|
||||||
|
and the remaining \code{nfold - 1} subsamples are used as training data.
|
||||||
|
|
||||||
The cross-validation process is then repeated \code{nrounds} times, with each of the \code{nfold} subsamples used exactly once as the validation data.
|
The cross-validation process is then repeated \code{nrounds} times, with each of the
|
||||||
|
\code{nfold} subsamples used exactly once as the validation data.
|
||||||
|
|
||||||
All observations are used for both training and validation.
|
All observations are used for both training and validation.
|
||||||
|
|
||||||
@@ -160,7 +160,7 @@ Adapted from \url{https://en.wikipedia.org/wiki/Cross-validation_\%28statistics\
|
|||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
dtrain <- xgb.DMatrix(agaricus.train$data, label = agaricus.train$label)
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
cv <- xgb.cv(data = dtrain, nrounds = 3, nthread = 2, nfold = 5, metrics = list("rmse","auc"),
|
cv <- xgb.cv(data = dtrain, nrounds = 3, nthread = 2, nfold = 5, metrics = list("rmse","auc"),
|
||||||
max_depth = 3, eta = 1, objective = "binary:logistic")
|
max_depth = 3, eta = 1, objective = "binary:logistic")
|
||||||
print(cv)
|
print(cv)
|
||||||
|
|||||||
@@ -20,8 +20,6 @@ xgb.dump(
|
|||||||
If not provided or set to \code{NULL}, the model is returned as a \code{character} vector.}
|
If not provided or set to \code{NULL}, the model is returned as a \code{character} vector.}
|
||||||
|
|
||||||
\item{fmap}{feature map file representing feature types.
|
\item{fmap}{feature map file representing feature types.
|
||||||
Detailed description could be found at
|
|
||||||
\url{https://github.com/dmlc/xgboost/wiki/Binary-Classification#dump-model}.
|
|
||||||
See demo/ for walkthrough example in R, and
|
See demo/ for walkthrough example in R, and
|
||||||
\url{https://github.com/dmlc/xgboost/blob/master/demo/data/featmap.txt}
|
\url{https://github.com/dmlc/xgboost/blob/master/demo/data/featmap.txt}
|
||||||
for example Format.}
|
for example Format.}
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ An object of \code{xgb.Booster} class.
|
|||||||
Load xgboost model from the binary model file.
|
Load xgboost model from the binary model file.
|
||||||
}
|
}
|
||||||
\details{
|
\details{
|
||||||
The input file is expected to contain a model saved in an xgboost-internal binary format
|
The input file is expected to contain a model saved in an xgboost model format
|
||||||
using either \code{\link{xgb.save}} or \code{\link{cb.save.model}} in R, or using some
|
using either \code{\link{xgb.save}} or \code{\link{cb.save.model}} in R, or using some
|
||||||
appropriate methods from other xgboost interfaces. E.g., a model trained in Python and
|
appropriate methods from other xgboost interfaces. E.g., a model trained in Python and
|
||||||
saved from there in xgboost format, could be loaded from R.
|
saved from there in xgboost format, could be loaded from R.
|
||||||
|
|||||||
@@ -4,10 +4,12 @@
|
|||||||
\alias{xgb.load.raw}
|
\alias{xgb.load.raw}
|
||||||
\title{Load serialised xgboost model from R's raw vector}
|
\title{Load serialised xgboost model from R's raw vector}
|
||||||
\usage{
|
\usage{
|
||||||
xgb.load.raw(buffer)
|
xgb.load.raw(buffer, as_booster = FALSE)
|
||||||
}
|
}
|
||||||
\arguments{
|
\arguments{
|
||||||
\item{buffer}{the buffer returned by xgb.save.raw}
|
\item{buffer}{the buffer returned by xgb.save.raw}
|
||||||
|
|
||||||
|
\item{as_booster}{Return the loaded model as xgb.Booster instead of xgb.Booster.handle.}
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
User can generate raw memory buffer by calling xgb.save.raw
|
User can generate raw memory buffer by calling xgb.save.raw
|
||||||
|
|||||||
@@ -10,7 +10,7 @@ xgb.ggplot.importance(
|
|||||||
top_n = NULL,
|
top_n = NULL,
|
||||||
measure = NULL,
|
measure = NULL,
|
||||||
rel_to_first = FALSE,
|
rel_to_first = FALSE,
|
||||||
n_clusters = c(1:10),
|
n_clusters = seq_len(10),
|
||||||
...
|
...
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -87,7 +87,7 @@ more than 5 distinct values.}
|
|||||||
|
|
||||||
\item{which}{whether to do univariate or bivariate plotting. NOTE: only 1D is implemented so far.}
|
\item{which}{whether to do univariate or bivariate plotting. NOTE: only 1D is implemented so far.}
|
||||||
|
|
||||||
\item{plot}{whether a plot should be drawn. If FALSE, only a lits of matrices is returned.}
|
\item{plot}{whether a plot should be drawn. If FALSE, only a list of matrices is returned.}
|
||||||
|
|
||||||
\item{...}{other parameters passed to \code{plot}.}
|
\item{...}{other parameters passed to \code{plot}.}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -67,12 +67,12 @@ Each point (observation) is coloured based on its feature value. The plot
|
|||||||
hence allows us to see which features have a negative / positive contribution
|
hence allows us to see which features have a negative / positive contribution
|
||||||
on the model prediction, and whether the contribution is different for larger
|
on the model prediction, and whether the contribution is different for larger
|
||||||
or smaller values of the feature. We effectively try to replicate the
|
or smaller values of the feature. We effectively try to replicate the
|
||||||
\code{summary_plot} function from https://github.com/slundberg/shap.
|
\code{summary_plot} function from https://github.com/shap/shap.
|
||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
# See \code{\link{xgb.plot.shap}}.
|
# See \code{\link{xgb.plot.shap}}.
|
||||||
}
|
}
|
||||||
\seealso{
|
\seealso{
|
||||||
\code{\link{xgb.plot.shap}}, \code{\link{xgb.ggplot.shap.summary}},
|
\code{\link{xgb.plot.shap}}, \code{\link{xgb.ggplot.shap.summary}},
|
||||||
\url{https://github.com/slundberg/shap}
|
\url{https://github.com/shap/shap}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -67,7 +67,7 @@ The "Yes" branches are marked by the "< split_value" label.
|
|||||||
The branches that also used for missing values are marked as bold
|
The branches that also used for missing values are marked as bold
|
||||||
(as in "carrying extra capacity").
|
(as in "carrying extra capacity").
|
||||||
|
|
||||||
This function uses \href{http://www.graphviz.org/}{GraphViz} as a backend of DiagrammeR.
|
This function uses \href{https://www.graphviz.org/}{GraphViz} as a backend of DiagrammeR.
|
||||||
}
|
}
|
||||||
\examples{
|
\examples{
|
||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
|
|||||||
@@ -5,10 +5,19 @@
|
|||||||
\title{Save xgboost model to R's raw vector,
|
\title{Save xgboost model to R's raw vector,
|
||||||
user can call xgb.load.raw to load the model back from raw vector}
|
user can call xgb.load.raw to load the model back from raw vector}
|
||||||
\usage{
|
\usage{
|
||||||
xgb.save.raw(model)
|
xgb.save.raw(model, raw_format = "deprecated")
|
||||||
}
|
}
|
||||||
\arguments{
|
\arguments{
|
||||||
\item{model}{the model object.}
|
\item{model}{the model object.}
|
||||||
|
|
||||||
|
\item{raw_format}{The format for encoding the booster. Available options are
|
||||||
|
\itemize{
|
||||||
|
\item \code{json}: Encode the booster into JSON text document.
|
||||||
|
\item \code{ubj}: Encode the booster into Universal Binary JSON.
|
||||||
|
\item \code{deprecated}: Encode the booster into old customized binary format.
|
||||||
|
}
|
||||||
|
|
||||||
|
Right now the default is \code{deprecated} but will be changed to \code{ubj} in upcoming release.}
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
Save xgboost model from xgboost or xgb.train
|
Save xgboost model from xgboost or xgb.train
|
||||||
|
|||||||
@@ -54,21 +54,43 @@ xgboost(
|
|||||||
|
|
||||||
2. Booster Parameters
|
2. Booster Parameters
|
||||||
|
|
||||||
2.1. Parameter for Tree Booster
|
2.1. Parameters for Tree Booster
|
||||||
|
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{eta} control the learning rate: scale the contribution of each tree by a factor of \code{0 < eta < 1} when it is added to the current approximation. Used to prevent overfitting by making the boosting process more conservative. Lower value for \code{eta} implies larger value for \code{nrounds}: low \code{eta} value means model more robust to overfitting but slower to compute. Default: 0.3
|
\item{ \code{eta} control the learning rate: scale the contribution of each tree by a factor of \code{0 < eta < 1}
|
||||||
\item \code{gamma} minimum loss reduction required to make a further partition on a leaf node of the tree. the larger, the more conservative the algorithm will be.
|
when it is added to the current approximation.
|
||||||
|
Used to prevent overfitting by making the boosting process more conservative.
|
||||||
|
Lower value for \code{eta} implies larger value for \code{nrounds}: low \code{eta} value means model
|
||||||
|
more robust to overfitting but slower to compute. Default: 0.3}
|
||||||
|
\item{ \code{gamma} minimum loss reduction required to make a further partition on a leaf node of the tree.
|
||||||
|
the larger, the more conservative the algorithm will be.}
|
||||||
\item \code{max_depth} maximum depth of a tree. Default: 6
|
\item \code{max_depth} maximum depth of a tree. Default: 6
|
||||||
\item \code{min_child_weight} minimum sum of instance weight (hessian) needed in a child. If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight, then the building process will give up further partitioning. In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node. The larger, the more conservative the algorithm will be. Default: 1
|
\item{\code{min_child_weight} minimum sum of instance weight (hessian) needed in a child.
|
||||||
\item \code{subsample} subsample ratio of the training instance. Setting it to 0.5 means that xgboost randomly collected half of the data instances to grow trees and this will prevent overfitting. It makes computation shorter (because less data to analyse). It is advised to use this parameter with \code{eta} and increase \code{nrounds}. Default: 1
|
If the tree partition step results in a leaf node with the sum of instance weight less than min_child_weight,
|
||||||
|
then the building process will give up further partitioning.
|
||||||
|
In linear regression mode, this simply corresponds to minimum number of instances needed to be in each node.
|
||||||
|
The larger, the more conservative the algorithm will be. Default: 1}
|
||||||
|
\item{ \code{subsample} subsample ratio of the training instance.
|
||||||
|
Setting it to 0.5 means that xgboost randomly collected half of the data instances to grow trees
|
||||||
|
and this will prevent overfitting. It makes computation shorter (because less data to analyse).
|
||||||
|
It is advised to use this parameter with \code{eta} and increase \code{nrounds}. Default: 1}
|
||||||
\item \code{colsample_bytree} subsample ratio of columns when constructing each tree. Default: 1
|
\item \code{colsample_bytree} subsample ratio of columns when constructing each tree. Default: 1
|
||||||
\item \code{num_parallel_tree} Experimental parameter. number of trees to grow per round. Useful to test Random Forest through Xgboost (set \code{colsample_bytree < 1}, \code{subsample < 1} and \code{round = 1}) accordingly. Default: 1
|
\item \code{lambda} L2 regularization term on weights. Default: 1
|
||||||
\item \code{monotone_constraints} A numerical vector consists of \code{1}, \code{0} and \code{-1} with its length equals to the number of features in the training data. \code{1} is increasing, \code{-1} is decreasing and \code{0} is no constraint.
|
\item \code{alpha} L1 regularization term on weights. (there is no L1 reg on bias because it is not important). Default: 0
|
||||||
\item \code{interaction_constraints} A list of vectors specifying feature indices of permitted interactions. Each item of the list represents one permitted interaction where specified features are allowed to interact with each other. Feature index values should start from \code{0} (\code{0} references the first column). Leave argument unspecified for no interaction constraints.
|
\item{ \code{num_parallel_tree} Experimental parameter. number of trees to grow per round.
|
||||||
|
Useful to test Random Forest through XGBoost
|
||||||
|
(set \code{colsample_bytree < 1}, \code{subsample < 1} and \code{round = 1}) accordingly.
|
||||||
|
Default: 1}
|
||||||
|
\item{ \code{monotone_constraints} A numerical vector consists of \code{1}, \code{0} and \code{-1} with its length
|
||||||
|
equals to the number of features in the training data.
|
||||||
|
\code{1} is increasing, \code{-1} is decreasing and \code{0} is no constraint.}
|
||||||
|
\item{ \code{interaction_constraints} A list of vectors specifying feature indices of permitted interactions.
|
||||||
|
Each item of the list represents one permitted interaction where specified features are allowed to interact with each other.
|
||||||
|
Feature index values should start from \code{0} (\code{0} references the first column).
|
||||||
|
Leave argument unspecified for no interaction constraints.}
|
||||||
}
|
}
|
||||||
|
|
||||||
2.2. Parameter for Linear Booster
|
2.2. Parameters for Linear Booster
|
||||||
|
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{lambda} L2 regularization term on weights. Default: 0
|
\item \code{lambda} L2 regularization term on weights. Default: 0
|
||||||
@@ -79,29 +101,53 @@ xgboost(
|
|||||||
3. Task Parameters
|
3. Task Parameters
|
||||||
|
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{objective} specify the learning task and the corresponding learning objective, users can pass a self-defined function to it. The default objective options are below:
|
\item{ \code{objective} specify the learning task and the corresponding learning objective, users can pass a self-defined function to it.
|
||||||
|
The default objective options are below:
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{reg:squarederror} Regression with squared loss (Default).
|
\item \code{reg:squarederror} Regression with squared loss (Default).
|
||||||
\item \code{reg:squaredlogerror}: regression with squared log loss \eqn{1/2 * (log(pred + 1) - log(label + 1))^2}. All inputs are required to be greater than -1. Also, see metric rmsle for possible issue with this objective.
|
\item{ \code{reg:squaredlogerror}: regression with squared log loss \eqn{1/2 * (log(pred + 1) - log(label + 1))^2}.
|
||||||
|
All inputs are required to be greater than -1.
|
||||||
|
Also, see metric rmsle for possible issue with this objective.}
|
||||||
\item \code{reg:logistic} logistic regression.
|
\item \code{reg:logistic} logistic regression.
|
||||||
\item \code{reg:pseudohubererror}: regression with Pseudo Huber loss, a twice differentiable alternative to absolute loss.
|
\item \code{reg:pseudohubererror}: regression with Pseudo Huber loss, a twice differentiable alternative to absolute loss.
|
||||||
\item \code{binary:logistic} logistic regression for binary classification. Output probability.
|
\item \code{binary:logistic} logistic regression for binary classification. Output probability.
|
||||||
\item \code{binary:logitraw} logistic regression for binary classification, output score before logistic transformation.
|
\item \code{binary:logitraw} logistic regression for binary classification, output score before logistic transformation.
|
||||||
\item \code{binary:hinge}: hinge loss for binary classification. This makes predictions of 0 or 1, rather than producing probabilities.
|
\item \code{binary:hinge}: hinge loss for binary classification. This makes predictions of 0 or 1, rather than producing probabilities.
|
||||||
\item \code{count:poisson}: poisson regression for count data, output mean of poisson distribution. \code{max_delta_step} is set to 0.7 by default in poisson regression (used to safeguard optimization).
|
\item{ \code{count:poisson}: Poisson regression for count data, output mean of Poisson distribution.
|
||||||
\item \code{survival:cox}: Cox regression for right censored survival time data (negative values are considered right censored). Note that predictions are returned on the hazard ratio scale (i.e., as HR = exp(marginal_prediction) in the proportional hazard function \code{h(t) = h0(t) * HR)}.
|
\code{max_delta_step} is set to 0.7 by default in poisson regression (used to safeguard optimization).}
|
||||||
\item \code{survival:aft}: Accelerated failure time model for censored survival time data. See \href{https://xgboost.readthedocs.io/en/latest/tutorials/aft_survival_analysis.html}{Survival Analysis with Accelerated Failure Time} for details.
|
\item{ \code{survival:cox}: Cox regression for right censored survival time data (negative values are considered right censored).
|
||||||
\item \code{aft_loss_distribution}: Probabilty Density Function used by \code{survival:aft} and \code{aft-nloglik} metric.
|
Note that predictions are returned on the hazard ratio scale (i.e., as HR = exp(marginal_prediction) in the proportional
|
||||||
\item \code{multi:softmax} set xgboost to do multiclass classification using the softmax objective. Class is represented by a number and should be from 0 to \code{num_class - 1}.
|
hazard function \code{h(t) = h0(t) * HR)}.}
|
||||||
\item \code{multi:softprob} same as softmax, but prediction outputs a vector of ndata * nclass elements, which can be further reshaped to ndata, nclass matrix. The result contains predicted probabilities of each data point belonging to each class.
|
\item{ \code{survival:aft}: Accelerated failure time model for censored survival time data. See
|
||||||
|
\href{https://xgboost.readthedocs.io/en/latest/tutorials/aft_survival_analysis.html}{Survival Analysis with Accelerated Failure Time}
|
||||||
|
for details.}
|
||||||
|
\item \code{aft_loss_distribution}: Probability Density Function used by \code{survival:aft} and \code{aft-nloglik} metric.
|
||||||
|
\item{ \code{multi:softmax} set xgboost to do multiclass classification using the softmax objective.
|
||||||
|
Class is represented by a number and should be from 0 to \code{num_class - 1}.}
|
||||||
|
\item{ \code{multi:softprob} same as softmax, but prediction outputs a vector of ndata * nclass elements, which can be
|
||||||
|
further reshaped to ndata, nclass matrix. The result contains predicted probabilities of each data point belonging
|
||||||
|
to each class.}
|
||||||
\item \code{rank:pairwise} set xgboost to do ranking task by minimizing the pairwise loss.
|
\item \code{rank:pairwise} set xgboost to do ranking task by minimizing the pairwise loss.
|
||||||
\item \code{rank:ndcg}: Use LambdaMART to perform list-wise ranking where \href{https://en.wikipedia.org/wiki/Discounted_cumulative_gain}{Normalized Discounted Cumulative Gain (NDCG)} is maximized.
|
\item{ \code{rank:ndcg}: Use LambdaMART to perform list-wise ranking where
|
||||||
\item \code{rank:map}: Use LambdaMART to perform list-wise ranking where \href{https://en.wikipedia.org/wiki/Evaluation_measures_(information_retrieval)#Mean_average_precision}{Mean Average Precision (MAP)} is maximized.
|
\href{https://en.wikipedia.org/wiki/Discounted_cumulative_gain}{Normalized Discounted Cumulative Gain (NDCG)} is maximized.}
|
||||||
\item \code{reg:gamma}: gamma regression with log-link. Output is a mean of gamma distribution. It might be useful, e.g., for modeling insurance claims severity, or for any outcome that might be \href{https://en.wikipedia.org/wiki/Gamma_distribution#Applications}{gamma-distributed}.
|
\item{ \code{rank:map}: Use LambdaMART to perform list-wise ranking where
|
||||||
\item \code{reg:tweedie}: Tweedie regression with log-link. It might be useful, e.g., for modeling total loss in insurance, or for any outcome that might be \href{https://en.wikipedia.org/wiki/Tweedie_distribution#Applications}{Tweedie-distributed}.
|
\href{https://en.wikipedia.org/wiki/Evaluation_measures_(information_retrieval)#Mean_average_precision}{Mean Average Precision (MAP)}
|
||||||
|
is maximized.}
|
||||||
|
\item{ \code{reg:gamma}: gamma regression with log-link.
|
||||||
|
Output is a mean of gamma distribution.
|
||||||
|
It might be useful, e.g., for modeling insurance claims severity, or for any outcome that might be
|
||||||
|
\href{https://en.wikipedia.org/wiki/Gamma_distribution#Applications}{gamma-distributed}.}
|
||||||
|
\item{ \code{reg:tweedie}: Tweedie regression with log-link.
|
||||||
|
It might be useful, e.g., for modeling total loss in insurance, or for any outcome that might be
|
||||||
|
\href{https://en.wikipedia.org/wiki/Tweedie_distribution#Applications}{Tweedie-distributed}.}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
\item \code{base_score} the initial prediction score of all instances, global bias. Default: 0.5
|
\item \code{base_score} the initial prediction score of all instances, global bias. Default: 0.5
|
||||||
\item \code{eval_metric} evaluation metrics for validation data. Users can pass a self-defined function to it. Default: metric will be assigned according to objective(rmse for regression, and error for classification, mean average precision for ranking). List is provided in detail section.
|
\item{ \code{eval_metric} evaluation metrics for validation data.
|
||||||
|
Users can pass a self-defined function to it.
|
||||||
|
Default: metric will be assigned according to objective
|
||||||
|
(rmse for regression, and error for classification, mean average precision for ranking).
|
||||||
|
List is provided in detail section.}
|
||||||
}}
|
}}
|
||||||
|
|
||||||
\item{data}{training dataset. \code{xgb.train} accepts only an \code{xgb.DMatrix} as the input.
|
\item{data}{training dataset. \code{xgb.train} accepts only an \code{xgb.DMatrix} as the input.
|
||||||
@@ -185,9 +231,6 @@ An object of class \code{xgb.Booster} with the following elements:
|
|||||||
explicitly passed.
|
explicitly passed.
|
||||||
\item \code{best_iteration} iteration number with the best evaluation metric value
|
\item \code{best_iteration} iteration number with the best evaluation metric value
|
||||||
(only available with early stopping).
|
(only available with early stopping).
|
||||||
\item \code{best_ntreelimit} the \code{ntreelimit} value corresponding to the best iteration,
|
|
||||||
which could further be used in \code{predict} method
|
|
||||||
(only available with early stopping).
|
|
||||||
\item \code{best_score} the best evaluation metric value during early stopping.
|
\item \code{best_score} the best evaluation metric value during early stopping.
|
||||||
(only available with early stopping).
|
(only available with early stopping).
|
||||||
\item \code{feature_names} names of the training dataset features
|
\item \code{feature_names} names of the training dataset features
|
||||||
@@ -209,11 +252,11 @@ than the \code{xgboost} interface.
|
|||||||
Parallelization is automatically enabled if \code{OpenMP} is present.
|
Parallelization is automatically enabled if \code{OpenMP} is present.
|
||||||
Number of threads can also be manually specified via \code{nthread} parameter.
|
Number of threads can also be manually specified via \code{nthread} parameter.
|
||||||
|
|
||||||
The evaluation metric is chosen automatically by Xgboost (according to the objective)
|
The evaluation metric is chosen automatically by XGBoost (according to the objective)
|
||||||
when the \code{eval_metric} parameter is not provided.
|
when the \code{eval_metric} parameter is not provided.
|
||||||
User may set one or several \code{eval_metric} parameters.
|
User may set one or several \code{eval_metric} parameters.
|
||||||
Note that when using a customized metric, only this single metric can be used.
|
Note that when using a customized metric, only this single metric can be used.
|
||||||
The following is the list of built-in metrics for which Xgboost provides optimized implementation:
|
The following is the list of built-in metrics for which XGBoost provides optimized implementation:
|
||||||
\itemize{
|
\itemize{
|
||||||
\item \code{rmse} root mean square error. \url{https://en.wikipedia.org/wiki/Root_mean_square_error}
|
\item \code{rmse} root mean square error. \url{https://en.wikipedia.org/wiki/Root_mean_square_error}
|
||||||
\item \code{logloss} negative log-likelihood. \url{https://en.wikipedia.org/wiki/Log-likelihood}
|
\item \code{logloss} negative log-likelihood. \url{https://en.wikipedia.org/wiki/Log-likelihood}
|
||||||
@@ -224,7 +267,8 @@ The following is the list of built-in metrics for which Xgboost provides optimiz
|
|||||||
\item \code{merror} Multiclass classification error rate. It is calculated as \code{(# wrong cases) / (# all cases)}.
|
\item \code{merror} Multiclass classification error rate. It is calculated as \code{(# wrong cases) / (# all cases)}.
|
||||||
\item \code{mae} Mean absolute error
|
\item \code{mae} Mean absolute error
|
||||||
\item \code{mape} Mean absolute percentage error
|
\item \code{mape} Mean absolute percentage error
|
||||||
\item \code{auc} Area under the curve. \url{https://en.wikipedia.org/wiki/Receiver_operating_characteristic#'Area_under_curve} for ranking evaluation.
|
\item{ \code{auc} Area under the curve.
|
||||||
|
\url{https://en.wikipedia.org/wiki/Receiver_operating_characteristic#'Area_under_curve} for ranking evaluation.}
|
||||||
\item \code{aucpr} Area under the PR curve. \url{https://en.wikipedia.org/wiki/Precision_and_recall} for ranking evaluation.
|
\item \code{aucpr} Area under the PR curve. \url{https://en.wikipedia.org/wiki/Precision_and_recall} for ranking evaluation.
|
||||||
\item \code{ndcg} Normalized Discounted Cumulative Gain (for ranking task). \url{https://en.wikipedia.org/wiki/NDCG}
|
\item \code{ndcg} Normalized Discounted Cumulative Gain (for ranking task). \url{https://en.wikipedia.org/wiki/NDCG}
|
||||||
}
|
}
|
||||||
@@ -242,8 +286,8 @@ The following callbacks are automatically created when certain parameters are se
|
|||||||
data(agaricus.train, package='xgboost')
|
data(agaricus.train, package='xgboost')
|
||||||
data(agaricus.test, package='xgboost')
|
data(agaricus.test, package='xgboost')
|
||||||
|
|
||||||
dtrain <- xgb.DMatrix(agaricus.train$data, label = agaricus.train$label)
|
dtrain <- with(agaricus.train, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
dtest <- xgb.DMatrix(agaricus.test$data, label = agaricus.test$label)
|
dtest <- with(agaricus.test, xgb.DMatrix(data, label = label, nthread = 2))
|
||||||
watchlist <- list(train = dtrain, eval = dtest)
|
watchlist <- list(train = dtrain, eval = dtest)
|
||||||
|
|
||||||
## A simple xgb.train example:
|
## A simple xgb.train example:
|
||||||
|
|||||||
@@ -4,10 +4,17 @@
|
|||||||
\alias{xgb.unserialize}
|
\alias{xgb.unserialize}
|
||||||
\title{Load the instance back from \code{\link{xgb.serialize}}}
|
\title{Load the instance back from \code{\link{xgb.serialize}}}
|
||||||
\usage{
|
\usage{
|
||||||
xgb.unserialize(buffer)
|
xgb.unserialize(buffer, handle = NULL)
|
||||||
}
|
}
|
||||||
\arguments{
|
\arguments{
|
||||||
\item{buffer}{the buffer containing booster instance saved by \code{\link{xgb.serialize}}}
|
\item{buffer}{the buffer containing booster instance saved by \code{\link{xgb.serialize}}}
|
||||||
|
|
||||||
|
\item{handle}{An \code{xgb.Booster.handle} object which will be overwritten with
|
||||||
|
the new deserialized object. Must be a null handle (e.g. when loading the model through
|
||||||
|
`readRDS`). If not provided, a new handle will be created.}
|
||||||
|
}
|
||||||
|
\value{
|
||||||
|
An \code{xgb.Booster.handle} object.
|
||||||
}
|
}
|
||||||
\description{
|
\description{
|
||||||
Load the instance back from \code{\link{xgb.serialize}}
|
Load the instance back from \code{\link{xgb.serialize}}
|
||||||
|
|||||||
@@ -3,12 +3,11 @@ PKGROOT=../../
|
|||||||
ENABLE_STD_THREAD=1
|
ENABLE_STD_THREAD=1
|
||||||
# _*_ mode: Makefile; _*_
|
# _*_ mode: Makefile; _*_
|
||||||
|
|
||||||
CXX_STD = CXX14
|
CXX_STD = CXX17
|
||||||
|
|
||||||
XGB_RFLAGS = -DXGBOOST_STRICT_R_MODE=1 -DDMLC_LOG_BEFORE_THROW=0\
|
XGB_RFLAGS = -DXGBOOST_STRICT_R_MODE=1 -DDMLC_LOG_BEFORE_THROW=0\
|
||||||
-DDMLC_ENABLE_STD_THREAD=$(ENABLE_STD_THREAD) -DDMLC_DISABLE_STDIN=1\
|
-DDMLC_ENABLE_STD_THREAD=$(ENABLE_STD_THREAD) -DDMLC_DISABLE_STDIN=1\
|
||||||
-DDMLC_LOG_CUSTOMIZE=1 -DXGBOOST_CUSTOMIZE_LOGGER=1\
|
-DDMLC_LOG_CUSTOMIZE=1
|
||||||
-DRABIT_CUSTOMIZE_MSG_
|
|
||||||
|
|
||||||
# disable the use of thread_local for 32 bit windows:
|
# disable the use of thread_local for 32 bit windows:
|
||||||
ifeq ($(R_OSTYPE)$(WIN),windows)
|
ifeq ($(R_OSTYPE)$(WIN),windows)
|
||||||
@@ -17,9 +16,94 @@ endif
|
|||||||
$(foreach v, $(XGB_RFLAGS), $(warning $(v)))
|
$(foreach v, $(XGB_RFLAGS), $(warning $(v)))
|
||||||
|
|
||||||
PKG_CPPFLAGS= -I$(PKGROOT)/include -I$(PKGROOT)/dmlc-core/include -I$(PKGROOT)/rabit/include -I$(PKGROOT) $(XGB_RFLAGS)
|
PKG_CPPFLAGS= -I$(PKGROOT)/include -I$(PKGROOT)/dmlc-core/include -I$(PKGROOT)/rabit/include -I$(PKGROOT) $(XGB_RFLAGS)
|
||||||
PKG_CXXFLAGS= @OPENMP_CXXFLAGS@ @ENDIAN_FLAG@ -pthread
|
PKG_CXXFLAGS= @OPENMP_CXXFLAGS@ @ENDIAN_FLAG@ -pthread $(CXX_VISIBILITY)
|
||||||
PKG_LIBS = @OPENMP_CXXFLAGS@ @OPENMP_LIB@ @ENDIAN_FLAG@ @BACKTRACE_LIB@ -pthread
|
PKG_LIBS = @OPENMP_CXXFLAGS@ @OPENMP_LIB@ @ENDIAN_FLAG@ @BACKTRACE_LIB@ -pthread
|
||||||
OBJECTS= ./xgboost_R.o ./xgboost_custom.o ./xgboost_assert.o ./init.o \
|
|
||||||
$(PKGROOT)/amalgamation/xgboost-all0.o $(PKGROOT)/amalgamation/dmlc-minimum0.o \
|
OBJECTS= \
|
||||||
$(PKGROOT)/rabit/src/engine.o $(PKGROOT)/rabit/src/c_api.o \
|
./xgboost_R.o \
|
||||||
|
./xgboost_custom.o \
|
||||||
|
./init.o \
|
||||||
|
$(PKGROOT)/src/metric/metric.o \
|
||||||
|
$(PKGROOT)/src/metric/elementwise_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/multiclass_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/rank_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/auc.o \
|
||||||
|
$(PKGROOT)/src/metric/survival_metric.o \
|
||||||
|
$(PKGROOT)/src/objective/objective.o \
|
||||||
|
$(PKGROOT)/src/objective/regression_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/multiclass_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/lambdarank_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/hinge.o \
|
||||||
|
$(PKGROOT)/src/objective/aft_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/adaptive.o \
|
||||||
|
$(PKGROOT)/src/objective/init_estimation.o \
|
||||||
|
$(PKGROOT)/src/objective/quantile_obj.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbm.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbtree.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbtree_model.o \
|
||||||
|
$(PKGROOT)/src/gbm/gblinear.o \
|
||||||
|
$(PKGROOT)/src/gbm/gblinear_model.o \
|
||||||
|
$(PKGROOT)/src/data/simple_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/data.o \
|
||||||
|
$(PKGROOT)/src/data/sparse_page_raw_format.o \
|
||||||
|
$(PKGROOT)/src/data/ellpack_page.o \
|
||||||
|
$(PKGROOT)/src/data/file_iterator.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index_page_source.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index_format.o \
|
||||||
|
$(PKGROOT)/src/data/sparse_page_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/proxy_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/iterative_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/predictor/predictor.o \
|
||||||
|
$(PKGROOT)/src/predictor/cpu_predictor.o \
|
||||||
|
$(PKGROOT)/src/predictor/cpu_treeshap.o \
|
||||||
|
$(PKGROOT)/src/tree/constraints.o \
|
||||||
|
$(PKGROOT)/src/tree/param.o \
|
||||||
|
$(PKGROOT)/src/tree/fit_stump.o \
|
||||||
|
$(PKGROOT)/src/tree/tree_model.o \
|
||||||
|
$(PKGROOT)/src/tree/tree_updater.o \
|
||||||
|
$(PKGROOT)/src/tree/multi_target_tree_model.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_approx.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_colmaker.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_prune.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_quantile_hist.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_refresh.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_sync.o \
|
||||||
|
$(PKGROOT)/src/tree/hist/param.o \
|
||||||
|
$(PKGROOT)/src/tree/hist/histogram.o \
|
||||||
|
$(PKGROOT)/src/linear/linear_updater.o \
|
||||||
|
$(PKGROOT)/src/linear/updater_coordinate.o \
|
||||||
|
$(PKGROOT)/src/linear/updater_shotgun.o \
|
||||||
|
$(PKGROOT)/src/learner.o \
|
||||||
|
$(PKGROOT)/src/context.o \
|
||||||
|
$(PKGROOT)/src/logging.o \
|
||||||
|
$(PKGROOT)/src/global_config.o \
|
||||||
|
$(PKGROOT)/src/collective/communicator.o \
|
||||||
|
$(PKGROOT)/src/collective/in_memory_communicator.o \
|
||||||
|
$(PKGROOT)/src/collective/in_memory_handler.o \
|
||||||
|
$(PKGROOT)/src/collective/socket.o \
|
||||||
|
$(PKGROOT)/src/common/charconv.o \
|
||||||
|
$(PKGROOT)/src/common/column_matrix.o \
|
||||||
|
$(PKGROOT)/src/common/common.o \
|
||||||
|
$(PKGROOT)/src/common/error_msg.o \
|
||||||
|
$(PKGROOT)/src/common/hist_util.o \
|
||||||
|
$(PKGROOT)/src/common/host_device_vector.o \
|
||||||
|
$(PKGROOT)/src/common/io.o \
|
||||||
|
$(PKGROOT)/src/common/json.o \
|
||||||
|
$(PKGROOT)/src/common/numeric.o \
|
||||||
|
$(PKGROOT)/src/common/pseudo_huber.o \
|
||||||
|
$(PKGROOT)/src/common/quantile.o \
|
||||||
|
$(PKGROOT)/src/common/random.o \
|
||||||
|
$(PKGROOT)/src/common/stats.o \
|
||||||
|
$(PKGROOT)/src/common/survival_util.o \
|
||||||
|
$(PKGROOT)/src/common/threading_utils.o \
|
||||||
|
$(PKGROOT)/src/common/ranking_utils.o \
|
||||||
|
$(PKGROOT)/src/common/quantile_loss_utils.o \
|
||||||
|
$(PKGROOT)/src/common/timer.o \
|
||||||
|
$(PKGROOT)/src/common/version.o \
|
||||||
|
$(PKGROOT)/src/c_api/c_api.o \
|
||||||
|
$(PKGROOT)/src/c_api/c_api_error.o \
|
||||||
|
$(PKGROOT)/amalgamation/dmlc-minimum0.o \
|
||||||
|
$(PKGROOT)/rabit/src/engine.o \
|
||||||
|
$(PKGROOT)/rabit/src/rabit_c_api.o \
|
||||||
$(PKGROOT)/rabit/src/allreduce_base.o
|
$(PKGROOT)/rabit/src/allreduce_base.o
|
||||||
|
|||||||
@@ -1,26 +1,13 @@
|
|||||||
# package root
|
# package root
|
||||||
PKGROOT=./
|
PKGROOT=../../
|
||||||
ENABLE_STD_THREAD=0
|
ENABLE_STD_THREAD=0
|
||||||
# _*_ mode: Makefile; _*_
|
# _*_ mode: Makefile; _*_
|
||||||
|
|
||||||
# This file is only used for windows compilation from github
|
CXX_STD = CXX17
|
||||||
# It will be replaced with Makevars.in for the CRAN version
|
|
||||||
.PHONY: all xgblib
|
|
||||||
all: $(SHLIB)
|
|
||||||
$(SHLIB): xgblib
|
|
||||||
xgblib:
|
|
||||||
cp -r ../../src .
|
|
||||||
cp -r ../../rabit .
|
|
||||||
cp -r ../../dmlc-core .
|
|
||||||
cp -r ../../include .
|
|
||||||
cp -r ../../amalgamation .
|
|
||||||
|
|
||||||
CXX_STD = CXX14
|
|
||||||
|
|
||||||
XGB_RFLAGS = -DXGBOOST_STRICT_R_MODE=1 -DDMLC_LOG_BEFORE_THROW=0\
|
XGB_RFLAGS = -DXGBOOST_STRICT_R_MODE=1 -DDMLC_LOG_BEFORE_THROW=0\
|
||||||
-DDMLC_ENABLE_STD_THREAD=$(ENABLE_STD_THREAD) -DDMLC_DISABLE_STDIN=1\
|
-DDMLC_ENABLE_STD_THREAD=$(ENABLE_STD_THREAD) -DDMLC_DISABLE_STDIN=1\
|
||||||
-DDMLC_LOG_CUSTOMIZE=1 -DXGBOOST_CUSTOMIZE_LOGGER=1\
|
-DDMLC_LOG_CUSTOMIZE=1
|
||||||
-DRABIT_CUSTOMIZE_MSG_
|
|
||||||
|
|
||||||
# disable the use of thread_local for 32 bit windows:
|
# disable the use of thread_local for 32 bit windows:
|
||||||
ifeq ($(R_OSTYPE)$(WIN),windows)
|
ifeq ($(R_OSTYPE)$(WIN),windows)
|
||||||
@@ -29,11 +16,94 @@ endif
|
|||||||
$(foreach v, $(XGB_RFLAGS), $(warning $(v)))
|
$(foreach v, $(XGB_RFLAGS), $(warning $(v)))
|
||||||
|
|
||||||
PKG_CPPFLAGS= -I$(PKGROOT)/include -I$(PKGROOT)/dmlc-core/include -I$(PKGROOT)/rabit/include -I$(PKGROOT) $(XGB_RFLAGS)
|
PKG_CPPFLAGS= -I$(PKGROOT)/include -I$(PKGROOT)/dmlc-core/include -I$(PKGROOT)/rabit/include -I$(PKGROOT) $(XGB_RFLAGS)
|
||||||
PKG_CXXFLAGS= $(SHLIB_OPENMP_CXXFLAGS) $(SHLIB_PTHREAD_FLAGS)
|
PKG_CXXFLAGS= $(SHLIB_OPENMP_CXXFLAGS) -DDMLC_CMAKE_LITTLE_ENDIAN=1 $(SHLIB_PTHREAD_FLAGS) $(CXX_VISIBILITY)
|
||||||
PKG_LIBS = $(SHLIB_OPENMP_CXXFLAGS) $(SHLIB_PTHREAD_FLAGS)
|
PKG_LIBS = $(SHLIB_OPENMP_CXXFLAGS) -DDMLC_CMAKE_LITTLE_ENDIAN=1 $(SHLIB_PTHREAD_FLAGS) -lwsock32 -lws2_32
|
||||||
OBJECTS= ./xgboost_R.o ./xgboost_custom.o ./xgboost_assert.o ./init.o \
|
|
||||||
$(PKGROOT)/amalgamation/xgboost-all0.o $(PKGROOT)/amalgamation/dmlc-minimum0.o \
|
|
||||||
$(PKGROOT)/rabit/src/engine.o $(PKGROOT)/rabit/src/c_api.o \
|
|
||||||
$(PKGROOT)/rabit/src/allreduce_base.o
|
|
||||||
|
|
||||||
$(OBJECTS) : xgblib
|
OBJECTS= \
|
||||||
|
./xgboost_R.o \
|
||||||
|
./xgboost_custom.o \
|
||||||
|
./init.o \
|
||||||
|
$(PKGROOT)/src/metric/metric.o \
|
||||||
|
$(PKGROOT)/src/metric/elementwise_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/multiclass_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/rank_metric.o \
|
||||||
|
$(PKGROOT)/src/metric/auc.o \
|
||||||
|
$(PKGROOT)/src/metric/survival_metric.o \
|
||||||
|
$(PKGROOT)/src/objective/objective.o \
|
||||||
|
$(PKGROOT)/src/objective/regression_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/multiclass_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/lambdarank_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/hinge.o \
|
||||||
|
$(PKGROOT)/src/objective/aft_obj.o \
|
||||||
|
$(PKGROOT)/src/objective/adaptive.o \
|
||||||
|
$(PKGROOT)/src/objective/init_estimation.o \
|
||||||
|
$(PKGROOT)/src/objective/quantile_obj.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbm.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbtree.o \
|
||||||
|
$(PKGROOT)/src/gbm/gbtree_model.o \
|
||||||
|
$(PKGROOT)/src/gbm/gblinear.o \
|
||||||
|
$(PKGROOT)/src/gbm/gblinear_model.o \
|
||||||
|
$(PKGROOT)/src/data/simple_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/data.o \
|
||||||
|
$(PKGROOT)/src/data/sparse_page_raw_format.o \
|
||||||
|
$(PKGROOT)/src/data/ellpack_page.o \
|
||||||
|
$(PKGROOT)/src/data/file_iterator.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index_page_source.o \
|
||||||
|
$(PKGROOT)/src/data/gradient_index_format.o \
|
||||||
|
$(PKGROOT)/src/data/sparse_page_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/proxy_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/data/iterative_dmatrix.o \
|
||||||
|
$(PKGROOT)/src/predictor/predictor.o \
|
||||||
|
$(PKGROOT)/src/predictor/cpu_predictor.o \
|
||||||
|
$(PKGROOT)/src/predictor/cpu_treeshap.o \
|
||||||
|
$(PKGROOT)/src/tree/constraints.o \
|
||||||
|
$(PKGROOT)/src/tree/param.o \
|
||||||
|
$(PKGROOT)/src/tree/fit_stump.o \
|
||||||
|
$(PKGROOT)/src/tree/tree_model.o \
|
||||||
|
$(PKGROOT)/src/tree/multi_target_tree_model.o \
|
||||||
|
$(PKGROOT)/src/tree/tree_updater.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_approx.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_colmaker.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_prune.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_quantile_hist.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_refresh.o \
|
||||||
|
$(PKGROOT)/src/tree/updater_sync.o \
|
||||||
|
$(PKGROOT)/src/tree/hist/param.o \
|
||||||
|
$(PKGROOT)/src/tree/hist/histogram.o \
|
||||||
|
$(PKGROOT)/src/linear/linear_updater.o \
|
||||||
|
$(PKGROOT)/src/linear/updater_coordinate.o \
|
||||||
|
$(PKGROOT)/src/linear/updater_shotgun.o \
|
||||||
|
$(PKGROOT)/src/learner.o \
|
||||||
|
$(PKGROOT)/src/context.o \
|
||||||
|
$(PKGROOT)/src/logging.o \
|
||||||
|
$(PKGROOT)/src/global_config.o \
|
||||||
|
$(PKGROOT)/src/collective/communicator.o \
|
||||||
|
$(PKGROOT)/src/collective/in_memory_communicator.o \
|
||||||
|
$(PKGROOT)/src/collective/in_memory_handler.o \
|
||||||
|
$(PKGROOT)/src/collective/socket.o \
|
||||||
|
$(PKGROOT)/src/common/charconv.o \
|
||||||
|
$(PKGROOT)/src/common/column_matrix.o \
|
||||||
|
$(PKGROOT)/src/common/common.o \
|
||||||
|
$(PKGROOT)/src/common/error_msg.o \
|
||||||
|
$(PKGROOT)/src/common/hist_util.o \
|
||||||
|
$(PKGROOT)/src/common/host_device_vector.o \
|
||||||
|
$(PKGROOT)/src/common/io.o \
|
||||||
|
$(PKGROOT)/src/common/json.o \
|
||||||
|
$(PKGROOT)/src/common/numeric.o \
|
||||||
|
$(PKGROOT)/src/common/pseudo_huber.o \
|
||||||
|
$(PKGROOT)/src/common/quantile.o \
|
||||||
|
$(PKGROOT)/src/common/random.o \
|
||||||
|
$(PKGROOT)/src/common/stats.o \
|
||||||
|
$(PKGROOT)/src/common/survival_util.o \
|
||||||
|
$(PKGROOT)/src/common/threading_utils.o \
|
||||||
|
$(PKGROOT)/src/common/ranking_utils.o \
|
||||||
|
$(PKGROOT)/src/common/quantile_loss_utils.o \
|
||||||
|
$(PKGROOT)/src/common/timer.o \
|
||||||
|
$(PKGROOT)/src/common/version.o \
|
||||||
|
$(PKGROOT)/src/c_api/c_api.o \
|
||||||
|
$(PKGROOT)/src/c_api/c_api_error.o \
|
||||||
|
$(PKGROOT)/amalgamation/dmlc-minimum0.o \
|
||||||
|
$(PKGROOT)/rabit/src/engine.o \
|
||||||
|
$(PKGROOT)/rabit/src/rabit_c_api.o \
|
||||||
|
$(PKGROOT)/rabit/src/allreduce_base.o
|
||||||
|
|||||||
@@ -9,6 +9,7 @@
|
|||||||
#include <Rinternals.h>
|
#include <Rinternals.h>
|
||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
#include <R_ext/Rdynload.h>
|
#include <R_ext/Rdynload.h>
|
||||||
|
#include <R_ext/Visibility.h>
|
||||||
|
|
||||||
/* FIXME:
|
/* FIXME:
|
||||||
Check these declarations against the C/Fortran source code.
|
Check these declarations against the C/Fortran source code.
|
||||||
@@ -17,73 +18,83 @@ Check these declarations against the C/Fortran source code.
|
|||||||
/* .Call calls */
|
/* .Call calls */
|
||||||
extern SEXP XGBoosterBoostOneIter_R(SEXP, SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterBoostOneIter_R(SEXP, SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterCreate_R(SEXP);
|
extern SEXP XGBoosterCreate_R(SEXP);
|
||||||
|
extern SEXP XGBoosterCreateInEmptyObj_R(SEXP, SEXP);
|
||||||
extern SEXP XGBoosterDumpModel_R(SEXP, SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterDumpModel_R(SEXP, SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterEvalOneIter_R(SEXP, SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterEvalOneIter_R(SEXP, SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterGetAttrNames_R(SEXP);
|
extern SEXP XGBoosterGetAttrNames_R(SEXP);
|
||||||
extern SEXP XGBoosterGetAttr_R(SEXP, SEXP);
|
extern SEXP XGBoosterGetAttr_R(SEXP, SEXP);
|
||||||
extern SEXP XGBoosterLoadModelFromRaw_R(SEXP, SEXP);
|
extern SEXP XGBoosterLoadModelFromRaw_R(SEXP, SEXP);
|
||||||
|
extern SEXP XGBoosterSaveModelToRaw_R(SEXP handle, SEXP config);
|
||||||
extern SEXP XGBoosterLoadModel_R(SEXP, SEXP);
|
extern SEXP XGBoosterLoadModel_R(SEXP, SEXP);
|
||||||
extern SEXP XGBoosterSaveJsonConfig_R(SEXP handle);
|
extern SEXP XGBoosterSaveJsonConfig_R(SEXP handle);
|
||||||
extern SEXP XGBoosterLoadJsonConfig_R(SEXP handle, SEXP value);
|
extern SEXP XGBoosterLoadJsonConfig_R(SEXP handle, SEXP value);
|
||||||
extern SEXP XGBoosterSerializeToBuffer_R(SEXP handle);
|
extern SEXP XGBoosterSerializeToBuffer_R(SEXP handle);
|
||||||
extern SEXP XGBoosterUnserializeFromBuffer_R(SEXP handle, SEXP raw);
|
extern SEXP XGBoosterUnserializeFromBuffer_R(SEXP handle, SEXP raw);
|
||||||
extern SEXP XGBoosterModelToRaw_R(SEXP);
|
extern SEXP XGBoosterPredictFromDMatrix_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterPredict_R(SEXP, SEXP, SEXP, SEXP, SEXP);
|
|
||||||
extern SEXP XGBoosterSaveModel_R(SEXP, SEXP);
|
extern SEXP XGBoosterSaveModel_R(SEXP, SEXP);
|
||||||
extern SEXP XGBoosterSetAttr_R(SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterSetAttr_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterSetParam_R(SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterSetParam_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGBoosterUpdateOneIter_R(SEXP, SEXP, SEXP);
|
extern SEXP XGBoosterUpdateOneIter_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGCheckNullPtr_R(SEXP);
|
extern SEXP XGCheckNullPtr_R(SEXP);
|
||||||
extern SEXP XGDMatrixCreateFromCSC_R(SEXP, SEXP, SEXP, SEXP);
|
extern SEXP XGDMatrixCreateFromCSC_R(SEXP, SEXP, SEXP, SEXP, SEXP, SEXP);
|
||||||
|
extern SEXP XGDMatrixCreateFromCSR_R(SEXP, SEXP, SEXP, SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixCreateFromFile_R(SEXP, SEXP);
|
extern SEXP XGDMatrixCreateFromFile_R(SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixCreateFromMat_R(SEXP, SEXP);
|
extern SEXP XGDMatrixCreateFromMat_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixGetInfo_R(SEXP, SEXP);
|
extern SEXP XGDMatrixGetInfo_R(SEXP, SEXP);
|
||||||
|
extern SEXP XGDMatrixGetStrFeatureInfo_R(SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixNumCol_R(SEXP);
|
extern SEXP XGDMatrixNumCol_R(SEXP);
|
||||||
extern SEXP XGDMatrixNumRow_R(SEXP);
|
extern SEXP XGDMatrixNumRow_R(SEXP);
|
||||||
extern SEXP XGDMatrixSaveBinary_R(SEXP, SEXP, SEXP);
|
extern SEXP XGDMatrixSaveBinary_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixSetInfo_R(SEXP, SEXP, SEXP);
|
extern SEXP XGDMatrixSetInfo_R(SEXP, SEXP, SEXP);
|
||||||
|
extern SEXP XGDMatrixSetStrFeatureInfo_R(SEXP, SEXP, SEXP);
|
||||||
extern SEXP XGDMatrixSliceDMatrix_R(SEXP, SEXP);
|
extern SEXP XGDMatrixSliceDMatrix_R(SEXP, SEXP);
|
||||||
extern SEXP XGBSetGlobalConfig_R(SEXP);
|
extern SEXP XGBSetGlobalConfig_R(SEXP);
|
||||||
extern SEXP XGBGetGlobalConfig_R();
|
extern SEXP XGBGetGlobalConfig_R(void);
|
||||||
|
extern SEXP XGBoosterFeatureScore_R(SEXP, SEXP);
|
||||||
|
|
||||||
static const R_CallMethodDef CallEntries[] = {
|
static const R_CallMethodDef CallEntries[] = {
|
||||||
{"XGBoosterBoostOneIter_R", (DL_FUNC) &XGBoosterBoostOneIter_R, 4},
|
{"XGBoosterBoostOneIter_R", (DL_FUNC) &XGBoosterBoostOneIter_R, 4},
|
||||||
{"XGBoosterCreate_R", (DL_FUNC) &XGBoosterCreate_R, 1},
|
{"XGBoosterCreate_R", (DL_FUNC) &XGBoosterCreate_R, 1},
|
||||||
|
{"XGBoosterCreateInEmptyObj_R", (DL_FUNC) &XGBoosterCreateInEmptyObj_R, 2},
|
||||||
{"XGBoosterDumpModel_R", (DL_FUNC) &XGBoosterDumpModel_R, 4},
|
{"XGBoosterDumpModel_R", (DL_FUNC) &XGBoosterDumpModel_R, 4},
|
||||||
{"XGBoosterEvalOneIter_R", (DL_FUNC) &XGBoosterEvalOneIter_R, 4},
|
{"XGBoosterEvalOneIter_R", (DL_FUNC) &XGBoosterEvalOneIter_R, 4},
|
||||||
{"XGBoosterGetAttrNames_R", (DL_FUNC) &XGBoosterGetAttrNames_R, 1},
|
{"XGBoosterGetAttrNames_R", (DL_FUNC) &XGBoosterGetAttrNames_R, 1},
|
||||||
{"XGBoosterGetAttr_R", (DL_FUNC) &XGBoosterGetAttr_R, 2},
|
{"XGBoosterGetAttr_R", (DL_FUNC) &XGBoosterGetAttr_R, 2},
|
||||||
{"XGBoosterLoadModelFromRaw_R", (DL_FUNC) &XGBoosterLoadModelFromRaw_R, 2},
|
{"XGBoosterLoadModelFromRaw_R", (DL_FUNC) &XGBoosterLoadModelFromRaw_R, 2},
|
||||||
|
{"XGBoosterSaveModelToRaw_R", (DL_FUNC) &XGBoosterSaveModelToRaw_R, 2},
|
||||||
{"XGBoosterLoadModel_R", (DL_FUNC) &XGBoosterLoadModel_R, 2},
|
{"XGBoosterLoadModel_R", (DL_FUNC) &XGBoosterLoadModel_R, 2},
|
||||||
{"XGBoosterSaveJsonConfig_R", (DL_FUNC) &XGBoosterSaveJsonConfig_R, 1},
|
{"XGBoosterSaveJsonConfig_R", (DL_FUNC) &XGBoosterSaveJsonConfig_R, 1},
|
||||||
{"XGBoosterLoadJsonConfig_R", (DL_FUNC) &XGBoosterLoadJsonConfig_R, 2},
|
{"XGBoosterLoadJsonConfig_R", (DL_FUNC) &XGBoosterLoadJsonConfig_R, 2},
|
||||||
{"XGBoosterSerializeToBuffer_R", (DL_FUNC) &XGBoosterSerializeToBuffer_R, 1},
|
{"XGBoosterSerializeToBuffer_R", (DL_FUNC) &XGBoosterSerializeToBuffer_R, 1},
|
||||||
{"XGBoosterUnserializeFromBuffer_R", (DL_FUNC) &XGBoosterUnserializeFromBuffer_R, 2},
|
{"XGBoosterUnserializeFromBuffer_R", (DL_FUNC) &XGBoosterUnserializeFromBuffer_R, 2},
|
||||||
{"XGBoosterModelToRaw_R", (DL_FUNC) &XGBoosterModelToRaw_R, 1},
|
{"XGBoosterPredictFromDMatrix_R", (DL_FUNC) &XGBoosterPredictFromDMatrix_R, 3},
|
||||||
{"XGBoosterPredict_R", (DL_FUNC) &XGBoosterPredict_R, 5},
|
|
||||||
{"XGBoosterSaveModel_R", (DL_FUNC) &XGBoosterSaveModel_R, 2},
|
{"XGBoosterSaveModel_R", (DL_FUNC) &XGBoosterSaveModel_R, 2},
|
||||||
{"XGBoosterSetAttr_R", (DL_FUNC) &XGBoosterSetAttr_R, 3},
|
{"XGBoosterSetAttr_R", (DL_FUNC) &XGBoosterSetAttr_R, 3},
|
||||||
{"XGBoosterSetParam_R", (DL_FUNC) &XGBoosterSetParam_R, 3},
|
{"XGBoosterSetParam_R", (DL_FUNC) &XGBoosterSetParam_R, 3},
|
||||||
{"XGBoosterUpdateOneIter_R", (DL_FUNC) &XGBoosterUpdateOneIter_R, 3},
|
{"XGBoosterUpdateOneIter_R", (DL_FUNC) &XGBoosterUpdateOneIter_R, 3},
|
||||||
{"XGCheckNullPtr_R", (DL_FUNC) &XGCheckNullPtr_R, 1},
|
{"XGCheckNullPtr_R", (DL_FUNC) &XGCheckNullPtr_R, 1},
|
||||||
{"XGDMatrixCreateFromCSC_R", (DL_FUNC) &XGDMatrixCreateFromCSC_R, 4},
|
{"XGDMatrixCreateFromCSC_R", (DL_FUNC) &XGDMatrixCreateFromCSC_R, 6},
|
||||||
|
{"XGDMatrixCreateFromCSR_R", (DL_FUNC) &XGDMatrixCreateFromCSR_R, 6},
|
||||||
{"XGDMatrixCreateFromFile_R", (DL_FUNC) &XGDMatrixCreateFromFile_R, 2},
|
{"XGDMatrixCreateFromFile_R", (DL_FUNC) &XGDMatrixCreateFromFile_R, 2},
|
||||||
{"XGDMatrixCreateFromMat_R", (DL_FUNC) &XGDMatrixCreateFromMat_R, 2},
|
{"XGDMatrixCreateFromMat_R", (DL_FUNC) &XGDMatrixCreateFromMat_R, 3},
|
||||||
{"XGDMatrixGetInfo_R", (DL_FUNC) &XGDMatrixGetInfo_R, 2},
|
{"XGDMatrixGetInfo_R", (DL_FUNC) &XGDMatrixGetInfo_R, 2},
|
||||||
|
{"XGDMatrixGetStrFeatureInfo_R", (DL_FUNC) &XGDMatrixGetStrFeatureInfo_R, 2},
|
||||||
{"XGDMatrixNumCol_R", (DL_FUNC) &XGDMatrixNumCol_R, 1},
|
{"XGDMatrixNumCol_R", (DL_FUNC) &XGDMatrixNumCol_R, 1},
|
||||||
{"XGDMatrixNumRow_R", (DL_FUNC) &XGDMatrixNumRow_R, 1},
|
{"XGDMatrixNumRow_R", (DL_FUNC) &XGDMatrixNumRow_R, 1},
|
||||||
{"XGDMatrixSaveBinary_R", (DL_FUNC) &XGDMatrixSaveBinary_R, 3},
|
{"XGDMatrixSaveBinary_R", (DL_FUNC) &XGDMatrixSaveBinary_R, 3},
|
||||||
{"XGDMatrixSetInfo_R", (DL_FUNC) &XGDMatrixSetInfo_R, 3},
|
{"XGDMatrixSetInfo_R", (DL_FUNC) &XGDMatrixSetInfo_R, 3},
|
||||||
|
{"XGDMatrixSetStrFeatureInfo_R", (DL_FUNC) &XGDMatrixSetStrFeatureInfo_R, 3},
|
||||||
{"XGDMatrixSliceDMatrix_R", (DL_FUNC) &XGDMatrixSliceDMatrix_R, 2},
|
{"XGDMatrixSliceDMatrix_R", (DL_FUNC) &XGDMatrixSliceDMatrix_R, 2},
|
||||||
{"XGBSetGlobalConfig_R", (DL_FUNC) &XGBSetGlobalConfig_R, 1},
|
{"XGBSetGlobalConfig_R", (DL_FUNC) &XGBSetGlobalConfig_R, 1},
|
||||||
{"XGBGetGlobalConfig_R", (DL_FUNC) &XGBGetGlobalConfig_R, 0},
|
{"XGBGetGlobalConfig_R", (DL_FUNC) &XGBGetGlobalConfig_R, 0},
|
||||||
|
{"XGBoosterFeatureScore_R", (DL_FUNC) &XGBoosterFeatureScore_R, 2},
|
||||||
{NULL, NULL, 0}
|
{NULL, NULL, 0}
|
||||||
};
|
};
|
||||||
|
|
||||||
#if defined(_WIN32)
|
#if defined(_WIN32)
|
||||||
__declspec(dllexport)
|
__declspec(dllexport)
|
||||||
#endif // defined(_WIN32)
|
#endif // defined(_WIN32)
|
||||||
void R_init_xgboost(DllInfo *dll) {
|
void attribute_visible R_init_xgboost(DllInfo *dll) {
|
||||||
R_registerRoutines(dll, NULL, CallEntries, NULL, NULL);
|
R_registerRoutines(dll, NULL, CallEntries, NULL, NULL);
|
||||||
R_useDynamicSymbols(dll, FALSE);
|
R_useDynamicSymbols(dll, FALSE);
|
||||||
}
|
}
|
||||||
|
|||||||
3
R-package/src/xgboost-win.def
Normal file
3
R-package/src/xgboost-win.def
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
LIBRARY xgboost.dll
|
||||||
|
EXPORTS
|
||||||
|
R_init_xgboost
|
||||||
@@ -1,15 +1,26 @@
|
|||||||
// Copyright (c) 2014 by Contributors
|
/**
|
||||||
#include <dmlc/logging.h>
|
* Copyright 2014-2023 by XGBoost Contributors
|
||||||
#include <dmlc/omp.h>
|
*/
|
||||||
#include <dmlc/common.h>
|
#include <dmlc/common.h>
|
||||||
|
#include <dmlc/omp.h>
|
||||||
#include <xgboost/c_api.h>
|
#include <xgboost/c_api.h>
|
||||||
#include <vector>
|
#include <xgboost/context.h>
|
||||||
|
#include <xgboost/data.h>
|
||||||
|
#include <xgboost/logging.h>
|
||||||
|
|
||||||
|
#include <cstdio>
|
||||||
|
#include <cstring>
|
||||||
|
#include <sstream>
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <utility>
|
#include <utility>
|
||||||
#include <cstring>
|
#include <vector>
|
||||||
#include <cstdio>
|
|
||||||
#include <sstream>
|
#include "../../src/c_api/c_api_error.h"
|
||||||
#include "./xgboost_R.h"
|
#include "../../src/c_api/c_api_utils.h" // MakeSparseFromPtr
|
||||||
|
#include "../../src/common/threading_utils.h"
|
||||||
|
|
||||||
|
#include "./xgboost_R.h" // Must follow other includes.
|
||||||
|
#include "Rinternals.h"
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief macro to annotate begin of api
|
* \brief macro to annotate begin of api
|
||||||
@@ -35,14 +46,27 @@
|
|||||||
error(XGBGetLastError()); \
|
error(XGBGetLastError()); \
|
||||||
}
|
}
|
||||||
|
|
||||||
|
using dmlc::BeginPtr;
|
||||||
|
|
||||||
using namespace dmlc;
|
xgboost::Context const *BoosterCtx(BoosterHandle handle) {
|
||||||
|
CHECK_HANDLE();
|
||||||
|
auto *learner = static_cast<xgboost::Learner *>(handle);
|
||||||
|
CHECK(learner);
|
||||||
|
return learner->Ctx();
|
||||||
|
}
|
||||||
|
|
||||||
SEXP XGCheckNullPtr_R(SEXP handle) {
|
xgboost::Context const *DMatrixCtx(DMatrixHandle handle) {
|
||||||
|
CHECK_HANDLE();
|
||||||
|
auto p_m = static_cast<std::shared_ptr<xgboost::DMatrix> *>(handle);
|
||||||
|
CHECK(p_m);
|
||||||
|
return p_m->get()->Ctx();
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGCheckNullPtr_R(SEXP handle) {
|
||||||
return ScalarLogical(R_ExternalPtrAddr(handle) == NULL);
|
return ScalarLogical(R_ExternalPtrAddr(handle) == NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
void _DMatrixFinalizer(SEXP ext) {
|
XGB_DLL void _DMatrixFinalizer(SEXP ext) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
if (R_ExternalPtrAddr(ext) == NULL) return;
|
if (R_ExternalPtrAddr(ext) == NULL) return;
|
||||||
CHECK_CALL(XGDMatrixFree(R_ExternalPtrAddr(ext)));
|
CHECK_CALL(XGDMatrixFree(R_ExternalPtrAddr(ext)));
|
||||||
@@ -50,14 +74,14 @@ void _DMatrixFinalizer(SEXP ext) {
|
|||||||
R_API_END();
|
R_API_END();
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBSetGlobalConfig_R(SEXP json_str) {
|
XGB_DLL SEXP XGBSetGlobalConfig_R(SEXP json_str) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBSetGlobalConfig(CHAR(asChar(json_str))));
|
CHECK_CALL(XGBSetGlobalConfig(CHAR(asChar(json_str))));
|
||||||
R_API_END();
|
R_API_END();
|
||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBGetGlobalConfig_R() {
|
XGB_DLL SEXP XGBGetGlobalConfig_R() {
|
||||||
const char* json_str;
|
const char* json_str;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBGetGlobalConfig(&json_str));
|
CHECK_CALL(XGBGetGlobalConfig(&json_str));
|
||||||
@@ -65,7 +89,7 @@ SEXP XGBGetGlobalConfig_R() {
|
|||||||
return mkString(json_str);
|
return mkString(json_str);
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixCreateFromFile_R(SEXP fname, SEXP silent) {
|
XGB_DLL SEXP XGDMatrixCreateFromFile_R(SEXP fname, SEXP silent) {
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
DMatrixHandle handle;
|
DMatrixHandle handle;
|
||||||
@@ -77,8 +101,7 @@ SEXP XGDMatrixCreateFromFile_R(SEXP fname, SEXP silent) {
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixCreateFromMat_R(SEXP mat,
|
XGB_DLL SEXP XGDMatrixCreateFromMat_R(SEXP mat, SEXP missing, SEXP n_threads) {
|
||||||
SEXP missing) {
|
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
SEXP dim = getAttrib(mat, R_DimSymbol);
|
SEXP dim = getAttrib(mat, R_DimSymbol);
|
||||||
@@ -93,18 +116,32 @@ SEXP XGDMatrixCreateFromMat_R(SEXP mat,
|
|||||||
din = REAL(mat);
|
din = REAL(mat);
|
||||||
}
|
}
|
||||||
std::vector<float> data(nrow * ncol);
|
std::vector<float> data(nrow * ncol);
|
||||||
dmlc::OMPException exc;
|
xgboost::Context ctx;
|
||||||
#pragma omp parallel for schedule(static)
|
ctx.nthread = asInteger(n_threads);
|
||||||
for (omp_ulong i = 0; i < nrow; ++i) {
|
std::int32_t threads = ctx.Threads();
|
||||||
exc.Run([&]() {
|
|
||||||
|
if (is_int) {
|
||||||
|
xgboost::common::ParallelFor(nrow, threads, [&](xgboost::omp_ulong i) {
|
||||||
for (size_t j = 0; j < ncol; ++j) {
|
for (size_t j = 0; j < ncol; ++j) {
|
||||||
data[i * ncol +j] = is_int ? static_cast<float>(iin[i + nrow * j]) : din[i + nrow * j];
|
auto v = iin[i + nrow * j];
|
||||||
|
if (v == NA_INTEGER) {
|
||||||
|
data[i * ncol + j] = std::numeric_limits<float>::quiet_NaN();
|
||||||
|
} else {
|
||||||
|
data[i * ncol + j] = static_cast<float>(v);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
xgboost::common::ParallelFor(nrow, threads, [&](xgboost::omp_ulong i) {
|
||||||
|
for (size_t j = 0; j < ncol; ++j) {
|
||||||
|
data[i * ncol + j] = din[i + nrow * j];
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
exc.Rethrow();
|
|
||||||
DMatrixHandle handle;
|
DMatrixHandle handle;
|
||||||
CHECK_CALL(XGDMatrixCreateFromMat(BeginPtr(data), nrow, ncol, asReal(missing), &handle));
|
CHECK_CALL(XGDMatrixCreateFromMat_omp(BeginPtr(data), nrow, ncol,
|
||||||
|
asReal(missing), &handle, threads));
|
||||||
ret = PROTECT(R_MakeExternalPtr(handle, R_NilValue, R_NilValue));
|
ret = PROTECT(R_MakeExternalPtr(handle, R_NilValue, R_NilValue));
|
||||||
R_RegisterCFinalizerEx(ret, _DMatrixFinalizer, TRUE);
|
R_RegisterCFinalizerEx(ret, _DMatrixFinalizer, TRUE);
|
||||||
R_API_END();
|
R_API_END();
|
||||||
@@ -112,46 +149,85 @@ SEXP XGDMatrixCreateFromMat_R(SEXP mat,
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixCreateFromCSC_R(SEXP indptr,
|
namespace {
|
||||||
SEXP indices,
|
void CreateFromSparse(SEXP indptr, SEXP indices, SEXP data, std::string *indptr_str,
|
||||||
SEXP data,
|
std::string *indices_str, std::string *data_str) {
|
||||||
SEXP num_row) {
|
|
||||||
SEXP ret;
|
|
||||||
R_API_BEGIN();
|
|
||||||
const int *p_indptr = INTEGER(indptr);
|
const int *p_indptr = INTEGER(indptr);
|
||||||
const int *p_indices = INTEGER(indices);
|
const int *p_indices = INTEGER(indices);
|
||||||
const double *p_data = REAL(data);
|
const double *p_data = REAL(data);
|
||||||
size_t nindptr = static_cast<size_t>(length(indptr));
|
|
||||||
size_t ndata = static_cast<size_t>(length(data));
|
|
||||||
size_t nrow = static_cast<size_t>(INTEGER(num_row)[0]);
|
|
||||||
std::vector<size_t> col_ptr_(nindptr);
|
|
||||||
std::vector<unsigned> indices_(ndata);
|
|
||||||
std::vector<float> data_(ndata);
|
|
||||||
|
|
||||||
for (size_t i = 0; i < nindptr; ++i) {
|
auto nindptr = static_cast<std::size_t>(length(indptr));
|
||||||
col_ptr_[i] = static_cast<size_t>(p_indptr[i]);
|
auto ndata = static_cast<std::size_t>(length(data));
|
||||||
|
CHECK_EQ(ndata, p_indptr[nindptr - 1]);
|
||||||
|
xgboost::detail::MakeSparseFromPtr(p_indptr, p_indices, p_data, nindptr, indptr_str, indices_str,
|
||||||
|
data_str);
|
||||||
}
|
}
|
||||||
dmlc::OMPException exc;
|
} // namespace
|
||||||
#pragma omp parallel for schedule(static)
|
|
||||||
for (int64_t i = 0; i < static_cast<int64_t>(ndata); ++i) {
|
XGB_DLL SEXP XGDMatrixCreateFromCSC_R(SEXP indptr, SEXP indices, SEXP data, SEXP num_row,
|
||||||
exc.Run([&]() {
|
SEXP missing, SEXP n_threads) {
|
||||||
indices_[i] = static_cast<unsigned>(p_indices[i]);
|
SEXP ret;
|
||||||
data_[i] = static_cast<float>(p_data[i]);
|
R_API_BEGIN();
|
||||||
});
|
std::int32_t threads = asInteger(n_threads);
|
||||||
}
|
|
||||||
exc.Rethrow();
|
using xgboost::Integer;
|
||||||
|
using xgboost::Json;
|
||||||
|
using xgboost::Object;
|
||||||
|
|
||||||
|
std::string sindptr, sindices, sdata;
|
||||||
|
CreateFromSparse(indptr, indices, data, &sindptr, &sindices, &sdata);
|
||||||
|
auto nrow = static_cast<std::size_t>(INTEGER(num_row)[0]);
|
||||||
|
|
||||||
DMatrixHandle handle;
|
DMatrixHandle handle;
|
||||||
CHECK_CALL(XGDMatrixCreateFromCSCEx(BeginPtr(col_ptr_), BeginPtr(indices_),
|
Json jconfig{Object{}};
|
||||||
BeginPtr(data_), nindptr, ndata,
|
// Construct configuration
|
||||||
nrow, &handle));
|
jconfig["nthread"] = Integer{threads};
|
||||||
|
jconfig["missing"] = xgboost::Number{asReal(missing)};
|
||||||
|
std::string config;
|
||||||
|
Json::Dump(jconfig, &config);
|
||||||
|
CHECK_CALL(XGDMatrixCreateFromCSC(sindptr.c_str(), sindices.c_str(), sdata.c_str(), nrow,
|
||||||
|
config.c_str(), &handle));
|
||||||
|
|
||||||
ret = PROTECT(R_MakeExternalPtr(handle, R_NilValue, R_NilValue));
|
ret = PROTECT(R_MakeExternalPtr(handle, R_NilValue, R_NilValue));
|
||||||
|
|
||||||
R_RegisterCFinalizerEx(ret, _DMatrixFinalizer, TRUE);
|
R_RegisterCFinalizerEx(ret, _DMatrixFinalizer, TRUE);
|
||||||
R_API_END();
|
R_API_END();
|
||||||
UNPROTECT(1);
|
UNPROTECT(1);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixSliceDMatrix_R(SEXP handle, SEXP idxset) {
|
XGB_DLL SEXP XGDMatrixCreateFromCSR_R(SEXP indptr, SEXP indices, SEXP data, SEXP num_col,
|
||||||
|
SEXP missing, SEXP n_threads) {
|
||||||
|
SEXP ret;
|
||||||
|
R_API_BEGIN();
|
||||||
|
std::int32_t threads = asInteger(n_threads);
|
||||||
|
|
||||||
|
using xgboost::Integer;
|
||||||
|
using xgboost::Json;
|
||||||
|
using xgboost::Object;
|
||||||
|
|
||||||
|
std::string sindptr, sindices, sdata;
|
||||||
|
CreateFromSparse(indptr, indices, data, &sindptr, &sindices, &sdata);
|
||||||
|
auto ncol = static_cast<std::size_t>(INTEGER(num_col)[0]);
|
||||||
|
|
||||||
|
DMatrixHandle handle;
|
||||||
|
Json jconfig{Object{}};
|
||||||
|
// Construct configuration
|
||||||
|
jconfig["nthread"] = Integer{threads};
|
||||||
|
jconfig["missing"] = xgboost::Number{asReal(missing)};
|
||||||
|
std::string config;
|
||||||
|
Json::Dump(jconfig, &config);
|
||||||
|
CHECK_CALL(XGDMatrixCreateFromCSR(sindptr.c_str(), sindices.c_str(), sdata.c_str(), ncol,
|
||||||
|
config.c_str(), &handle));
|
||||||
|
ret = PROTECT(R_MakeExternalPtr(handle, R_NilValue, R_NilValue));
|
||||||
|
|
||||||
|
R_RegisterCFinalizerEx(ret, _DMatrixFinalizer, TRUE);
|
||||||
|
R_API_END();
|
||||||
|
UNPROTECT(1);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGDMatrixSliceDMatrix_R(SEXP handle, SEXP idxset) {
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
int len = length(idxset);
|
int len = length(idxset);
|
||||||
@@ -171,7 +247,7 @@ SEXP XGDMatrixSliceDMatrix_R(SEXP handle, SEXP idxset) {
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixSaveBinary_R(SEXP handle, SEXP fname, SEXP silent) {
|
XGB_DLL SEXP XGDMatrixSaveBinary_R(SEXP handle, SEXP fname, SEXP silent) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGDMatrixSaveBinary(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGDMatrixSaveBinary(R_ExternalPtrAddr(handle),
|
||||||
CHAR(asChar(fname)),
|
CHAR(asChar(fname)),
|
||||||
@@ -180,49 +256,76 @@ SEXP XGDMatrixSaveBinary_R(SEXP handle, SEXP fname, SEXP silent) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixSetInfo_R(SEXP handle, SEXP field, SEXP array) {
|
XGB_DLL SEXP XGDMatrixSetInfo_R(SEXP handle, SEXP field, SEXP array) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
int len = length(array);
|
int len = length(array);
|
||||||
const char *name = CHAR(asChar(field));
|
const char *name = CHAR(asChar(field));
|
||||||
dmlc::OMPException exc;
|
auto ctx = DMatrixCtx(R_ExternalPtrAddr(handle));
|
||||||
if (!strcmp("group", name)) {
|
if (!strcmp("group", name)) {
|
||||||
std::vector<unsigned> vec(len);
|
std::vector<unsigned> vec(len);
|
||||||
#pragma omp parallel for schedule(static)
|
xgboost::common::ParallelFor(len, ctx->Threads(), [&](xgboost::omp_ulong i) {
|
||||||
for (int i = 0; i < len; ++i) {
|
|
||||||
exc.Run([&]() {
|
|
||||||
vec[i] = static_cast<unsigned>(INTEGER(array)[i]);
|
vec[i] = static_cast<unsigned>(INTEGER(array)[i]);
|
||||||
});
|
});
|
||||||
}
|
CHECK_CALL(
|
||||||
exc.Rethrow();
|
XGDMatrixSetUIntInfo(R_ExternalPtrAddr(handle), CHAR(asChar(field)), BeginPtr(vec), len));
|
||||||
CHECK_CALL(XGDMatrixSetUIntInfo(R_ExternalPtrAddr(handle),
|
|
||||||
CHAR(asChar(field)),
|
|
||||||
BeginPtr(vec), len));
|
|
||||||
} else {
|
} else {
|
||||||
std::vector<float> vec(len);
|
std::vector<float> vec(len);
|
||||||
#pragma omp parallel for schedule(static)
|
xgboost::common::ParallelFor(len, ctx->Threads(),
|
||||||
for (int i = 0; i < len; ++i) {
|
[&](xgboost::omp_ulong i) { vec[i] = REAL(array)[i]; });
|
||||||
exc.Run([&]() {
|
CHECK_CALL(
|
||||||
vec[i] = REAL(array)[i];
|
XGDMatrixSetFloatInfo(R_ExternalPtrAddr(handle), CHAR(asChar(field)), BeginPtr(vec), len));
|
||||||
});
|
|
||||||
}
|
|
||||||
exc.Rethrow();
|
|
||||||
CHECK_CALL(XGDMatrixSetFloatInfo(R_ExternalPtrAddr(handle),
|
|
||||||
CHAR(asChar(field)),
|
|
||||||
BeginPtr(vec), len));
|
|
||||||
}
|
}
|
||||||
R_API_END();
|
R_API_END();
|
||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixGetInfo_R(SEXP handle, SEXP field) {
|
XGB_DLL SEXP XGDMatrixSetStrFeatureInfo_R(SEXP handle, SEXP field, SEXP array) {
|
||||||
|
R_API_BEGIN();
|
||||||
|
size_t len{0};
|
||||||
|
if (!isNull(array)) {
|
||||||
|
len = length(array);
|
||||||
|
}
|
||||||
|
|
||||||
|
const char *name = CHAR(asChar(field));
|
||||||
|
std::vector<std::string> str_info;
|
||||||
|
for (size_t i = 0; i < len; ++i) {
|
||||||
|
str_info.emplace_back(CHAR(asChar(VECTOR_ELT(array, i))));
|
||||||
|
}
|
||||||
|
std::vector<char const*> vec(len);
|
||||||
|
std::transform(str_info.cbegin(), str_info.cend(), vec.begin(),
|
||||||
|
[](std::string const &str) { return str.c_str(); });
|
||||||
|
CHECK_CALL(XGDMatrixSetStrFeatureInfo(R_ExternalPtrAddr(handle), name, vec.data(), len));
|
||||||
|
R_API_END();
|
||||||
|
return R_NilValue;
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGDMatrixGetStrFeatureInfo_R(SEXP handle, SEXP field) {
|
||||||
|
SEXP ret;
|
||||||
|
R_API_BEGIN();
|
||||||
|
char const **out_features{nullptr};
|
||||||
|
bst_ulong len{0};
|
||||||
|
const char *name = CHAR(asChar(field));
|
||||||
|
XGDMatrixGetStrFeatureInfo(R_ExternalPtrAddr(handle), name, &len, &out_features);
|
||||||
|
|
||||||
|
if (len > 0) {
|
||||||
|
ret = PROTECT(allocVector(STRSXP, len));
|
||||||
|
for (size_t i = 0; i < len; ++i) {
|
||||||
|
SET_STRING_ELT(ret, i, mkChar(out_features[i]));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
ret = PROTECT(R_NilValue);
|
||||||
|
}
|
||||||
|
R_API_END();
|
||||||
|
UNPROTECT(1);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGDMatrixGetInfo_R(SEXP handle, SEXP field) {
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong olen;
|
bst_ulong olen;
|
||||||
const float *res;
|
const float *res;
|
||||||
CHECK_CALL(XGDMatrixGetFloatInfo(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGDMatrixGetFloatInfo(R_ExternalPtrAddr(handle), CHAR(asChar(field)), &olen, &res));
|
||||||
CHAR(asChar(field)),
|
|
||||||
&olen,
|
|
||||||
&res));
|
|
||||||
ret = PROTECT(allocVector(REALSXP, olen));
|
ret = PROTECT(allocVector(REALSXP, olen));
|
||||||
for (size_t i = 0; i < olen; ++i) {
|
for (size_t i = 0; i < olen; ++i) {
|
||||||
REAL(ret)[i] = res[i];
|
REAL(ret)[i] = res[i];
|
||||||
@@ -232,7 +335,7 @@ SEXP XGDMatrixGetInfo_R(SEXP handle, SEXP field) {
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixNumRow_R(SEXP handle) {
|
XGB_DLL SEXP XGDMatrixNumRow_R(SEXP handle) {
|
||||||
bst_ulong nrow;
|
bst_ulong nrow;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGDMatrixNumRow(R_ExternalPtrAddr(handle), &nrow));
|
CHECK_CALL(XGDMatrixNumRow(R_ExternalPtrAddr(handle), &nrow));
|
||||||
@@ -240,7 +343,7 @@ SEXP XGDMatrixNumRow_R(SEXP handle) {
|
|||||||
return ScalarInteger(static_cast<int>(nrow));
|
return ScalarInteger(static_cast<int>(nrow));
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGDMatrixNumCol_R(SEXP handle) {
|
XGB_DLL SEXP XGDMatrixNumCol_R(SEXP handle) {
|
||||||
bst_ulong ncol;
|
bst_ulong ncol;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGDMatrixNumCol(R_ExternalPtrAddr(handle), &ncol));
|
CHECK_CALL(XGDMatrixNumCol(R_ExternalPtrAddr(handle), &ncol));
|
||||||
@@ -255,7 +358,7 @@ void _BoosterFinalizer(SEXP ext) {
|
|||||||
R_ClearExternalPtr(ext);
|
R_ClearExternalPtr(ext);
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterCreate_R(SEXP dmats) {
|
XGB_DLL SEXP XGBoosterCreate_R(SEXP dmats) {
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
int len = length(dmats);
|
int len = length(dmats);
|
||||||
@@ -272,7 +375,22 @@ SEXP XGBoosterCreate_R(SEXP dmats) {
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterSetParam_R(SEXP handle, SEXP name, SEXP val) {
|
XGB_DLL SEXP XGBoosterCreateInEmptyObj_R(SEXP dmats, SEXP R_handle) {
|
||||||
|
R_API_BEGIN();
|
||||||
|
int len = length(dmats);
|
||||||
|
std::vector<void*> dvec;
|
||||||
|
for (int i = 0; i < len; ++i) {
|
||||||
|
dvec.push_back(R_ExternalPtrAddr(VECTOR_ELT(dmats, i)));
|
||||||
|
}
|
||||||
|
BoosterHandle handle;
|
||||||
|
CHECK_CALL(XGBoosterCreate(BeginPtr(dvec), dvec.size(), &handle));
|
||||||
|
R_SetExternalPtrAddr(R_handle, handle);
|
||||||
|
R_RegisterCFinalizerEx(R_handle, _BoosterFinalizer, TRUE);
|
||||||
|
R_API_END();
|
||||||
|
return R_NilValue;
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGBoosterSetParam_R(SEXP handle, SEXP name, SEXP val) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterSetParam(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterSetParam(R_ExternalPtrAddr(handle),
|
||||||
CHAR(asChar(name)),
|
CHAR(asChar(name)),
|
||||||
@@ -281,7 +399,7 @@ SEXP XGBoosterSetParam_R(SEXP handle, SEXP name, SEXP val) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterUpdateOneIter_R(SEXP handle, SEXP iter, SEXP dtrain) {
|
XGB_DLL SEXP XGBoosterUpdateOneIter_R(SEXP handle, SEXP iter, SEXP dtrain) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterUpdateOneIter(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterUpdateOneIter(R_ExternalPtrAddr(handle),
|
||||||
asInteger(iter),
|
asInteger(iter),
|
||||||
@@ -290,21 +408,17 @@ SEXP XGBoosterUpdateOneIter_R(SEXP handle, SEXP iter, SEXP dtrain) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterBoostOneIter_R(SEXP handle, SEXP dtrain, SEXP grad, SEXP hess) {
|
XGB_DLL SEXP XGBoosterBoostOneIter_R(SEXP handle, SEXP dtrain, SEXP grad, SEXP hess) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_EQ(length(grad), length(hess))
|
CHECK_EQ(length(grad), length(hess))
|
||||||
<< "gradient and hess must have same length";
|
<< "gradient and hess must have same length";
|
||||||
int len = length(grad);
|
int len = length(grad);
|
||||||
std::vector<float> tgrad(len), thess(len);
|
std::vector<float> tgrad(len), thess(len);
|
||||||
dmlc::OMPException exc;
|
auto ctx = BoosterCtx(R_ExternalPtrAddr(handle));
|
||||||
#pragma omp parallel for schedule(static)
|
xgboost::common::ParallelFor(len, ctx->Threads(), [&](xgboost::omp_ulong j) {
|
||||||
for (int j = 0; j < len; ++j) {
|
|
||||||
exc.Run([&]() {
|
|
||||||
tgrad[j] = REAL(grad)[j];
|
tgrad[j] = REAL(grad)[j];
|
||||||
thess[j] = REAL(hess)[j];
|
thess[j] = REAL(hess)[j];
|
||||||
});
|
});
|
||||||
}
|
|
||||||
exc.Rethrow();
|
|
||||||
CHECK_CALL(XGBoosterBoostOneIter(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterBoostOneIter(R_ExternalPtrAddr(handle),
|
||||||
R_ExternalPtrAddr(dtrain),
|
R_ExternalPtrAddr(dtrain),
|
||||||
BeginPtr(tgrad), BeginPtr(thess),
|
BeginPtr(tgrad), BeginPtr(thess),
|
||||||
@@ -313,7 +427,7 @@ SEXP XGBoosterBoostOneIter_R(SEXP handle, SEXP dtrain, SEXP grad, SEXP hess) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterEvalOneIter_R(SEXP handle, SEXP iter, SEXP dmats, SEXP evnames) {
|
XGB_DLL SEXP XGBoosterEvalOneIter_R(SEXP handle, SEXP iter, SEXP dmats, SEXP evnames) {
|
||||||
const char *ret;
|
const char *ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_EQ(length(dmats), length(evnames))
|
CHECK_EQ(length(dmats), length(evnames))
|
||||||
@@ -338,57 +452,59 @@ SEXP XGBoosterEvalOneIter_R(SEXP handle, SEXP iter, SEXP dmats, SEXP evnames) {
|
|||||||
return mkString(ret);
|
return mkString(ret);
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterPredict_R(SEXP handle, SEXP dmat, SEXP option_mask,
|
XGB_DLL SEXP XGBoosterPredictFromDMatrix_R(SEXP handle, SEXP dmat, SEXP json_config) {
|
||||||
SEXP ntree_limit, SEXP training) {
|
SEXP r_out_shape;
|
||||||
SEXP ret;
|
SEXP r_out_result;
|
||||||
|
SEXP r_out;
|
||||||
|
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong olen;
|
char const *c_json_config = CHAR(asChar(json_config));
|
||||||
const float *res;
|
|
||||||
CHECK_CALL(XGBoosterPredict(R_ExternalPtrAddr(handle),
|
bst_ulong out_dim;
|
||||||
R_ExternalPtrAddr(dmat),
|
bst_ulong const *out_shape;
|
||||||
asInteger(option_mask),
|
float const *out_result;
|
||||||
asInteger(ntree_limit),
|
CHECK_CALL(XGBoosterPredictFromDMatrix(R_ExternalPtrAddr(handle),
|
||||||
asInteger(training),
|
R_ExternalPtrAddr(dmat), c_json_config,
|
||||||
&olen, &res));
|
&out_shape, &out_dim, &out_result));
|
||||||
ret = PROTECT(allocVector(REALSXP, olen));
|
|
||||||
for (size_t i = 0; i < olen; ++i) {
|
r_out_shape = PROTECT(allocVector(INTSXP, out_dim));
|
||||||
REAL(ret)[i] = res[i];
|
size_t len = 1;
|
||||||
|
for (size_t i = 0; i < out_dim; ++i) {
|
||||||
|
INTEGER(r_out_shape)[i] = out_shape[i];
|
||||||
|
len *= out_shape[i];
|
||||||
}
|
}
|
||||||
|
r_out_result = PROTECT(allocVector(REALSXP, len));
|
||||||
|
auto ctx = BoosterCtx(R_ExternalPtrAddr(handle));
|
||||||
|
xgboost::common::ParallelFor(len, ctx->Threads(), [&](xgboost::omp_ulong i) {
|
||||||
|
REAL(r_out_result)[i] = out_result[i];
|
||||||
|
});
|
||||||
|
|
||||||
|
r_out = PROTECT(allocVector(VECSXP, 2));
|
||||||
|
|
||||||
|
SET_VECTOR_ELT(r_out, 0, r_out_shape);
|
||||||
|
SET_VECTOR_ELT(r_out, 1, r_out_result);
|
||||||
|
|
||||||
R_API_END();
|
R_API_END();
|
||||||
UNPROTECT(1);
|
UNPROTECT(3);
|
||||||
return ret;
|
|
||||||
|
return r_out;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterLoadModel_R(SEXP handle, SEXP fname) {
|
XGB_DLL SEXP XGBoosterLoadModel_R(SEXP handle, SEXP fname) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterLoadModel(R_ExternalPtrAddr(handle), CHAR(asChar(fname))));
|
CHECK_CALL(XGBoosterLoadModel(R_ExternalPtrAddr(handle), CHAR(asChar(fname))));
|
||||||
R_API_END();
|
R_API_END();
|
||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterSaveModel_R(SEXP handle, SEXP fname) {
|
XGB_DLL SEXP XGBoosterSaveModel_R(SEXP handle, SEXP fname) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterSaveModel(R_ExternalPtrAddr(handle), CHAR(asChar(fname))));
|
CHECK_CALL(XGBoosterSaveModel(R_ExternalPtrAddr(handle), CHAR(asChar(fname))));
|
||||||
R_API_END();
|
R_API_END();
|
||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterModelToRaw_R(SEXP handle) {
|
XGB_DLL SEXP XGBoosterLoadModelFromRaw_R(SEXP handle, SEXP raw) {
|
||||||
SEXP ret;
|
|
||||||
R_API_BEGIN();
|
|
||||||
bst_ulong olen;
|
|
||||||
const char *raw;
|
|
||||||
CHECK_CALL(XGBoosterGetModelRaw(R_ExternalPtrAddr(handle), &olen, &raw));
|
|
||||||
ret = PROTECT(allocVector(RAWSXP, olen));
|
|
||||||
if (olen != 0) {
|
|
||||||
memcpy(RAW(ret), raw, olen);
|
|
||||||
}
|
|
||||||
R_API_END();
|
|
||||||
UNPROTECT(1);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
SEXP XGBoosterLoadModelFromRaw_R(SEXP handle, SEXP raw) {
|
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterLoadModelFromBuffer(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterLoadModelFromBuffer(R_ExternalPtrAddr(handle),
|
||||||
RAW(raw),
|
RAW(raw),
|
||||||
@@ -397,7 +513,23 @@ SEXP XGBoosterLoadModelFromRaw_R(SEXP handle, SEXP raw) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterSaveJsonConfig_R(SEXP handle) {
|
XGB_DLL SEXP XGBoosterSaveModelToRaw_R(SEXP handle, SEXP json_config) {
|
||||||
|
SEXP ret;
|
||||||
|
R_API_BEGIN();
|
||||||
|
bst_ulong olen;
|
||||||
|
char const *c_json_config = CHAR(asChar(json_config));
|
||||||
|
char const *raw;
|
||||||
|
CHECK_CALL(XGBoosterSaveModelToBuffer(R_ExternalPtrAddr(handle), c_json_config, &olen, &raw))
|
||||||
|
ret = PROTECT(allocVector(RAWSXP, olen));
|
||||||
|
if (olen != 0) {
|
||||||
|
std::memcpy(RAW(ret), raw, olen);
|
||||||
|
}
|
||||||
|
R_API_END();
|
||||||
|
UNPROTECT(1);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGBoosterSaveJsonConfig_R(SEXP handle) {
|
||||||
const char* ret;
|
const char* ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong len {0};
|
bst_ulong len {0};
|
||||||
@@ -408,14 +540,14 @@ SEXP XGBoosterSaveJsonConfig_R(SEXP handle) {
|
|||||||
return mkString(ret);
|
return mkString(ret);
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterLoadJsonConfig_R(SEXP handle, SEXP value) {
|
XGB_DLL SEXP XGBoosterLoadJsonConfig_R(SEXP handle, SEXP value) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterLoadJsonConfig(R_ExternalPtrAddr(handle), CHAR(asChar(value))));
|
CHECK_CALL(XGBoosterLoadJsonConfig(R_ExternalPtrAddr(handle), CHAR(asChar(value))));
|
||||||
R_API_END();
|
R_API_END();
|
||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterSerializeToBuffer_R(SEXP handle) {
|
XGB_DLL SEXP XGBoosterSerializeToBuffer_R(SEXP handle) {
|
||||||
SEXP ret;
|
SEXP ret;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong out_len;
|
bst_ulong out_len;
|
||||||
@@ -430,7 +562,7 @@ SEXP XGBoosterSerializeToBuffer_R(SEXP handle) {
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterUnserializeFromBuffer_R(SEXP handle, SEXP raw) {
|
XGB_DLL SEXP XGBoosterUnserializeFromBuffer_R(SEXP handle, SEXP raw) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
CHECK_CALL(XGBoosterUnserializeFromBuffer(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterUnserializeFromBuffer(R_ExternalPtrAddr(handle),
|
||||||
RAW(raw),
|
RAW(raw),
|
||||||
@@ -439,7 +571,7 @@ SEXP XGBoosterUnserializeFromBuffer_R(SEXP handle, SEXP raw) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterDumpModel_R(SEXP handle, SEXP fmap, SEXP with_stats, SEXP dump_format) {
|
XGB_DLL SEXP XGBoosterDumpModel_R(SEXP handle, SEXP fmap, SEXP with_stats, SEXP dump_format) {
|
||||||
SEXP out;
|
SEXP out;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong olen;
|
bst_ulong olen;
|
||||||
@@ -476,7 +608,7 @@ SEXP XGBoosterDumpModel_R(SEXP handle, SEXP fmap, SEXP with_stats, SEXP dump_for
|
|||||||
return out;
|
return out;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterGetAttr_R(SEXP handle, SEXP name) {
|
XGB_DLL SEXP XGBoosterGetAttr_R(SEXP handle, SEXP name) {
|
||||||
SEXP out;
|
SEXP out;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
int success;
|
int success;
|
||||||
@@ -496,7 +628,7 @@ SEXP XGBoosterGetAttr_R(SEXP handle, SEXP name) {
|
|||||||
return out;
|
return out;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterSetAttr_R(SEXP handle, SEXP name, SEXP val) {
|
XGB_DLL SEXP XGBoosterSetAttr_R(SEXP handle, SEXP name, SEXP val) {
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
const char *v = isNull(val) ? nullptr : CHAR(asChar(val));
|
const char *v = isNull(val) ? nullptr : CHAR(asChar(val));
|
||||||
CHECK_CALL(XGBoosterSetAttr(R_ExternalPtrAddr(handle),
|
CHECK_CALL(XGBoosterSetAttr(R_ExternalPtrAddr(handle),
|
||||||
@@ -505,7 +637,7 @@ SEXP XGBoosterSetAttr_R(SEXP handle, SEXP name, SEXP val) {
|
|||||||
return R_NilValue;
|
return R_NilValue;
|
||||||
}
|
}
|
||||||
|
|
||||||
SEXP XGBoosterGetAttrNames_R(SEXP handle) {
|
XGB_DLL SEXP XGBoosterGetAttrNames_R(SEXP handle) {
|
||||||
SEXP out;
|
SEXP out;
|
||||||
R_API_BEGIN();
|
R_API_BEGIN();
|
||||||
bst_ulong len;
|
bst_ulong len;
|
||||||
@@ -524,3 +656,50 @@ SEXP XGBoosterGetAttrNames_R(SEXP handle) {
|
|||||||
UNPROTECT(1);
|
UNPROTECT(1);
|
||||||
return out;
|
return out;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
XGB_DLL SEXP XGBoosterFeatureScore_R(SEXP handle, SEXP json_config) {
|
||||||
|
SEXP out_features_sexp;
|
||||||
|
SEXP out_scores_sexp;
|
||||||
|
SEXP out_shape_sexp;
|
||||||
|
SEXP r_out;
|
||||||
|
|
||||||
|
R_API_BEGIN();
|
||||||
|
char const *c_json_config = CHAR(asChar(json_config));
|
||||||
|
bst_ulong out_n_features;
|
||||||
|
char const **out_features;
|
||||||
|
|
||||||
|
bst_ulong out_dim;
|
||||||
|
bst_ulong const *out_shape;
|
||||||
|
float const *out_scores;
|
||||||
|
|
||||||
|
CHECK_CALL(XGBoosterFeatureScore(R_ExternalPtrAddr(handle), c_json_config,
|
||||||
|
&out_n_features, &out_features,
|
||||||
|
&out_dim, &out_shape, &out_scores));
|
||||||
|
out_shape_sexp = PROTECT(allocVector(INTSXP, out_dim));
|
||||||
|
size_t len = 1;
|
||||||
|
for (size_t i = 0; i < out_dim; ++i) {
|
||||||
|
INTEGER(out_shape_sexp)[i] = out_shape[i];
|
||||||
|
len *= out_shape[i];
|
||||||
|
}
|
||||||
|
|
||||||
|
out_scores_sexp = PROTECT(allocVector(REALSXP, len));
|
||||||
|
auto ctx = BoosterCtx(R_ExternalPtrAddr(handle));
|
||||||
|
xgboost::common::ParallelFor(len, ctx->Threads(), [&](xgboost::omp_ulong i) {
|
||||||
|
REAL(out_scores_sexp)[i] = out_scores[i];
|
||||||
|
});
|
||||||
|
|
||||||
|
out_features_sexp = PROTECT(allocVector(STRSXP, out_n_features));
|
||||||
|
for (size_t i = 0; i < out_n_features; ++i) {
|
||||||
|
SET_STRING_ELT(out_features_sexp, i, mkChar(out_features[i]));
|
||||||
|
}
|
||||||
|
|
||||||
|
r_out = PROTECT(allocVector(VECSXP, 3));
|
||||||
|
SET_VECTOR_ELT(r_out, 0, out_features_sexp);
|
||||||
|
SET_VECTOR_ELT(r_out, 1, out_shape_sexp);
|
||||||
|
SET_VECTOR_ELT(r_out, 2, out_scores_sexp);
|
||||||
|
|
||||||
|
R_API_END();
|
||||||
|
UNPROTECT(4);
|
||||||
|
|
||||||
|
return r_out;
|
||||||
|
}
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
/*!
|
/*!
|
||||||
* Copyright 2014 (c) by Contributors
|
* Copyright 2014-2022 by XGBoost Contributors
|
||||||
* \file xgboost_R.h
|
* \file xgboost_R.h
|
||||||
* \author Tianqi Chen
|
* \author Tianqi Chen
|
||||||
* \brief R wrapper of xgboost
|
* \brief R wrapper of xgboost
|
||||||
@@ -47,22 +47,37 @@ XGB_DLL SEXP XGDMatrixCreateFromFile_R(SEXP fname, SEXP silent);
|
|||||||
* This assumes the matrix is stored in column major format
|
* This assumes the matrix is stored in column major format
|
||||||
* \param data R Matrix object
|
* \param data R Matrix object
|
||||||
* \param missing which value to represent missing value
|
* \param missing which value to represent missing value
|
||||||
|
* \param n_threads Number of threads used to construct DMatrix from dense matrix.
|
||||||
* \return created dmatrix
|
* \return created dmatrix
|
||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGDMatrixCreateFromMat_R(SEXP mat,
|
XGB_DLL SEXP XGDMatrixCreateFromMat_R(SEXP mat,
|
||||||
SEXP missing);
|
SEXP missing,
|
||||||
|
SEXP n_threads);
|
||||||
/*!
|
/*!
|
||||||
* \brief create a matrix content from CSC format
|
* \brief create a matrix content from CSC format
|
||||||
* \param indptr pointer to column headers
|
* \param indptr pointer to column headers
|
||||||
* \param indices row indices
|
* \param indices row indices
|
||||||
* \param data content of the data
|
* \param data content of the data
|
||||||
* \param num_row numer of rows (when it's set to 0, then guess from data)
|
* \param num_row numer of rows (when it's set to 0, then guess from data)
|
||||||
|
* \param missing which value to represent missing value
|
||||||
|
* \param n_threads Number of threads used to construct DMatrix from csc matrix.
|
||||||
* \return created dmatrix
|
* \return created dmatrix
|
||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGDMatrixCreateFromCSC_R(SEXP indptr,
|
XGB_DLL SEXP XGDMatrixCreateFromCSC_R(SEXP indptr, SEXP indices, SEXP data, SEXP num_row,
|
||||||
SEXP indices,
|
SEXP missing, SEXP n_threads);
|
||||||
SEXP data,
|
|
||||||
SEXP num_row);
|
/*!
|
||||||
|
* \brief create a matrix content from CSR format
|
||||||
|
* \param indptr pointer to row headers
|
||||||
|
* \param indices column indices
|
||||||
|
* \param data content of the data
|
||||||
|
* \param num_col numer of columns (when it's set to 0, then guess from data)
|
||||||
|
* \param missing which value to represent missing value
|
||||||
|
* \param n_threads Number of threads used to construct DMatrix from csr matrix.
|
||||||
|
* \return created dmatrix
|
||||||
|
*/
|
||||||
|
XGB_DLL SEXP XGDMatrixCreateFromCSR_R(SEXP indptr, SEXP indices, SEXP data, SEXP num_col,
|
||||||
|
SEXP missing, SEXP n_threads);
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief create a new dmatrix from sliced content of existing matrix
|
* \brief create a new dmatrix from sliced content of existing matrix
|
||||||
@@ -116,6 +131,14 @@ XGB_DLL SEXP XGDMatrixNumCol_R(SEXP handle);
|
|||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGBoosterCreate_R(SEXP dmats);
|
XGB_DLL SEXP XGBoosterCreate_R(SEXP dmats);
|
||||||
|
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief create xgboost learner, saving the pointer into an existing R object
|
||||||
|
* \param dmats a list of dmatrix handles that will be cached
|
||||||
|
* \param R_handle a clean R external pointer (not holding any object)
|
||||||
|
*/
|
||||||
|
XGB_DLL SEXP XGBoosterCreateInEmptyObj_R(SEXP dmats, SEXP R_handle);
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief set parameters
|
* \brief set parameters
|
||||||
* \param handle handle
|
* \param handle handle
|
||||||
@@ -156,15 +179,14 @@ XGB_DLL SEXP XGBoosterBoostOneIter_R(SEXP handle, SEXP dtrain, SEXP grad, SEXP h
|
|||||||
XGB_DLL SEXP XGBoosterEvalOneIter_R(SEXP handle, SEXP iter, SEXP dmats, SEXP evnames);
|
XGB_DLL SEXP XGBoosterEvalOneIter_R(SEXP handle, SEXP iter, SEXP dmats, SEXP evnames);
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief make prediction based on dmat
|
* \brief Run prediction on DMatrix, replacing `XGBoosterPredict_R`
|
||||||
* \param handle handle
|
* \param handle handle
|
||||||
* \param dmat data matrix
|
* \param dmat data matrix
|
||||||
* \param option_mask output_margin:1 predict_leaf:2
|
* \param json_config See `XGBoosterPredictFromDMatrix` in xgboost c_api.h
|
||||||
* \param ntree_limit limit number of trees used in prediction
|
*
|
||||||
* \param training Whether the prediction value is used for training.
|
* \return A list containing 2 vectors, first one for shape while second one for prediction result.
|
||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGBoosterPredict_R(SEXP handle, SEXP dmat, SEXP option_mask,
|
XGB_DLL SEXP XGBoosterPredictFromDMatrix_R(SEXP handle, SEXP dmat, SEXP json_config);
|
||||||
SEXP ntree_limit, SEXP training);
|
|
||||||
/*!
|
/*!
|
||||||
* \brief load model from existing file
|
* \brief load model from existing file
|
||||||
* \param handle handle
|
* \param handle handle
|
||||||
@@ -189,11 +211,21 @@ XGB_DLL SEXP XGBoosterSaveModel_R(SEXP handle, SEXP fname);
|
|||||||
XGB_DLL SEXP XGBoosterLoadModelFromRaw_R(SEXP handle, SEXP raw);
|
XGB_DLL SEXP XGBoosterLoadModelFromRaw_R(SEXP handle, SEXP raw);
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief save model into R's raw array
|
* \brief Save model into R's raw array
|
||||||
|
*
|
||||||
* \param handle handle
|
* \param handle handle
|
||||||
* \return raw array
|
* \param json_config JSON encoded string storing parameters for the function. Following
|
||||||
|
* keys are expected in the JSON document:
|
||||||
|
*
|
||||||
|
* "format": str
|
||||||
|
* - json: Output booster will be encoded as JSON.
|
||||||
|
* - ubj: Output booster will be encoded as Univeral binary JSON.
|
||||||
|
* - deprecated: Output booster will be encoded as old custom binary format. Do now use
|
||||||
|
* this format except for compatibility reasons.
|
||||||
|
*
|
||||||
|
* \return Raw array
|
||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGBoosterModelToRaw_R(SEXP handle);
|
XGB_DLL SEXP XGBoosterSaveModelToRaw_R(SEXP handle, SEXP json_config);
|
||||||
|
|
||||||
/*!
|
/*!
|
||||||
* \brief Save internal parameters as a JSON string
|
* \brief Save internal parameters as a JSON string
|
||||||
@@ -257,4 +289,12 @@ XGB_DLL SEXP XGBoosterSetAttr_R(SEXP handle, SEXP name, SEXP val);
|
|||||||
*/
|
*/
|
||||||
XGB_DLL SEXP XGBoosterGetAttrNames_R(SEXP handle);
|
XGB_DLL SEXP XGBoosterGetAttrNames_R(SEXP handle);
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief Get feature scores from the model.
|
||||||
|
* \param json_config See `XGBoosterFeatureScore` in xgboost c_api.h
|
||||||
|
* \return A vector with the first element as feature names, second element as shape of
|
||||||
|
* feature scores and thrid element as feature scores.
|
||||||
|
*/
|
||||||
|
XGB_DLL SEXP XGBoosterFeatureScore_R(SEXP handle, SEXP json_config);
|
||||||
|
|
||||||
#endif // XGBOOST_WRAPPER_R_H_ // NOLINT(*)
|
#endif // XGBOOST_WRAPPER_R_H_ // NOLINT(*)
|
||||||
|
|||||||
@@ -1,26 +0,0 @@
|
|||||||
// Copyright (c) 2014 by Contributors
|
|
||||||
#include <stdio.h>
|
|
||||||
#include <stdarg.h>
|
|
||||||
#include <Rinternals.h>
|
|
||||||
|
|
||||||
// implements error handling
|
|
||||||
void XGBoostAssert_R(int exp, const char *fmt, ...) {
|
|
||||||
char buf[1024];
|
|
||||||
if (exp == 0) {
|
|
||||||
va_list args;
|
|
||||||
va_start(args, fmt);
|
|
||||||
vsprintf(buf, fmt, args);
|
|
||||||
va_end(args);
|
|
||||||
error("AssertError:%s\n", buf);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
void XGBoostCheck_R(int exp, const char *fmt, ...) {
|
|
||||||
char buf[1024];
|
|
||||||
if (exp == 0) {
|
|
||||||
va_list args;
|
|
||||||
va_start(args, fmt);
|
|
||||||
vsprintf(buf, fmt, args);
|
|
||||||
va_end(args);
|
|
||||||
error("%s\n", buf);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
51
R-package/tests/helper_scripts/install_deps.R
Normal file
51
R-package/tests/helper_scripts/install_deps.R
Normal file
@@ -0,0 +1,51 @@
|
|||||||
|
## Install dependencies of R package for testing. The list might not be
|
||||||
|
## up-to-date, check DESCRIPTION for the latest list and update this one if
|
||||||
|
## inconsistent is found.
|
||||||
|
pkgs <- c(
|
||||||
|
## CI
|
||||||
|
"caret",
|
||||||
|
"pkgbuild",
|
||||||
|
"roxygen2",
|
||||||
|
"XML",
|
||||||
|
"cplm",
|
||||||
|
"e1071",
|
||||||
|
## suggests
|
||||||
|
"knitr",
|
||||||
|
"rmarkdown",
|
||||||
|
"ggplot2",
|
||||||
|
"DiagrammeR",
|
||||||
|
"Ckmeans.1d.dp",
|
||||||
|
"vcd",
|
||||||
|
"lintr",
|
||||||
|
"testthat",
|
||||||
|
"igraph",
|
||||||
|
"float",
|
||||||
|
"titanic",
|
||||||
|
## imports
|
||||||
|
"Matrix",
|
||||||
|
"methods",
|
||||||
|
"data.table",
|
||||||
|
"jsonlite"
|
||||||
|
)
|
||||||
|
|
||||||
|
ncpus <- parallel::detectCores()
|
||||||
|
print(paste0("Using ", ncpus, " cores to install dependencies."))
|
||||||
|
|
||||||
|
if (.Platform$OS.type == "unix") {
|
||||||
|
print("Installing source packages on unix.")
|
||||||
|
install.packages(
|
||||||
|
pkgs,
|
||||||
|
repo = "https://cloud.r-project.org",
|
||||||
|
dependencies = c("Depends", "Imports", "LinkingTo"),
|
||||||
|
Ncpus = parallel::detectCores()
|
||||||
|
)
|
||||||
|
} else {
|
||||||
|
print("Installing binary packages on Windows.")
|
||||||
|
install.packages(
|
||||||
|
pkgs,
|
||||||
|
repo = "https://cloud.r-project.org",
|
||||||
|
dependencies = c("Depends", "Imports", "LinkingTo"),
|
||||||
|
Ncpus = parallel::detectCores(),
|
||||||
|
type = "binary"
|
||||||
|
)
|
||||||
|
}
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
library(lintr)
|
|
||||||
library(crayon)
|
|
||||||
|
|
||||||
my_linters <- list(
|
|
||||||
absolute_path_linter = lintr::absolute_path_linter,
|
|
||||||
assignment_linter = lintr::assignment_linter,
|
|
||||||
closed_curly_linter = lintr::closed_curly_linter,
|
|
||||||
commas_linter = lintr::commas_linter,
|
|
||||||
equals_na = lintr::equals_na_linter,
|
|
||||||
infix_spaces_linter = lintr::infix_spaces_linter,
|
|
||||||
line_length_linter = lintr::line_length_linter,
|
|
||||||
no_tab_linter = lintr::no_tab_linter,
|
|
||||||
object_usage_linter = lintr::object_usage_linter,
|
|
||||||
object_length_linter = lintr::object_length_linter,
|
|
||||||
open_curly_linter = lintr::open_curly_linter,
|
|
||||||
semicolon = lintr::semicolon_terminator_linter,
|
|
||||||
seq = lintr::seq_linter,
|
|
||||||
spaces_inside_linter = lintr::spaces_inside_linter,
|
|
||||||
spaces_left_parentheses_linter = lintr::spaces_left_parentheses_linter,
|
|
||||||
trailing_blank_lines_linter = lintr::trailing_blank_lines_linter,
|
|
||||||
trailing_whitespace_linter = lintr::trailing_whitespace_linter,
|
|
||||||
true_false = lintr::T_and_F_symbol_linter,
|
|
||||||
unneeded_concatenation = lintr::unneeded_concatenation_linter
|
|
||||||
)
|
|
||||||
|
|
||||||
results <- lapply(
|
|
||||||
list.files(path = '.', pattern = '\\.[Rr]$', recursive = TRUE),
|
|
||||||
function (r_file) {
|
|
||||||
cat(sprintf("Processing %s ...\n", r_file))
|
|
||||||
list(r_file = r_file,
|
|
||||||
output = lintr::lint(filename = r_file, linters = my_linters))
|
|
||||||
})
|
|
||||||
num_issue <- Reduce(sum, lapply(results, function (e) length(e$output)))
|
|
||||||
|
|
||||||
lint2str <- function(lint_entry) {
|
|
||||||
color <- function(type) {
|
|
||||||
switch(type,
|
|
||||||
"warning" = crayon::magenta,
|
|
||||||
"error" = crayon::red,
|
|
||||||
"style" = crayon::blue,
|
|
||||||
crayon::bold
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
paste0(
|
|
||||||
lapply(lint_entry$output,
|
|
||||||
function (lint_line) {
|
|
||||||
paste0(
|
|
||||||
crayon::bold(lint_entry$r_file, ":",
|
|
||||||
as.character(lint_line$line_number), ":",
|
|
||||||
as.character(lint_line$column_number), ": ", sep = ""),
|
|
||||||
color(lint_line$type)(lint_line$type, ": ", sep = ""),
|
|
||||||
crayon::bold(lint_line$message), "\n",
|
|
||||||
lint_line$line, "\n",
|
|
||||||
lintr:::highlight_string(lint_line$message, lint_line$column_number, lint_line$ranges),
|
|
||||||
"\n",
|
|
||||||
collapse = "")
|
|
||||||
}),
|
|
||||||
collapse = "")
|
|
||||||
}
|
|
||||||
|
|
||||||
if (num_issue > 0) {
|
|
||||||
cat(sprintf('R linters found %d issues:\n', num_issue))
|
|
||||||
for (entry in results) {
|
|
||||||
if (length(entry$output)) {
|
|
||||||
cat(paste0('**** ', crayon::bold(entry$r_file), '\n'))
|
|
||||||
cat(paste0(lint2str(entry), collapse = ''))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
quit(save = 'no', status = 1) # Signal error to parent shell
|
|
||||||
}
|
|
||||||
@@ -1,5 +1,3 @@
|
|||||||
require(xgboost)
|
|
||||||
|
|
||||||
context("basic functions")
|
context("basic functions")
|
||||||
|
|
||||||
data(agaricus.train, package = 'xgboost')
|
data(agaricus.train, package = 'xgboost')
|
||||||
@@ -34,6 +32,10 @@ test_that("train and predict binary classification", {
|
|||||||
err_pred1 <- sum((pred1 > 0.5) != train$label) / length(train$label)
|
err_pred1 <- sum((pred1 > 0.5) != train$label) / length(train$label)
|
||||||
err_log <- bst$evaluation_log[1, train_error]
|
err_log <- bst$evaluation_log[1, train_error]
|
||||||
expect_lt(abs(err_pred1 - err_log), 10e-6)
|
expect_lt(abs(err_pred1 - err_log), 10e-6)
|
||||||
|
|
||||||
|
pred2 <- predict(bst, train$data, iterationrange = c(1, 2))
|
||||||
|
expect_length(pred1, 6513)
|
||||||
|
expect_equal(pred1, pred2)
|
||||||
})
|
})
|
||||||
|
|
||||||
test_that("parameter validation works", {
|
test_that("parameter validation works", {
|
||||||
@@ -83,9 +85,18 @@ test_that("dart prediction works", {
|
|||||||
rnorm(100)
|
rnorm(100)
|
||||||
|
|
||||||
set.seed(1994)
|
set.seed(1994)
|
||||||
booster_by_xgboost <- xgboost(data = d, label = y, max_depth = 2, booster = "dart",
|
booster_by_xgboost <- xgboost(
|
||||||
rate_drop = 0.5, one_drop = TRUE,
|
data = d,
|
||||||
eta = 1, nthread = 2, nrounds = nrounds, objective = "reg:squarederror")
|
label = y,
|
||||||
|
max_depth = 2,
|
||||||
|
booster = "dart",
|
||||||
|
rate_drop = 0.5,
|
||||||
|
one_drop = TRUE,
|
||||||
|
eta = 1,
|
||||||
|
nthread = 2,
|
||||||
|
nrounds = nrounds,
|
||||||
|
objective = "reg:squarederror"
|
||||||
|
)
|
||||||
pred_by_xgboost_0 <- predict(booster_by_xgboost, newdata = d, ntreelimit = 0)
|
pred_by_xgboost_0 <- predict(booster_by_xgboost, newdata = d, ntreelimit = 0)
|
||||||
pred_by_xgboost_1 <- predict(booster_by_xgboost, newdata = d, ntreelimit = nrounds)
|
pred_by_xgboost_1 <- predict(booster_by_xgboost, newdata = d, ntreelimit = nrounds)
|
||||||
expect_true(all(matrix(pred_by_xgboost_0, byrow = TRUE) == matrix(pred_by_xgboost_1, byrow = TRUE)))
|
expect_true(all(matrix(pred_by_xgboost_0, byrow = TRUE) == matrix(pred_by_xgboost_1, byrow = TRUE)))
|
||||||
@@ -95,14 +106,14 @@ test_that("dart prediction works", {
|
|||||||
|
|
||||||
set.seed(1994)
|
set.seed(1994)
|
||||||
dtrain <- xgb.DMatrix(data = d, info = list(label = y))
|
dtrain <- xgb.DMatrix(data = d, info = list(label = y))
|
||||||
booster_by_train <- xgb.train(params = list(
|
booster_by_train <- xgb.train(
|
||||||
|
params = list(
|
||||||
booster = "dart",
|
booster = "dart",
|
||||||
max_depth = 2,
|
max_depth = 2,
|
||||||
eta = 1,
|
eta = 1,
|
||||||
rate_drop = 0.5,
|
rate_drop = 0.5,
|
||||||
one_drop = TRUE,
|
one_drop = TRUE,
|
||||||
nthread = 1,
|
nthread = 1,
|
||||||
tree_method = "exact",
|
|
||||||
objective = "reg:squarederror"
|
objective = "reg:squarederror"
|
||||||
),
|
),
|
||||||
data = dtrain,
|
data = dtrain,
|
||||||
@@ -143,6 +154,24 @@ test_that("train and predict softprob", {
|
|||||||
pred_labels <- max.col(mpred) - 1
|
pred_labels <- max.col(mpred) - 1
|
||||||
err <- sum(pred_labels != lb) / length(lb)
|
err <- sum(pred_labels != lb) / length(lb)
|
||||||
expect_equal(bst$evaluation_log[1, train_merror], err, tolerance = 5e-6)
|
expect_equal(bst$evaluation_log[1, train_merror], err, tolerance = 5e-6)
|
||||||
|
|
||||||
|
mpred1 <- predict(bst, as.matrix(iris[, -5]), reshape = TRUE, iterationrange = c(1, 2))
|
||||||
|
expect_equal(mpred, mpred1)
|
||||||
|
|
||||||
|
d <- cbind(
|
||||||
|
x1 = rnorm(100),
|
||||||
|
x2 = rnorm(100),
|
||||||
|
x3 = rnorm(100)
|
||||||
|
)
|
||||||
|
y <- sample.int(10, 100, replace = TRUE) - 1
|
||||||
|
dtrain <- xgb.DMatrix(data = d, info = list(label = y))
|
||||||
|
booster <- xgb.train(
|
||||||
|
params = list(tree_method = "hist"), data = dtrain, nrounds = 4, num_class = 10,
|
||||||
|
objective = "multi:softprob"
|
||||||
|
)
|
||||||
|
predt <- predict(booster, as.matrix(d), reshape = TRUE, strict_shape = FALSE)
|
||||||
|
expect_equal(ncol(predt), 10)
|
||||||
|
expect_equal(rowSums(predt), rep(1, 100), tolerance = 1e-7)
|
||||||
})
|
})
|
||||||
|
|
||||||
test_that("train and predict softmax", {
|
test_that("train and predict softmax", {
|
||||||
@@ -182,10 +211,8 @@ test_that("train and predict RF", {
|
|||||||
pred_err_20 <- sum((pred > 0.5) != lb) / length(lb)
|
pred_err_20 <- sum((pred > 0.5) != lb) / length(lb)
|
||||||
expect_equal(pred_err_20, pred_err)
|
expect_equal(pred_err_20, pred_err)
|
||||||
|
|
||||||
#pred <- predict(bst, train$data, ntreelimit = 1)
|
pred1 <- predict(bst, train$data, iterationrange = c(1, 2))
|
||||||
#pred_err_1 <- sum((pred > 0.5) != lb)/length(lb)
|
expect_equal(pred, pred1)
|
||||||
#expect_lt(pred_err, pred_err_1)
|
|
||||||
#expect_lt(pred_err, 0.08)
|
|
||||||
})
|
})
|
||||||
|
|
||||||
test_that("train and predict RF with softprob", {
|
test_that("train and predict RF with softprob", {
|
||||||
@@ -220,6 +247,14 @@ test_that("use of multiple eval metrics works", {
|
|||||||
expect_false(is.null(bst$evaluation_log))
|
expect_false(is.null(bst$evaluation_log))
|
||||||
expect_equal(dim(bst$evaluation_log), c(2, 4))
|
expect_equal(dim(bst$evaluation_log), c(2, 4))
|
||||||
expect_equal(colnames(bst$evaluation_log), c("iter", "train_error", "train_auc", "train_logloss"))
|
expect_equal(colnames(bst$evaluation_log), c("iter", "train_error", "train_auc", "train_logloss"))
|
||||||
|
expect_output(
|
||||||
|
bst2 <- xgboost(data = train$data, label = train$label, max_depth = 2,
|
||||||
|
eta = 1, nthread = 2, nrounds = 2, objective = "binary:logistic",
|
||||||
|
eval_metric = list("error", "auc", "logloss"))
|
||||||
|
, "train-error.*train-auc.*train-logloss")
|
||||||
|
expect_false(is.null(bst2$evaluation_log))
|
||||||
|
expect_equal(dim(bst2$evaluation_log), c(2, 4))
|
||||||
|
expect_equal(colnames(bst2$evaluation_log), c("iter", "train_error", "train_auc", "train_logloss"))
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
||||||
@@ -331,7 +366,7 @@ test_that("train and predict with non-strict classes", {
|
|||||||
expect_error(pr <- predict(bst, train_dense), regexp = NA)
|
expect_error(pr <- predict(bst, train_dense), regexp = NA)
|
||||||
expect_equal(pr0, pr)
|
expect_equal(pr0, pr)
|
||||||
|
|
||||||
# when someone inhertis from xgb.Booster, it should still be possible to use it as xgb.Booster
|
# when someone inherits from xgb.Booster, it should still be possible to use it as xgb.Booster
|
||||||
class(bst) <- c('super.Booster', 'xgb.Booster')
|
class(bst) <- c('super.Booster', 'xgb.Booster')
|
||||||
expect_error(pr <- predict(bst, train_dense), regexp = NA)
|
expect_error(pr <- predict(bst, train_dense), regexp = NA)
|
||||||
expect_equal(pr0, pr)
|
expect_equal(pr0, pr)
|
||||||
@@ -346,7 +381,7 @@ test_that("max_delta_step works", {
|
|||||||
bst1 <- xgb.train(param, dtrain, nrounds, watchlist, verbose = 1)
|
bst1 <- xgb.train(param, dtrain, nrounds, watchlist, verbose = 1)
|
||||||
# model with restricted max_delta_step
|
# model with restricted max_delta_step
|
||||||
bst2 <- xgb.train(param, dtrain, nrounds, watchlist, verbose = 1, max_delta_step = 1)
|
bst2 <- xgb.train(param, dtrain, nrounds, watchlist, verbose = 1, max_delta_step = 1)
|
||||||
# the no-restriction model is expected to have consistently lower loss during the initial interations
|
# the no-restriction model is expected to have consistently lower loss during the initial iterations
|
||||||
expect_true(all(bst1$evaluation_log$train_logloss < bst2$evaluation_log$train_logloss))
|
expect_true(all(bst1$evaluation_log$train_logloss < bst2$evaluation_log$train_logloss))
|
||||||
expect_lt(mean(bst1$evaluation_log$train_logloss) / mean(bst2$evaluation_log$train_logloss), 0.8)
|
expect_lt(mean(bst1$evaluation_log$train_logloss) / mean(bst2$evaluation_log$train_logloss), 0.8)
|
||||||
})
|
})
|
||||||
@@ -373,7 +408,7 @@ test_that("colsample_bytree works", {
|
|||||||
xgb.importance(model = bst)
|
xgb.importance(model = bst)
|
||||||
# If colsample_bytree works properly, a variety of features should be used
|
# If colsample_bytree works properly, a variety of features should be used
|
||||||
# in the 100 trees
|
# in the 100 trees
|
||||||
expect_gte(nrow(xgb.importance(model = bst)), 30)
|
expect_gte(nrow(xgb.importance(model = bst)), 28)
|
||||||
})
|
})
|
||||||
|
|
||||||
test_that("Configuration works", {
|
test_that("Configuration works", {
|
||||||
@@ -383,5 +418,74 @@ test_that("Configuration works", {
|
|||||||
config <- xgb.config(bst)
|
config <- xgb.config(bst)
|
||||||
xgb.config(bst) <- config
|
xgb.config(bst) <- config
|
||||||
reloaded_config <- xgb.config(bst)
|
reloaded_config <- xgb.config(bst)
|
||||||
expect_equal(config, reloaded_config);
|
expect_equal(config, reloaded_config)
|
||||||
|
})
|
||||||
|
|
||||||
|
test_that("strict_shape works", {
|
||||||
|
n_rounds <- 2
|
||||||
|
|
||||||
|
test_strict_shape <- function(bst, X, n_groups) {
|
||||||
|
predt <- predict(bst, X, strict_shape = TRUE)
|
||||||
|
margin <- predict(bst, X, outputmargin = TRUE, strict_shape = TRUE)
|
||||||
|
contri <- predict(bst, X, predcontrib = TRUE, strict_shape = TRUE)
|
||||||
|
interact <- predict(bst, X, predinteraction = TRUE, strict_shape = TRUE)
|
||||||
|
leaf <- predict(bst, X, predleaf = TRUE, strict_shape = TRUE)
|
||||||
|
|
||||||
|
n_rows <- nrow(X)
|
||||||
|
n_cols <- ncol(X)
|
||||||
|
|
||||||
|
expect_equal(dim(predt), c(n_groups, n_rows))
|
||||||
|
expect_equal(dim(margin), c(n_groups, n_rows))
|
||||||
|
expect_equal(dim(contri), c(n_cols + 1, n_groups, n_rows))
|
||||||
|
expect_equal(dim(interact), c(n_cols + 1, n_cols + 1, n_groups, n_rows))
|
||||||
|
expect_equal(dim(leaf), c(1, n_groups, n_rounds, n_rows))
|
||||||
|
|
||||||
|
if (n_groups != 1) {
|
||||||
|
for (g in seq_len(n_groups)) {
|
||||||
|
expect_lt(max(abs(colSums(contri[, g, ]) - margin[g, ])), 1e-5)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
test_iris <- function() {
|
||||||
|
y <- as.numeric(iris$Species) - 1
|
||||||
|
X <- as.matrix(iris[, -5])
|
||||||
|
|
||||||
|
bst <- xgboost(data = X, label = y,
|
||||||
|
max_depth = 2, nrounds = n_rounds,
|
||||||
|
objective = "multi:softprob", num_class = 3, eval_metric = "merror")
|
||||||
|
|
||||||
|
test_strict_shape(bst, X, 3)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
test_agaricus <- function() {
|
||||||
|
data(agaricus.train, package = 'xgboost')
|
||||||
|
X <- agaricus.train$data
|
||||||
|
y <- agaricus.train$label
|
||||||
|
|
||||||
|
bst <- xgboost(data = X, label = y, max_depth = 2,
|
||||||
|
nrounds = n_rounds, objective = "binary:logistic",
|
||||||
|
eval_metric = 'error', eval_metric = 'auc', eval_metric = "logloss")
|
||||||
|
|
||||||
|
test_strict_shape(bst, X, 1)
|
||||||
|
}
|
||||||
|
|
||||||
|
test_iris()
|
||||||
|
test_agaricus()
|
||||||
|
})
|
||||||
|
|
||||||
|
test_that("'predict' accepts CSR data", {
|
||||||
|
X <- agaricus.train$data
|
||||||
|
y <- agaricus.train$label
|
||||||
|
x_csc <- as(X[1L, , drop = FALSE], "CsparseMatrix")
|
||||||
|
x_csr <- as(x_csc, "RsparseMatrix")
|
||||||
|
x_spv <- as(x_csc, "sparseVector")
|
||||||
|
bst <- xgboost(data = X, label = y, objective = "binary:logistic",
|
||||||
|
nrounds = 5L, verbose = FALSE)
|
||||||
|
p_csc <- predict(bst, x_csc)
|
||||||
|
p_csr <- predict(bst, x_csr)
|
||||||
|
p_spv <- predict(bst, x_spv)
|
||||||
|
expect_equal(p_csc, p_csr)
|
||||||
|
expect_equal(p_csc, p_spv)
|
||||||
})
|
})
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user