From 4bf23cfbb1ed607d8062174520b074b5083bd2d2 Mon Sep 17 00:00:00 2001 From: antinucleon Date: Fri, 16 May 2014 18:05:17 -0600 Subject: [PATCH] new speed test --- demo/kaggle-higgs/speedtest.py | 17 ++++++++++------- 1 file changed, 10 insertions(+), 7 deletions(-) diff --git a/demo/kaggle-higgs/speedtest.py b/demo/kaggle-higgs/speedtest.py index 1b07d4619..d17c2680b 100755 --- a/demo/kaggle-higgs/speedtest.py +++ b/demo/kaggle-higgs/speedtest.py @@ -42,22 +42,25 @@ param['eval_metric'] = 'auc' param['silent'] = 1 param['nthread'] = 4 -# you can directly throw param in, though we want to watch multiple metrics here plst = param.items()+[('eval_metric', 'ams@0.15')] watchlist = [ (xgmat,'train') ] -# boost 135 tres -num_round = 135 +# boost 10 tres +num_round = 10 print 'loading data end, start to boost trees' print "training GBM from sklearn" tmp = time.time() -gbm = GradientBoostingClassifier(n_estimators=135, max_depth=6, verbose=2) +gbm = GradientBoostingClassifier(n_estimators=num_round, max_depth=6, verbose=2) gbm.fit(data, label) print "GBM costs: %s seconds" % str(time.time() - tmp) #raw_input() print "training xgboost" -tmp = time.time() -bst = xgb.train( plst, xgmat, num_round, watchlist ); -print "XGBoost costs: %s seconds" % str(time.time() - tmp) +threads = [1, 2, 4, 16] +for i in threads: + param['nthread'] = i + tmp = time.time() + plst = param.items()+[('eval_metric', 'ams@0.15')] + bst = xgb.train( plst, xgmat, num_round, watchlist ); + print "XGBoost with %d thread costs: %s seconds" % (i, str(time.time() - tmp)) print 'finish training'