Skip to content

Commit 24ffaf7

Browse files
RAMitchelltqchen
authored andcommitted
Update benchmarks (#34)
1 parent 104922d commit 24ffaf7

File tree

1 file changed

+9
-7
lines changed

1 file changed

+9
-7
lines changed

_posts/2016-12-14-GPU-accelerated-xgboost.markdown

Lines changed: 9 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,10 @@ date: 2016-12-14 00.00.00 -0800
55
author: Rory Mitchell
66
---
77

8-
# GPU Accelerated XGBoost
8+
***Update 2016/12/23:***
9+
Some of our benchmarks were incorrect due to a wrong compiler flag. These have all been updated below.
10+
11+
---
912

1013
Decision tree learning and gradient boosting have until recently been the domain of multicore CPUs. Here we showcase a new plugin providing GPU acceleration for the [XGBoost library](https://github.com/dmlc/xgboost). The plugin provides significant speedups over multicore CPUs for large datasets.
1114

@@ -41,15 +44,15 @@ Luckily the decision tree construction algorithm may be formulated in parallel,
4144

4245

4346
## How fast is it?
44-
The following benchmarks show a performance comparison of GPUs against multicore CPUs for 500 boosting iterations. The new Pascal Titan X shows some nice performance improvements of up to 9.4x as compared to an i7 CPU. The Titan is also able to process the entire 10M row Higgs dataset in its 12GB of memory.
47+
The following benchmarks show a performance comparison of GPUs against multicore CPUs for 500 boosting iterations. The new Pascal Titan X shows some nice performance improvements of up to 5.57x as compared to an i7 CPU. The Titan is also able to process the entire 10M row Higgs dataset in its 12GB of memory.
4548

4649
Dataset | Instances | Features | i7-6700K | Titan X (pascal) | Speedup
4750
--- | --- | --- | --- | --- | ---
48-
Yahoo LTR | 473,134 | 700 | 3738 | 507 | 7.37
49-
Higgs | 10,500,000 | 28 | 31352 | 4173 | 7.51
50-
Bosch | 1,183,747 | 968 | 9460 | 1009 | 9.38
51+
Yahoo LTR | 473,134 | 700 | 877 | 277 | 3.16
52+
Higgs | 10,500,000 | 28 | 14504 | 3052 | 4.75
53+
Bosch | 1,183,747 | 968 | 3294 | 591 | 5.57
5154

52-
We also tested the Titan X against a server with 2x Xeon E5-2695v2 CPUs (24 cores in total) on the Yahoo learning to rank dataset. The CPUs outperform the GPU by about 1.5x when using all available cores. This is still a nice result however considering the Titan X costs $1200 and the 2x Xeon CPUs cost almost $5000.
55+
We also tested the Titan X against a server with 2x Xeon E5-2695v2 CPUs (24 cores in total) on the Yahoo learning to rank dataset. The GPU outperforms the CPUs by about 1.2x. This is a nice result considering the Titan X costs $1200 and the 2x Xeon CPUs cost almost $5000.
5356

5457
![](https://github.com/dmlc/web-data/raw/master/xgboost/gpu/yahooltr_xeon_titan.png)
5558

@@ -60,7 +63,6 @@ The algorithm also switches between two modes. The first mode processes node gro
6063

6164
## How do I use it?
6265
To use the GPU algorithm add the single parameter:
63-
6466
```python
6567
# Python example
6668
param['updater'] = 'grow_gpu'

0 commit comments

Comments
 (0)