GPU Recommendation for XGBoost with MATLAB R2022b
Show older comments
Hi,
I'm currently planning to use XGBoost with GPU acceleration, mainly for parameter optimization. Right now, I'm performing leave-one-out cross-validation, and my feature dimensions are around 1000.
I haven’t purchased a GPU yet, but I'm considering either:
GeForce RTX 5060 Ti/ 5060 (though I couldn’t find its compute capability listed on the official NVIDIA site: https://developer.nvidia.com/cuda-gpus ), or
GeForce RTX 4070 Ti, which has a compute capability of 8.9.
I'm using MATLAB R2022b, and according to the documentation, "MATLAB® supports NVIDIA® GPU architectures with compute capability 3.5 to 8.x."
Could anyone clarify whether either (or both) of these GPUs would be supported? Or perhaps suggest an alternative GPU that would work well with my setup?
Thanks in advance!
3 Comments
Walter Roberson
on 18 Apr 2025
Note that the 40x0 and 40x0 Ti series all have floating point double precision (FP64) performance at 1/64 of the FP32 performance.
- 4070 Ti Super: 689.0 GFLOPS
- 5070 Ti: 686.6 GFLOPS
- 4070 Ti: 626.4 GFLOPS
Compare this to the end-of-life 1080 Ti, at 354.4 (a 1:32 rating insead of a 1:64 rating), or the 3080 Ti at 532.8 GFLOPS (1:64). This leads to questions about whether it is more cost effective to purchase multiple lower-rated GPUs if your intention is to do double precison work
Walter Roberson
on 19 Apr 2025
I find claims that XGBoost treats all values as single precision floating point, so my concerns about double precision performance are perhaps misplaced here.
- 4070 Ti Super: 44.10 TFLOPS fp32
- 4060 Ti: 22.06 TFLOPS fp32
- 5070 Ti: 43.94 TFLOPS fp32
- 5060 Ti: 23.70 TFLOPS fp32
- 1080 Ti: 11.34 TFLOPS fp32
JIA
on 20 Apr 2025
Answers (1)
Joss Knight
on 18 Apr 2025
Edited: Joss Knight
on 18 Apr 2025
2 votes
Yes, MATLAB R2022b supports these GPUs.

7 Comments
JIA
on 19 Apr 2025
Walter Roberson
on 19 Apr 2025
The 5060 Ti is Compute Capacity 12.0 .
This device is not fully supported by MATLAB R2022b (not even by R2025a Pre-release).
There is a forward-support mode that can be enabled that will cause kernels to be automatically re-built on demand. However, problems can be expected, the more so for older versions such as R2022b. It is unclear at the moment whether the problems will be trivial or "a nuisance to work around" or "quite serious".
The 5060 Ti is rated for 370.4 GFLOPS of FP64, which is considerably worse than the FP64 performance of the 4070 Ti.
Joss Knight
on 19 Apr 2025
I think it's fair to say that R2022b isn't going to work well with the RTX 5060 in forward compatibility mode. R2024a, R2024b and R2025a should be okay (although be wary of the caveats in this post).
Walter Roberson
on 19 Apr 2025
By the way, to enable forward compatibility see https://www.mathworks.com/help/parallel-computing/parallel.gpu.enablecudaforwardcompatibility.html
JIA
on 20 Apr 2025
Walter Roberson
on 20 Apr 2025
- RTX 3060 fp32 12.74 TFLOPS
- RTX 3060 Ti fp32 16.20 TFLOPS
- RTX 3070 fp32 20.31 TFLOPS
- RTX 3070 ti fp32 21.75 TFLOPS
- RTX 3080 fp32 29.77 TFLOPS
- RTX 3080 ti fp32 34.10 TFLOPS
- RTX 3090 fp32 35.58 TFLOPS
- RTX 3090 ti fp32 40.00 TFLOPS
Joss Knight
on 20 Apr 2025
Good choice. The 30xx series is still pretty recent and most of the improvements since have been in reduced precision that you are not using.
Categories
Find more on Get Started with GPU Coder in Help Center and File Exchange
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!