-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Build fails for -DUSE_CUDA=1
#5785
Comments
@jmakov Is it possible to see more error message? For example, why the compilation of cuda_best_split_finder.cu fail? |
@shiyu1994 there seems to be only 1 type of error: /tmp/lib/LightGBM/include/LightGBM/utils/../../../external_libs/fmt/include/fmt/format-inl.h(85): here
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with '...':
435 | function(_Functor&& __f)
|
^
/usr/include/c++/11/bits/std_function.h:435:145: note: '_ArgTypes'
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with '...':
530 | operator=(_Functor&& __f)
|
^
/usr/include/c++/11/bits/std_function.h:530:146: note: '_ArgTypes' whole log: |
This is kinda a blocker for me. Would be great to have some more insight into what can be done about it. |
I've been having similar problems I think when trying to install v4.0. Builds were failing until I switched gcc (and g++ for good measure) to version 10 for compiling. Found solution from this reference: NVIDIA/nccl#650 |
Sorry for the long delay in response. I believe recent changes in LightGBM have fixed this. I was able to build latest LightGBM (1443548) in the latest stable ( docker run \
--rm \
--user root \
-it rapidsai/base:24.04-cuda12.0-py3.10 \
bash
mkdir /tmp/lib
cd /tmp/lib
# install build tools (rapidsai/core doesn't ship these)
apt-get update
apt-get install -y \
build-essential \
cmake \
git
# build LightGBM
git clone --recursive https://github.com/microsoft/LightGBM
cd ./LightGBM
cmake -B build -S . -DUSE_CUDA=1
cmake --build build --target _lightgbm -j2
sh build-python.sh install --precompile That built successfully for me. full logs (click me)Configure step:
Build step:
Python build + install logs.
|
This issue has been automatically closed because it has been awaiting a response for too long. When you have time to to work with the maintainers to resolve this issue, please post a new comment and it will be re-opened. If the issue has been locked for editing by the time you return to it, please open a new issue and reference this one. Thank you for taking the time to improve LightGBM! |
Wondering if it's possible to enforce architecture somehow. Trying to reproduce your commands on NVIDIA RTX 6000 Ada (SM 8.9) & CUDA Version: 12.4, Ubuntu 20.04.6 LTS leads to
|
Nevermind. I had to remove nvidia-cuda-toolkit (which I installed 'cause it allowed open CL version of lightgbm to work, only to find out it's buggy on big datasets and overall an abandoned branch). Currently stuck at
Why is it so hard to get lightgbm working with GPU? Catboost & Xgboost teams somehow managed to solve it with single "pip install" command ;-) |
You have to have Python installed and a I strongly suspect that you aren't using the exact example I provided in #5785 (comment), but you haven't described your setup here so it's not possible to help much more.
We're doing the best we can with a much smaller amount of maintainer availability. Those projects both have multiple maintainers being paid to work on them full-time... LightGBM does not. You're welcome to come contribute here any time. |
Yeah, I know. Thanks a lot for your hard work, guys. I hope getting an easier access to GPU training is on the roadmap. Not experienced myself in that, otherwise would contribute for sure. |
Description
#5089 is marked as resolved but this is still the case trying to build in RAPIDS Docker container:
Reproducible example
Environment info
LightGBM version or commit hash:
Command(s) you used to install LightGBM
Build in docker
FROM rapidsai/rapidsai-core:23.02-cuda11.8-runtime-ubuntu22.04-py3.10
GCC 11.3
Additional Comments
The text was updated successfully, but these errors were encountered: