-
Notifications
You must be signed in to change notification settings - Fork 128
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU Stats #262
Comments
We may be able to pull data for Nvidia and AMD GPUs from It would probably only work with the binary version of the agent, though. |
Experimental support for Nvidia and AMD will be added in 0.7.4. This works for the binary agent only and requires To enable, set the environment variable If you used the install script, you can do this by adding sudo systemctl daemon-reload
sudo systemctl restart beszel-agent Any feedback is appreciated. If it works then I'll enable it by default in the next minor release. I don't have a device using an Intel GPU, unfortunately, so I won't be able to add that. Tip Installing sudo ln -s /opt/rocm/bin/rocm-smi /usr/local/bin/rocm-smi |
Argh sad you can't get intel stats being most are no doubt using iGPUs 🙁 Amazing work for the Nvidia and AMD guys though. |
If it's useful here is a JSON output from intel_gpu_top
https://manpages.debian.org/testing/intel-gpu-tools/intel_gpu_top.1.en.html |
Maybe |
How do I do this if I'm running beszel-agent as a Docker service through docker-compose.yaml? WARN GPU err="no GPU found - install nvidia-smi or rocm-smi" # cat docker-compose.yaml
services:
beszel-agent:
image: "henrygd/beszel-agent"
container_name: "beszel-agent"
restart: unless-stopped
network_mode: host
volumes:
- /var/run/docker.sock:/var/run/docker.sock:ro
environment:
PORT: 45876
KEY: "ssh-ed25519 AAAAC3NzaC1lZ......adrfOpvRdFLD6p"
GPU: "true" # nvidia-smi
Sun Nov 10 13:59:38 2024
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 525.125.06 Driver Version: 525.125.06 CUDA Version: 12.0 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 NVIDIA GeForce ... Off | 00000000:09:00.0 Off | N/A |
| 35% 38C P8 14W / 215W | 1121MiB / 8192MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
| 1 NVIDIA GeForce ... Off | 00000000:0A:00.0 Off | N/A |
| 33% 35C P8 2W / 215W | 3MiB / 8192MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=============================================================================|
| 0 N/A N/A 1472954 C python 1118MiB |
+-----------------------------------------------------------------------------+ |
You can't at the moment, see above May it will be possible in future. The image needs to include nvidia-smi or rocm-smi, then it should be possible to mount the gpu like this: services:
beszel-agent:
image: "henrygd/beszel-agent"
container_name: "beszel-agent"
restart: unless-stopped
network_mode: host
volumes:
- /var/run/docker.sock:/var/run/docker.sock:ro
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: all
capabilities:
- gpu
environment:
PORT: 45876
KEY: "ssh-ed25519 AAAAC3NzaC1lZ......adrfOpvRdFLD6p"
GPU: "true" NVIDIA container toolkit should be installed. |
I don't plan to include But I can add example dockerfiles and compose configs. Maybe even build them automatically as separate images or tags. |
As separate images would be the best way 👍🏻 |
@Morethanevil This didn't work for me. I have container toolkit installed, we are using GPU in our other applications deployed by docker-compose and the
So it can't be done if it's not build as separated images? |
As I said before: it is not possible at the moment with docker, since there is no nvidia-smi included. You need to wait for a separate image. I just provided an example how to include the GPU into the compose.yaml (docker-compose.yml) |
I see, sorry for the misunderstanding. Haven't got a coffee yet. (cheap excuse 😅) |
Interesting. Didn't know this existed. Most guides point you towards jntel_gpu_top. I'm happy to test but appreciate its tricky to code it you don't have the hardware. |
I think I was wrong about Does anyone know if it works with newer iGPUs and Arc cards? If someone with Intel can look into this further and compare We need JSON or CSV output and ideally all the same info as Nvidia / AMD -- GPU name, utilization, VRAM usage, power draw, and temperature. Maybe next week I'll have some time to try doing it blind with sample output. |
I'm running a Intel Arc A750 and I can read out the following informations as a JSON:
|
#262 (comment) - my output here from an Intel ARC A310 |
Would be nice to be able to see GPU stats using intel_gpu_top or AMD equivalent.
Running the hub in docker but agent on containers I'm yet to install the agent on the Proxmox host.
The text was updated successfully, but these errors were encountered: