Replies: 1 comment 1 reply
-
If you get bored and need something to do, I would love to see a side by side comparison between the RPI5 and the RK1. It would be interesting to see how the RK3588 stacks up against the A76. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Moved from this issue. Author: @segabor
I promised to share results of Turing RK1 module. It arrived yesterday so I took the chance to run distributed llama on it.
Capability: 8 cores, 32 GB RAM. Storage: 1 TB NVMe SSD
OS: custom Ubuntu Server
Model: llama-2-7b
Command
sudo nice -n -20 ./main inference \ --model /mnt/bigdata/llama-2-7b/dllama_llama-2-7b_q40.bin \ --tokenizer ./tokenizer.bin \ --weights-float-type q40 \ --buffer-float-type q80 \ --prompt "Hello world" \ --steps 16 \ --nthreads 4
Result
Beta Was this translation helpful? Give feedback.
All reactions