fix(llama.cpp): set -1 as default for max tokens #2330
test-extra.yml
on: pull_request
tests-transformers
5m 13s
tests-sentencetransformers
5m 54s
tests-diffusers
4m 5s
tests-parler-tts
4m 34s
tests-transformers-musicgen
6m 21s
tests-vallex
6m 13s
tests-coqui
5m 21s