Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
examples/jais/README.md
.examples/dit/README.md
.Video NeVA
section inexamples/multimodal/README.md
.distil-whisper/distil-large-v3
, thanks to the contribution from @IbrahimAmin1 in [feat]: Add Option to convert and run distil-whisper large-v3 #1337.trtllm-build
command), see documents: examples/whisper/README.md.free_gpu_memory_fraction
inModelRunnerCpp
tokv_cache_free_gpu_memory_fraction
ModelRunnerCpp
, includingmax_tokens_in_paged_kv_cache
,kv_cache_enable_block_reuse
andenable_chunked_context
enable_executor
fromtensorrt_llm.LLM
API as it is using the C++Executor
API now.OutputConfig
ingenerate
API.BuildConfig
to thetensorrt_llm.LLM
API.LLM
construction phase, remove most of the trivial logs.SpeculativeDecodingMode.h
to choose between different speculative decoding techniques.SpeculativeDecodingModule.h
base class for speculative decoding techniquesdecodingMode.h
nvcr.io/nvidia/pytorch:24.04-py3
.nvcr.io/nvidia/tritonserver:24.04-py3
.