Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TensorRT-LLM v0.17 Release #2725

Merged
merged 1 commit into from
Jan 30, 2025
Merged

TensorRT-LLM v0.17 Release #2725

merged 1 commit into from
Jan 30, 2025

Conversation

schetlur-nv
Copy link
Collaborator

@schetlur-nv schetlur-nv commented Jan 30, 2025

TensorRT-LLM Release 0.17.0 [UPDATED 1/31]

Key Features and Enhancements

  • Blackwell support
    • NOTE: pip installation is not supported for TRT-LLM 0.17 on Blackwell platforms only. Instead, it is recommended that the user build from source using NVIDIA NGC 25.01 PyTorch container.
    • Added support for B200.
    • Added support for GeForce RTX 50 series using Windows Subsystem for Linux (WSL) for limited models.
    • Added NVFP4 Gemm support for Llama and Mixtral models.
    • Added NVFP4 support for the LLM API and trtllm-bench command.
    • GB200 NVL is not fully supported.
    • Added benchmark script to measure perf benefits of KV cache host offload with expected runtime improvements from GH200.
  • PyTorch workflow
    • The PyTorch workflow is an experimental feature in tensorrt_llm._torch. The following is a list of supported infrastructure, models, and features that can be used with the PyTorch workflow.
    • Added support for H100/H200/B200.
    • Added support for Llama models, Mixtral, QWen, Vila.
    • Added support for FP16/BF16/FP8/NVFP4 Gemm and fused Mixture-Of-Experts (MOE), FP16/BF16/FP8 KVCache.
    • Added custom context and decoding attention kernels support via PyTorch custom op.
    • Added support for chunked context (default off).
    • Added CudaGraph support for decoding only.
    • Added overlap scheduler support to overlap prepare inputs and model forward by decoding 1 extra token.
  • Added FP8 context FMHA support for the W4A8 quantization workflow.
  • Added ModelOpt quantized checkpoint support for the LLM API.
  • Added FP8 support for the Llama-3.2 VLM model. Refer to the “MLLaMA” section in examples/multimodal/README.md.
  • Added PDL support for userbuffer based AllReduce-Norm fusion kernel.
  • Added runtime support for seamless lookahead decoding.
  • Added token-aligned arbitrary output tensors support for the C++ executor API.

API Changes

  • [BREAKING CHANGE] KV cache reuse is enabled automatically when paged_context_fmha is enabled.
  • Added --concurrency support for the throughput subcommand of trtllm-bench.

Fixed Issues

Infrastructure Changes

  • The base Docker image for TensorRT-LLM is updated to nvcr.io/nvidia/pytorch:25.01-py3.
  • The base Docker image for TensorRT-LLM Backend is updated to nvcr.io/nvidia/tritonserver:25.01-py3.
  • The dependent TensorRT version is updated to 10.8.0.
  • The dependent CUDA version is updated to 12.8.0.
  • The dependent ModelOpt version is updated to 0.23 for Linux platform, while 0.17 is still used on Windows platform.

@zeroepoch zeroepoch merged commit 71213f7 into rel Jan 30, 2025
@zeroepoch zeroepoch deleted the preview/rel branch January 30, 2025 21:32
@aikitoria
Copy link

This release mentions min_p parameter but it is nowhere to be found in the actual code.

@juney-nvidia
Copy link
Collaborator

This release mentions min_p parameter but it is nowhere to be found in the actual code.

Sorry for the mis-leading. The min_p support is already done in our internal repo and not merged timely into the 0.17 release. It will be part of the next release. I just updated the release announcement to make it more precise. Thanks for the reminder!

June

@aikitoria
Copy link

I see. Perhaps the main branch could be updated with the new preview code ahead of the next release? 👀

@MahmoudAshraf97
Copy link
Contributor

Are you sure you included the enc dec FP8 in the readme? Because it's not in the 0.17 branch or the main branch

@juney-nvidia
Copy link
Collaborator

I see. Perhaps the main branch could be updated with the new preview code ahead of the next release? 👀

Thanks for the suggestion. The main branch will be updated later in February with incorporating the min_p change(BTW, thanks for your contribution of min_p to TRT-LLM which is the foundation of this feature support in TRT-LLM).

The reason that this time the release/0.17 branch contains more new features than main is due to the Blackwell release which makes the github update process a little bit special :)

Thanks
June

@juney-nvidia
Copy link
Collaborator

Are you sure you included the enc dec FP8 in the readme? Because it's not in the 0.17 branch or the main branch

Thanks for catching this, @MahmoudAshraf97 . I just communicated with the team and there are something wrong about the doc update process for 0.17 release. We just updated the 0.17 release info now. Pls help review it to see whether there is anything else incorrect.

Thanks again for reminding us about this doc issue.

June

kaiyux pushed a commit that referenced this pull request Feb 11, 2025
Co-authored-by: Kaiyu Xie <[email protected]>

open source f8c0381a2bc50ee2739c3d8c2be481b31e5f00bd (#2736)

Co-authored-by: Kaiyu Xie <[email protected]>

Add note for blackwell (#2742)

Update the docs to workaround the extra-index-url issue (#2744)

update README.md (#2751)

Fix github io pages (#2761)

Update
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants