Skip to content

Actions: NVIDIA/TensorRT

Actions

Blossom-CI

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
5,739 workflow runs
5,739 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

How to remove signal and wait layer in the engine?
Blossom-CI #6360: Issue comment #4232 (comment) created by lijinghaooo
November 5, 2024 07:52 5s
November 5, 2024 07:52 5s
Blossom-CI
Blossom-CI #6359: created by lijinghaooo
November 5, 2024 07:39 5s
November 5, 2024 07:39 5s
Tensor Parallel and Context Parallel
Blossom-CI #6358: Issue comment #4231 (comment) created by lix19937
November 5, 2024 03:23 5s
November 5, 2024 03:23 5s
State of affairs for NestedTensor (NJT) inference?
Blossom-CI #6357: Issue comment #4234 (comment) created by lix19937
November 5, 2024 03:19 5s
November 5, 2024 03:19 5s
How to convert model.onnx and model.onnx_data to trt model
Blossom-CI #6356: Issue comment #4235 (comment) created by Sdamuu
November 5, 2024 02:53 5s
November 5, 2024 02:53 5s
How to remove signal and wait layer in the engine?
Blossom-CI #6355: Issue comment #4232 (comment) created by lix19937
November 5, 2024 02:12 4s
November 5, 2024 02:12 4s
How to convert model.onnx and model.onnx_data to trt model
Blossom-CI #6354: Issue comment #4235 (comment) created by lix19937
November 5, 2024 00:32 4s
November 5, 2024 00:32 4s
November 4, 2024 09:06 4s
Given an engine file, how to know what GPU model it is generated on?
Blossom-CI #6352: Issue comment #4233 (comment) created by lix19937
November 4, 2024 01:42 4s
November 4, 2024 01:42 4s
Given an engine file, how to know what GPU model it is generated on?
Blossom-CI #6351: Issue comment #4233 (comment) created by yangdong02
November 4, 2024 01:31 4s
November 4, 2024 01:31 4s
Given an engine file, how to know what GPU model it is generated on?
Blossom-CI #6350: Issue comment #4233 (comment) created by lix19937
November 4, 2024 00:02 3s
November 4, 2024 00:02 3s
Optimize Dynamic Shape Inference for TTS Model with HiFi-GAN Vocoder
Blossom-CI #6349: Issue comment #4230 (comment) created by AntixK
November 3, 2024 11:08 4s
November 3, 2024 11:08 4s
Support UINT8 input / output and casting from UINT8 to FP16 and back
Blossom-CI #6348: Issue comment #3026 (comment) created by jax11235
November 2, 2024 15:27 5s
November 2, 2024 15:27 5s
TensorRT 10.5 Flux Dit BF16 precision
Blossom-CI #6347: Issue comment #4215 (comment) created by asfiyab-nvidia
November 1, 2024 17:15 4s
November 1, 2024 17:15 4s
November 1, 2024 13:39 4s
November 1, 2024 10:30 4s
wrong results of TensorRT 10.0 when running on GPU Tesla T4
Blossom-CI #6344: Issue comment #3999 (comment) created by yflv-yanxia
November 1, 2024 10:11 4s
November 1, 2024 10:11 4s
wrong results of TensorRT 10.0 when running on GPU Tesla T4
Blossom-CI #6343: Issue comment #3999 (comment) created by yuanyao-nv
November 1, 2024 05:38 4s
November 1, 2024 05:38 4s
Engine built failure of TensorRT 10.5 when running kSTRONGLY_TYPED model on GPU A10
Blossom-CI #6342: Issue comment #4228 (comment) created by tp-nan
November 1, 2024 03:45 4s
November 1, 2024 03:45 4s
November 1, 2024 03:34 5s
Engine built failure of TensorRT 10.5 when running kSTRONGLY_TYPED model on GPU A10
Blossom-CI #6340: Issue comment #4228 (comment) created by tp-nan
November 1, 2024 02:50 4s
November 1, 2024 02:50 4s
ERROR: No matching distribution found for numpy>=2.0
Blossom-CI #6339: Issue comment #4025 (comment) created by AnonymousPls
November 1, 2024 02:40 4s
November 1, 2024 02:40 4s
wrong results of TensorRT 10.0 when running on GPU Tesla T4
Blossom-CI #6338: Issue comment #3999 (comment) created by yflv-yanxia
November 1, 2024 02:20 4s
November 1, 2024 02:20 4s