-
Notifications
You must be signed in to change notification settings - Fork 2k
Pull requests: NVIDIA/TensorRT-LLM
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[None][feat] Support Ulysses CP for torch flow
#10276
opened Dec 24, 2025 by
DylanChen-NV
Loading…
1 task
[https://nvbugs/5594703][infra] Unwaive the failed case to test
#10275
opened Dec 24, 2025 by
EmmaQiaoCh
Loading…
1 task done
[TRTLLM-10029][scheduler] Re-implement MicroBatchScheduler and CapacityScheduler in Python
#10273
opened Dec 24, 2025 by
lancelly
Loading…
[https://nvbugs/5760740][fix] Enable ray tests
#10272
opened Dec 24, 2025 by
shuyixiong
•
Draft
1 task
[https://nvbugs/5766986][fix] fixed the shard_all_unprocessed default value to align with the default.yml
#10271
opened Dec 24, 2025 by
MrGeva
Loading…
1 task done
[None][fix] Add ISOLATION flag for phi4mm MMMU test
#10270
opened Dec 24, 2025 by
Wanli-Jiang
•
Draft
1 task
[https://nvbugs/5701425][chore] Unwaive tests.
#10269
opened Dec 24, 2025 by
yuxianq
Loading…
1 task done
[TRTLLM-7877][test] Add Deepseek V3.2 Test cases and adjust slurm log directory
#10268
opened Dec 24, 2025 by
fredricz-20070104
Loading…
[https://nvbugs/5760737][test] only skip mooncake+indexerkcache test
#10266
opened Dec 24, 2025 by
zhengd-nv
Loading…
1 task done
[TRTLLM-10022][feat] Add hopper xqa decode support for skip softmax attention
#10264
opened Dec 24, 2025 by
pengbowang-nv
Loading…
1 task done
[TRTLLM-10143][feat] Reuse previous draft requests if possible
#10263
opened Dec 24, 2025 by
ziyixiong-nv
Loading…
1 task
[None][test] Add disag-serving auto scaling qa test
#10262
opened Dec 24, 2025 by
StanleySun639
Loading…
1 task done
[None][feat] update trtllm-gen to support groupsTokensHeadsQ
#10261
opened Dec 24, 2025 by
PerkzZheng
•
Draft
1 task done
[https://nvbugs/5740359][chore] Unwaive tests.
#10260
opened Dec 24, 2025 by
yuxianq
Loading…
1 task done
[https://nvbugs//5584607][fix] Ray supports nixl backend
#10259
opened Dec 24, 2025 by
chuangz0
Loading…
1 task done
[None][feat] Drop non-deepgemm fp8 block scale gemm
#10256
opened Dec 24, 2025 by
lucifer1004
Loading…
1 task done
[#10244][feat] AutoDeploy: separate prefill/decode in flashinfer
#10252
opened Dec 24, 2025 by
lucaslie
Loading…
1 task done
[None][feat] Not CUDA graph captured eagle3 one-model draft loop
#10251
opened Dec 24, 2025 by
jhaotingc
Loading…
1 task done
[None][chore] Upgrade transformers to 4.57.3
#10250
opened Dec 24, 2025 by
nv-guomingz
Loading…
1 task done
[#8391][chore] added llama_v3.3_70b_instruct AutoDeploy perf test to L0
#10242
opened Dec 23, 2025 by
MrGeva
Loading…
1 task done
Previous Next
ProTip!
Exclude everything labeled
bug with -label:bug.