Skip to content

Actions: triton-inference-server/server

pre-commit

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
1,985 workflow runs
1,985 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

feat: Support LLM API backend in OpenAI frontend
pre-commit #2943: Pull request #8023 opened by krishung5
February 20, 2025 01:25 45s krish-llmapi-openai
February 20, 2025 01:25 45s
ci: Fix L0_batch related flaky tests
pre-commit #2942: Pull request #7999 synchronize by yinggeh
February 19, 2025 21:07 1m 2s yinggeh-DLIS-7826-fix-L0-batcher
February 19, 2025 21:07 1m 2s
Commit for Draft Changes
pre-commit #2941: Pull request #8018 synchronize by indrajit96
February 19, 2025 16:03 25s ibhosale_grpc_callback
February 19, 2025 16:03 25s
feat: ORCA Format KV Cache Utilization in Inference Response Header
pre-commit #2940: Pull request #7839 synchronize by BenjaminBraunDev
February 18, 2025 21:24 Action required BenjaminBraunDev:r24.10
February 18, 2025 21:24 Action required
Commit for Draft Changes
pre-commit #2939: Pull request #8018 opened by indrajit96
February 18, 2025 20:11 54s ibhosale_grpc_callback
February 18, 2025 20:11 54s
refactor: Update the response queue in the server to reuse response slots
pre-commit #2929: Pull request #7879 synchronize by pskiran1
February 16, 2025 19:34 3m 6s spolisetty_dlis_7657
February 16, 2025 19:34 3m 6s
[build] Bumping vllm version to 0.7.0 (#7978)
pre-commit #2925: Pull request #8015 opened by nv-kmcgill53
February 14, 2025 23:20 3m 10s cherry-pick-vllm-0.7.0
February 14, 2025 23:20 3m 10s
fix: build-secret flag not being set breaking build.py
pre-commit #2924: Pull request #7993 synchronize by BenjaminBraunDev
February 14, 2025 22:15 48s BenjaminBraunDev:main
February 14, 2025 22:15 48s
feat: ORCA Format KV Cache Utilization in Inference Response Header
pre-commit #2921: Pull request #7839 synchronize by BenjaminBraunDev
February 14, 2025 19:04 Action required BenjaminBraunDev:r24.10
February 14, 2025 19:04 Action required
feat: ORCA Format KV Cache Utilization in Inference Response Header
pre-commit #2920: Pull request #7839 synchronize by BenjaminBraunDev
February 14, 2025 18:56 Action required BenjaminBraunDev:r24.10
February 14, 2025 18:56 Action required