GitHub / triton-inference-server/tensorrtllm_backend issues and pull requests
Labelled with: feature request
#202 - Add usage in response like openai?
Issue -
State: open - Opened by npuichigo almost 2 years ago
- 8 comments
Labels: feature request
#105 - Feature request: support multiple model instances on TensorRT LLM triton backend.
Issue -
State: closed - Opened by wengsnow about 2 years ago
- 16 comments
Labels: triaged, feature request
#89 - Feature request: Output only generated text
Issue -
State: open - Opened by jiangshining about 2 years ago
- 3 comments
Labels: triaged, feature request
#87 - Feature request: Flag indicate end of stream
Issue -
State: open - Opened by yunfeng-scale about 2 years ago
- 2 comments
Labels: triaged, feature request
#58 - feature request: Expose logprob of output tokens
Issue -
State: closed - Opened by yunfeng-scale about 2 years ago
- 3 comments
Labels: triaged, feature request