An open API service for providing issue and pull request metadata for open source projects.

GitHub / triton-inference-server/tensorrtllm_backend issues and pull requests

Labelled with: feature request

#202 - Add usage in response like openai?

Issue - State: open - Opened by npuichigo almost 2 years ago - 8 comments
Labels: feature request

#105 - Feature request: support multiple model instances on TensorRT LLM triton backend.

Issue - State: closed - Opened by wengsnow about 2 years ago - 16 comments
Labels: triaged, feature request

#89 - Feature request: Output only generated text

Issue - State: open - Opened by jiangshining about 2 years ago - 3 comments
Labels: triaged, feature request

#87 - Feature request: Flag indicate end of stream

Issue - State: open - Opened by yunfeng-scale about 2 years ago - 2 comments
Labels: triaged, feature request

#58 - feature request: Expose logprob of output tokens

Issue - State: closed - Opened by yunfeng-scale about 2 years ago - 3 comments
Labels: triaged, feature request