Ecosyste.ms: Issues

An open API service for providing issue and pull request metadata for open source projects.

GitHub / InternLM/lmdeploy issues and pull requests

#2129 - [Bug] `slice_image` attribute error with MiniCPM-Llama3-V-2_5

Issue - State: open - Opened by thisiskofi about 2 months ago - 12 comments
Labels: awaiting response, Stale, mllm

#2125 - [Bug] topk (43930832) is larger

Issue - State: closed - Opened by WCwalker about 2 months ago - 3 comments

#2120 - [Bug] Gradio serve不能正确识别chat template

Issue - State: closed - Opened by cmpute about 2 months ago - 4 comments

#2117 - [Bug] Llama 3.1 Support

Issue - State: open - Opened by vladrad about 2 months ago - 18 comments
Labels: awaiting response

#2110 - 如何最大化pipline batch infer的吞吐量/速度[Bug]

Issue - State: closed - Opened by hitzhu about 2 months ago - 3 comments
Labels: awaiting response, Stale

#2104 - Refactor pytorch engine

Pull Request - State: closed - Opened by grimoire about 2 months ago - 21 comments
Labels: enhancement

#2103 - Renew a session for reset button

Pull Request - State: closed - Opened by AllentDan about 2 months ago

#2103 - Renew a session for reset button

Pull Request - State: closed - Opened by AllentDan about 2 months ago

#2101 - [Bug] serve的时候event loop报错

Issue - State: open - Opened by cmpute about 2 months ago - 27 comments

#2101 - [Bug] serve的时候event loop报错

Issue - State: open - Opened by cmpute about 2 months ago - 27 comments

#2095 - [Bug] glm-4v=9b速度特别慢

Issue - State: closed - Opened by bltcn 2 months ago - 2 comments

#2090 - New GEMM kernels for weight-only quantization

Pull Request - State: closed - Opened by lzhangzz 2 months ago - 14 comments
Labels: enhancement

#2084 - Add user guide about slora serving

Pull Request - State: closed - Opened by AllentDan 2 months ago - 1 comment
Labels: documentation

#2060 - [Bug] 使用官方镜像v0.5.1进行GLM4v的部署会有一个报错

Issue - State: closed - Opened by ZhiyuYUE 2 months ago - 8 comments
Labels: awaiting response, Stale

#2057 - 想问下Lmdeploy支持base model加多lora的部署方式么

Issue - State: closed - Opened by will-wiki 2 months ago - 11 comments
Labels: awaiting response, Stale

#2052 - Florence 2 support :)

Issue - State: open - Opened by SinanAkkoyun 2 months ago - 2 comments

#2050 - 多卡部署internvl2-8b报错

Issue - State: closed - Opened by haoduoyu1203 2 months ago - 5 comments

#2046 - Support custom attention backend

Pull Request - State: closed - Opened by grimoire 2 months ago

#2043 - [Bug] gradio reset button stucked after I cancel a response.

Issue - State: closed - Opened by zhulinJulia24 2 months ago - 2 comments

#2038 - Reorganize the user guide and update the get_started section

Pull Request - State: closed - Opened by lvhan028 2 months ago
Labels: documentation

#2036 - [Bug] lmdeploy has other questions about server for lora_merge_model

Issue - State: closed - Opened by Volta-lemon 2 months ago - 20 comments

#2036 - [Bug] lmdeploy has other questions about server for lora_merge_model

Issue - State: closed - Opened by Volta-lemon 2 months ago - 20 comments

#2018 - Add prefix cache stats to usage

Pull Request - State: open - Opened by ispobock 2 months ago - 10 comments

#2001 - [Bug] 访问一段时间后服务卡死/无响应

Issue - State: open - Opened by hezeli123 2 months ago - 13 comments

#1990 - Could not use my local internVL mini model for inference

Issue - State: open - Opened by shiva-vardhineedi 2 months ago - 4 comments

#1989 - [Feature] 我们支持gptq量化模型的推理么

Issue - State: closed - Opened by eigen2017 2 months ago - 15 comments

#1984 - Phi3 awq

Pull Request - State: closed - Opened by grimoire 2 months ago - 4 comments
Labels: enhancement

#1981 - [Bug] MiniCPMV的推理有问题

Issue - State: closed - Opened by LDLINGLINGLING 2 months ago - 17 comments
Labels: awaiting response, Stale

#1970 - [Feature] tubromind有计划支持cogvlm2吗?

Issue - State: closed - Opened by jidechao 2 months ago - 1 comment
Labels: backlog

#1966 - support min_p sampling & do_sample setting

Pull Request - State: closed - Opened by irexyc 2 months ago - 4 comments
Labels: WIP

#1966 - support min_p sampling & do_sample setting

Pull Request - State: closed - Opened by irexyc 2 months ago - 4 comments
Labels: WIP

#1962 - torch engine optimize prefill for long context

Pull Request - State: closed - Opened by grimoire 2 months ago - 8 comments
Labels: improvement

#1931 - Remove deprecated arguments from API and clarify model_name and chat_template_name

Pull Request - State: closed - Opened by lvhan028 3 months ago - 2 comments
Labels: BC-breaking, improvement

#1906 - minicpm-v采用W4A16量化,推理速度没什么变化

Issue - State: closed - Opened by DankoZhang 3 months ago - 14 comments

#1862 - [Bug] 单轮的图文交错对话的实现原理

Issue - State: open - Opened by stay-leave 3 months ago - 1 comment

#1856 - Support guided decoding for pytorch backend

Pull Request - State: closed - Opened by AllentDan 3 months ago - 8 comments
Labels: enhancement

#1846 - 如何指定模型的数据类型为f16

Issue - State: open - Opened by Yang-bug-star 3 months ago - 6 comments

#1846 - 如何指定模型的数据类型为f16

Issue - State: closed - Opened by Yang-bug-star 3 months ago - 9 comments

#1844 - Maybe a workaround for qwen2 quantization Nan error

Pull Request - State: closed - Opened by AllentDan 3 months ago - 5 comments

#1836 - [Bug] qwen2 awq量化微调后的模型报错

Issue - State: open - Opened by qiuxuezhe123 3 months ago - 12 comments

#1831 - [Bug] smoothquant量化Bacihuan2-7B-Chat模型,无法正常量化

Issue - State: closed - Opened by CodexDive 3 months ago - 11 comments

#1826 - [Bug] awq for Qwen2-72B-instruct

Issue - State: open - Opened by Vincent131499 3 months ago - 25 comments

#1815 - [Bug] lmdeploy部署intermlm2-chat-20b,遇到<|im_end|>不会停止

Issue - State: open - Opened by jeinlee1991 3 months ago - 11 comments

#1745 - [Feature] `min_p` sampling parameter

Issue - State: closed - Opened by josephrocca 3 months ago - 4 comments

#1738 - [Feature] Speculative Decoding

Issue - State: open - Opened by josephrocca 3 months ago - 15 comments

#1711 - [Feature] V100量化推理

Issue - State: closed - Opened by QwertyJack 4 months ago - 16 comments
Labels: v100

#1615 - Check base64 image validation

Pull Request - State: closed - Opened by AllentDan 4 months ago - 2 comments
Labels: Bug:P2

#1587 - [Feature] Support W4A8KV4 Quantization(QServe/QoQ)

Issue - State: closed - Opened by wanzhenchn 4 months ago - 3 comments

#1565 - add more model into benchmark and evaluate workflow

Pull Request - State: closed - Opened by zhulinJulia24 4 months ago - 1 comment

#1332 - Add docs of support new vl model

Pull Request - State: closed - Opened by irexyc 6 months ago
Labels: documentation

#1315 - [Feature] 建议训练internlm2-chat-7b 的 GPTQ-4bit 量化模型并支持llmdeploy部署

Issue - State: closed - Opened by wwewwt 6 months ago - 4 comments
Labels: backlog

#1177 - [Bug] 3090 部署internlm2-chat-20b-4bits,提问卡住不懂

Issue - State: closed - Opened by makefree3 7 months ago - 9 comments
Labels: awaiting response, Stale

#1035 - Compatible with Gradio 4.x

Pull Request - State: closed - Opened by irexyc 8 months ago - 1 comment
Labels: WIP, improvement

#759 - [Bug] tried to compile and run on aarch64 with iGPU

Issue - State: closed - Opened by cj401 10 months ago - 6 comments

#671 - [Feature] Request for Support for Ascend Series Graphics Cards

Issue - State: closed - Opened by junior-zsy 11 months ago - 15 comments

#301 - [Bug] Debug version run failed

Issue - State: open - Opened by sleepwalker2017 about 1 year ago - 11 comments

#100 - test issue bot

Issue - State: closed - Opened by tpoisonooo about 1 year ago

#100 - test issue bot

Issue - State: closed - Opened by tpoisonooo about 1 year ago

#99 - Comparasion with vllm

Issue - State: closed - Opened by lucasjinreal about 1 year ago - 15 comments

#99 - Comparasion with vllm

Issue - State: closed - Opened by lucasjinreal about 1 year ago - 13 comments

#98 - Can we have support for GGML as triton backend

Issue - State: closed - Opened by tikikun about 1 year ago - 12 comments

#98 - Can we have support for GGML as triton backend

Issue - State: closed - Opened by tikikun about 1 year ago - 12 comments

#97 - add docstring for turbomind

Pull Request - State: closed - Opened by lvhan028 about 1 year ago

#97 - add docstring for turbomind

Pull Request - State: closed - Opened by lvhan028 about 1 year ago

#96 - 启动时报错

Issue - State: closed - Opened by CocaColaKing about 1 year ago - 1 comment

#96 - 启动时报错

Issue - State: closed - Opened by CocaColaKing about 1 year ago - 1 comment

#95 - I am confused about KV Cache Manager, How does it work?

Issue - State: closed - Opened by randomseed713 about 1 year ago - 3 comments

#95 - I am confused about KV Cache Manager, How does it work?

Issue - State: closed - Opened by randomseed713 about 1 year ago - 3 comments

#94 - set chuk_size=1 and export ‘tp’ to config.ini

Pull Request - State: closed - Opened by lvhan028 about 1 year ago

#94 - set chuk_size=1 and export ‘tp’ to config.ini

Pull Request - State: closed - Opened by lvhan028 about 1 year ago

#93 - [Improve] Add docstrings to pytorch submodule

Pull Request - State: closed - Opened by wangruohui about 1 year ago

#93 - [Improve] Add docstrings to pytorch submodule

Pull Request - State: closed - Opened by wangruohui about 1 year ago

#92 - docs(serving.md): typo

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago

#92 - docs(serving.md): typo

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago

#91 - 支持tritonserver

Issue - State: closed - Opened by coderchem about 1 year ago - 6 comments

#91 - 支持tritonserver

Issue - State: closed - Opened by coderchem about 1 year ago - 6 comments

#90 - 请问未来会支持 Beam search 吗?

Issue - State: closed - Opened by B-201 about 1 year ago - 2 comments

#90 - 请问未来会支持 Beam search 吗?

Issue - State: closed - Opened by B-201 about 1 year ago - 2 comments

#89 - an error about llama-65b

Issue - State: closed - Opened by leizhao1234 about 1 year ago - 5 comments

#89 - an error about llama-65b

Issue - State: closed - Opened by leizhao1234 about 1 year ago - 5 comments

#88 - 请问这个部署支持多卡吗?

Issue - State: closed - Opened by alexw994 about 1 year ago - 2 comments

#88 - 请问这个部署支持多卡吗?

Issue - State: closed - Opened by alexw994 about 1 year ago - 2 comments

#86 - update contribution.md

Pull Request - State: closed - Opened by grimoire about 1 year ago

#86 - update contribution.md

Pull Request - State: closed - Opened by grimoire about 1 year ago

#84 - feat(quantization): kv cache use asymmetric

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago - 1 comment
Labels: enhancement

#84 - feat(quantization): kv cache use asymmetric

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago - 1 comment
Labels: enhancement

#83 - feat(deploy.py): support w pack qkv

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago - 1 comment

#83 - feat(deploy.py): support w pack qkv

Pull Request - State: closed - Opened by tpoisonooo about 1 year ago - 1 comment

#82 - Tensor Parallel python api

Pull Request - State: closed - Opened by grimoire about 1 year ago - 3 comments