An open API service for providing issue and pull request metadata for open source projects.

GitHub / xusenlinzy/api-for-open-llm issues and pull requests

#318 - 大佬,这个项目还更新吗?或者有啥平替的项目吗

Issue - State: open - Opened by FreeRotate 6 months ago - 4 comments

#317 - 运行qwen:7b出现错误,该如何解决?

Issue - State: open - Opened by test16553 7 months ago - 2 comments

#315 - 4*4090 显卡部署glm4-9b 使用dify 的api调用报错

Issue - State: open - Opened by he498 10 months ago - 1 comment

#313 - support cogvlm2 model

Pull Request - State: closed - Opened by white2018 10 months ago

#312 - Support MiniMonkey model

Pull Request - State: closed - Opened by santlchogva 10 months ago

#311 - 运行glm4v请求报错

Issue - State: open - Opened by 760485464 11 months ago - 2 comments

#310 - 执行streamlit_app.py报错

Issue - State: closed - Opened by louan1998 11 months ago

#309 - not support sglang backend

Issue - State: open - Opened by colinsongf 11 months ago

#307 - 部署gte-qwen2-1.5b-instruct请求rerank接口报错

Issue - State: open - Opened by cowcomic 12 months ago

#306 - vllm 接口支持vision(minicpm-v)

Pull Request - State: open - Opened by baisong666 12 months ago

#303 - 使用Qwen2-7B-Instrut模型出现问题-使用Vllm

Issue - State: closed - Opened by Empress7211 12 months ago - 3 comments

#302 - RuntimeError: CUDA error: device-side assert triggered

Issue - State: closed - Opened by ChaoPeng13 12 months ago

#296 - llama3-8B回答后自我交流,不停止

Issue - State: open - Opened by yd9038074 about 1 year ago - 1 comment

#292 - minicpm启动没问题,推理访问报错

Issue - State: open - Opened by 760485464 about 1 year ago - 2 comments

#291 - glm-4v启动正常 访问推理报错

Issue - State: open - Opened by 760485464 about 1 year ago - 10 comments

#288 - glm4 接入dify后无法触发使用工具

Issue - State: open - Opened by he498 about 1 year ago - 1 comment

#189 - ChatGLM3的输入长度超过8k时建议拦截中断执行

Issue - State: closed - Opened by lzhfe over 1 year ago - 3 comments

#100 - 模型输出<|im_start|> <|im_start|>

Issue - State: closed - Opened by bh4ffu almost 2 years ago - 2 comments

#99 - baichuan 13b 部署后 使用langchain调用报错

Issue - State: closed - Opened by zhouzhou0322 almost 2 years ago - 5 comments

#98 - 启动报错 TypeError: issubclass() arg 1 must be a class

Issue - State: closed - Opened by zhouzhou0322 almost 2 years ago - 5 comments

#97 - codellama-34b-instruct-hf 回复内容出现截断

Issue - State: closed - Opened by anyshu almost 2 years ago - 3 comments

#96 - Fixed baremetal startup process

Pull Request - State: closed - Opened by wey-gu almost 2 years ago

#95 - 流式接口输出,role为null

Issue - State: closed - Opened by bh4ffu almost 2 years ago

#92 - 13B模型量化加载提示显存溢出

Issue - State: closed - Opened by chelovek21 almost 2 years ago - 2 comments

#91 - 新代码运行报错

Issue - State: closed - Opened by xsun15 almost 2 years ago - 2 comments

#90 - [help]考虑支持一下codellama吧?

Issue - State: closed - Opened by bh4ffu almost 2 years ago - 4 comments
Labels: enhancement

#89 - ValueError: Out of range float values are not JSON compliant

Issue - State: closed - Opened by 143heyan almost 2 years ago - 4 comments

#88 - usage.first_tokens = content["usage"].get("first_tokens", None)

Issue - State: closed - Opened by uulichen almost 2 years ago - 5 comments

#86 - 对于qwen模型,在使用completion接口时,decode参数存在的问题

Issue - State: closed - Opened by nlfiasel almost 2 years ago - 1 comment

#85 - Unable to find image 'llm-api:vllm' locally

Issue - State: closed - Opened by xiechengmude almost 2 years ago - 1 comment

#84 - "GET /v1 HTTP/1.1" 404 Not Found

Issue - State: closed - Opened by YunFenLei almost 2 years ago - 1 comment

#83 - Qwen model outputs are different from Vllm server vs normal server.

Issue - State: closed - Opened by monksgoyal almost 2 years ago - 4 comments

#82 - 💡 [REQUEST] - 是否支持chatglm2的多轮对话?

Issue - State: closed - Opened by huanglx27 almost 2 years ago - 1 comment
Labels: question

#81 - Update requirements.txt

Pull Request - State: closed - Opened by luchenwei9266 almost 2 years ago

#80 - 模型输出没有完毕前,客户端停止输出,服务端报错。

Issue - State: closed - Opened by jinghai almost 2 years ago - 10 comments

#79 - starchat模型使用vllm推理乱码

Issue - State: closed - Opened by skingko almost 2 years ago - 3 comments

#78 - 试了下,显存占用明显增加,模型用的Qwen-7B-chat,不知道啥原因

Issue - State: closed - Opened by bh4ffu almost 2 years ago - 2 comments

#76 - 可不可以建个群交流啊

Issue - State: closed - Opened by queensking almost 2 years ago - 2 comments

#75 - VLLM运行Qwen报错

Issue - State: closed - Opened by jinghai almost 2 years ago - 8 comments

#74 - 💡 [REQUEST] - <title>建议增加并发量配置

Issue - State: closed - Opened by jinghai almost 2 years ago
Labels: question

#73 - 构建vllm的镜像卡在 Installing build dependencies: started

Issue - State: closed - Opened by bh4ffu almost 2 years ago - 4 comments

#72 - 💡 [REQUEST] - <title>可以单独配置embedding模型使用CPU资源吗?

Issue - State: closed - Opened by jinghai almost 2 years ago - 2 comments
Labels: question

#71 - 可以提供一个公共的api供其他电脑使用吗?

Issue - State: closed - Opened by 15899885850 almost 2 years ago - 2 comments

#70 - 关于模型异步调用的问题

Issue - State: closed - Opened by Isfate almost 2 years ago - 2 comments

#69 - vllm下,启动Qwen-7B-Chat 报错

Issue - State: closed - Opened by xcpuma almost 2 years ago - 6 comments

#68 - 💡 [REQUEST] - <title>QWen模型流失问答输出个不停,不知道什么问题

Issue - State: closed - Opened by jinghai almost 2 years ago - 20 comments
Labels: question

#67 - Merge pull request #1 from xusenlinzy/master

Pull Request - State: closed - Opened by xysnqdd almost 2 years ago

#65 - 使用vllm推理starchat报错,错误如下:

Issue - State: closed - Opened by foxxxx001 almost 2 years ago - 11 comments

#64 - Dev

Pull Request - State: closed - Opened by xusenlinzy almost 2 years ago

#62 - Merge pull request #61 from xusenlinzy/master

Pull Request - State: closed - Opened by xusenlinzy almost 2 years ago

#61 - Fix protocol, improve react code, support stream mode for function call

Pull Request - State: closed - Opened by xusenlinzy almost 2 years ago

#60 - vllm启动方式添加embedding模型报错

Issue - State: closed - Opened by youzhonghui almost 2 years ago - 1 comment
Labels: bug, help wanted

#59 - 安装成功,调用几次就会报错。

Issue - State: closed - Opened by Hkaisense almost 2 years ago - 2 comments
Labels: environment

#58 - Torch not compiled with CUDA enabled

Issue - State: closed - Opened by happyfire almost 2 years ago - 3 comments
Labels: environment

#57 - My Qwen LLM startup failed

Issue - State: closed - Opened by Hkaisense almost 2 years ago - 7 comments
Labels: solved

#56 - 求助:Baichuan-13b-chat运行报错

Issue - State: closed - Opened by happyfire almost 2 years ago - 2 comments

#54 - tiktoken.model.encoding_for_model 需要联网?

Issue - State: closed - Opened by TheBobbyliu almost 2 years ago - 5 comments

#53 - 建议向量模型考虑增加bge-large-zh

Issue - State: closed - Opened by berwinjoule almost 2 years ago - 1 comment
Labels: solved

#52 - stream模式下,function call 不兼容

Issue - State: closed - Opened by zhengxiang5965 almost 2 years ago - 4 comments

#51 - 为什么每次回答的内容都不一样阿

Issue - State: closed - Opened by lucheng07082221 almost 2 years ago - 1 comment

#50 - docker启动没有报错,但是端口没有监听,无法访问

Issue - State: closed - Opened by zengzhenhui almost 2 years ago - 2 comments

#48 - 希望并发处理多个请求

Issue - State: closed - Opened by Huangyajuan-123 almost 2 years ago - 3 comments
Labels: enhancement

#47 - 千问的加速依赖是要加到docker中吗?

Issue - State: closed - Opened by jinghai almost 2 years ago - 2 comments

#46 - 报错:更新了最新代码后报错

Issue - State: closed - Opened by jinghai almost 2 years ago - 5 comments

#45 - 建议增加llama2系列模型的call function功能

Issue - State: closed - Opened by skingko almost 2 years ago - 3 comments
Labels: enhancement

#44 - Update prompt_adapter.py 修改Qwen bug

Pull Request - State: closed - Opened by markliuyuxiang almost 2 years ago - 1 comment

#43 - chat/completions和completions接口的结果差异较大

Issue - State: closed - Opened by TheBobbyliu almost 2 years ago - 4 comments

#42 - 推理速度疑问(很快)

Issue - State: closed - Opened by onlyfish79 almost 2 years ago - 8 comments

#41 - 可以支持llama2-hf么?

Issue - State: closed - Opened by Smile-L almost 2 years ago - 2 comments

#39 - Support new hope model

Issue - State: closed - Opened by 2214962083 almost 2 years ago - 2 comments

#38 - 尝试加载本地firefly-baichuan13b的模型时显存占用翻倍

Issue - State: closed - Opened by bswaterb about 2 years ago - 3 comments

#36 - 为什么不支持word呢?

Issue - State: closed - Opened by lucheng07082221 about 2 years ago - 2 comments

#35 - 知识库支持上传文件夹吗?

Issue - State: closed - Opened by lucheng07082221 about 2 years ago - 2 comments

#34 - 访问出错

Issue - State: closed - Opened by lucheng07082221 about 2 years ago - 2 comments

#32 - 请问如何使用指定的卡

Issue - State: closed - Opened by jinghai about 2 years ago - 1 comment

#31 - 对baichuan 13b chat的代码存在疑问。

Issue - State: closed - Opened by askintution about 2 years ago - 3 comments

#30 - 建议增加文心一言和星火商业模型的支持

Issue - State: closed - Opened by jinghai about 2 years ago - 1 comment

#28 - 添加知识库后报错

Issue - State: closed - Opened by whh1009 about 2 years ago - 4 comments