Ecosyste.ms: Issues
An open API service for providing issue and pull request metadata for open source projects.
GitHub / horseee/LLM-Pruner issues and pull requests
#78 - 关于consecutive_groups
Issue -
State: open - Opened by VincentZ-2020 7 days ago
#77 - UnicodeDecodeError: 'utf-8' codec can't decode byte 0x8b in position 1
Issue -
State: closed - Opened by BrownTan 17 days ago
- 4 comments
#76 - Taylor pruner under-utilizing resources
Issue -
State: open - Opened by marianbasti 18 days ago
#75 - Creating custom configuration files in hgging face format
Issue -
State: open - Opened by sriyachakravarthy 19 days ago
#74 - No such file or directory: pytorch_model.bin
Issue -
State: open - Opened by yaolu-zjut 22 days ago
- 2 comments
#73 - Can the pruned model be converted into a gguf file?
Issue -
State: open - Opened by pxrgod 22 days ago
#72 - Custom Model pruning
Issue -
State: open - Opened by saidineshpola 23 days ago
#71 - Does it support qwen2?
Issue -
State: open - Opened by yangxue-1 24 days ago
- 1 comment
#70 - Update ppl_dataset.py with trust_remote_code
Pull Request -
State: open - Opened by manickavela29 26 days ago
- 1 comment
#69 - Llama3 reports shape error after pruning
Issue -
State: open - Opened by WentaoTan 28 days ago
- 4 comments
#68 - Loading pruned model for causal llm
Issue -
State: open - Opened by sriyachakravarthy about 1 month ago
#67 - Update hf_prune.py
Pull Request -
State: open - Opened by aritralegndery about 1 month ago
- 1 comment
#66 - I would like to ask if the current version is suitable for qwen.
Issue -
State: open - Opened by wangxiaoxue about 2 months ago
- 3 comments
#65 - Support GQA &Llama 3 & Llama 3.1.
Pull Request -
State: closed - Opened by horseee about 2 months ago
#64 - Adaptation of GQA
Issue -
State: open - Opened by junzhang-zj about 2 months ago
- 7 comments
#63 - No pytorch_model.bin file in the tune_log/llama_0.2/checkpoint-200 folder
Issue -
State: open - Opened by hebowei2000 3 months ago
- 3 comments
#62 - 请问能裁剪普通的transformer模型吗
Issue -
State: open - Opened by SKY072410 3 months ago
#61 - 请问可以支持chatglm3剪枝吗
Issue -
State: open - Opened by Franklin-L 3 months ago
#60 - Difference in Perplexity Values
Issue -
State: open - Opened by nikhil-ghosh-berkeley 4 months ago
#59 - No random seed Settings found in post_training.py
Issue -
State: open - Opened by JunKong5 4 months ago
#58 - Evaluation:UnicodeDecodeError: 'utf-8' codec can't decode byte 0x8b in position 1: invalid start byte
Issue -
State: open - Opened by manlenzzz 5 months ago
- 1 comment
#57 - Pruning llama3
Issue -
State: open - Opened by yinwangsong 5 months ago
#56 - I tired Mistral 7b model, but I got this issue
Issue -
State: open - Opened by TejasLidhure 5 months ago
#55 - How to prune the embedding and lm_head?
Issue -
State: open - Opened by L-hongbin 6 months ago
#54 - Is this method implementable on multi-GPUs?
Issue -
State: open - Opened by LeonCheng0129 6 months ago
#53 - RecursionError: maximum recursion depth exceeded
Issue -
State: open - Opened by Zhenyu001225 6 months ago
- 2 comments
#52 - Unable to reproduce the results for param_first and param_second in the paper after finetuning.
Issue -
State: open - Opened by danyal97 7 months ago
#51 - 剪枝率值的问题
Issue -
State: closed - Opened by qxpBlog 8 months ago
#50 - 延迟评估
Issue -
State: closed - Opened by qxpBlog 8 months ago
- 4 comments
#49 - The quantization of the compressed models
Issue -
State: open - Opened by lihuang258 8 months ago
#48 - ConnectionError: Couldn't reach https://raw.githubusercontent.com/wojzaremba/lstm/master/data/ptb.train.txt (ReadTimeout(ReadTimeoutError("HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Read timed out. (read timeout=100)")))
Issue -
State: closed - Opened by qxpBlog 8 months ago
- 2 comments
#47 - OSError: Can't load tokenizer for 'baffo32/decapoda-research-llama-7B-hf'.
Issue -
State: closed - Opened by qxpBlog 8 months ago
- 2 comments
#46 - Cannot use huggface to load
Issue -
State: open - Opened by coderchem 9 months ago
#45 - Issue: Missing Generation of `pytorch_model.bin` File During Model Tuning
Issue -
State: closed - Opened by WilliamYi96 9 months ago
- 5 comments
#44 - cannot import name 'SiLUActivation' from 'transformers.activations'
Issue -
State: closed - Opened by azuryl 9 months ago
- 1 comment
#43 - 401 Client Error: Unauthorized for url: https://huggingface.co/decapoda-research/llama-7b-hf/resolve/main/tokenizer_config.json
Issue -
State: open - Opened by azuryl 9 months ago
- 1 comment
#42 - Adding a tutorial for adapting new models?
Issue -
State: open - Opened by zhichaoxu-shufe 10 months ago
#41 - 在将部分层进行剪枝之后,不能直接通过tgi加载模型
Issue -
State: open - Opened by coderchem 10 months ago
#40 - Pruning MQA?
Issue -
State: open - Opened by jianyuheng 10 months ago
#39 - Question related to the model tuning
Issue -
State: open - Opened by shawnricecake 10 months ago
- 2 comments
#38 - 为什么num_examples默认是10?
Issue -
State: open - Opened by coderchem 10 months ago
- 2 comments
#37 - the new pytorch.bin is bigger than original model issue
Issue -
State: open - Opened by lb553024300 10 months ago
- 4 comments
#36 - Supporting device_map = 'auto' similar to the one in .from_pretrained method from Huggingface
Issue -
State: open - Opened by Ahmed-Roushdy 10 months ago
- 3 comments
#35 - a post-training issue
Issue -
State: open - Opened by cmnfriend 10 months ago
- 2 comments
#34 - Reproducing paper results
Issue -
State: open - Opened by grigorn 10 months ago
- 6 comments
#33 - Latency code
Issue -
State: closed - Opened by tuidan 10 months ago
- 2 comments
#32 - Can not import LlamaConfig
Issue -
State: closed - Opened by Ahmed-Roushdy 11 months ago
- 1 comment
#31 - Examples on the Huggingface Hub
Issue -
State: open - Opened by vgoklani 11 months ago
#30 - When will you support ChatGLM?
Issue -
State: open - Opened by AboveParadise 11 months ago
#29 - Force even pruning across layers
Issue -
State: open - Opened by thedarkzeno 12 months ago
- 1 comment
#28 - Calculating Importance of 'param_mix'
Issue -
State: closed - Opened by kiucho 12 months ago
- 2 comments
#27 - When would the code for GPT-J-6B be released?
Issue -
State: closed - Opened by mumuyeye 12 months ago
- 1 comment
#26 - Added ability to load custom dataset
Pull Request -
State: closed - Opened by michaelvin1322 about 1 year ago
#25 - Evaluation metric (acc vs. acc_norm) for lm-evaluation-harness tasks
Issue -
State: closed - Opened by bokyeong1015 about 1 year ago
- 4 comments
#24 - Sparse Mask question
Issue -
State: closed - Opened by coldplayers about 1 year ago
- 2 comments
#23 - param_first and param_mix result the same ppl
Issue -
State: open - Opened by Kausal-Lei about 1 year ago
- 1 comment
#22 - Error occurs when pruning LLaMa2-7b
Issue -
State: closed - Opened by moonlightian about 1 year ago
- 6 comments
#21 - I encountered the following error message when I assign iterative_steps = 2 during baichuan-7B pruning
Issue -
State: open - Opened by yangd85 about 1 year ago
- 1 comment
Labels: bug
#20 - Adding quantization
Issue -
State: open - Opened by Duncan1115 about 1 year ago
- 9 comments
#19 - evaluate
Issue -
State: open - Opened by StevensPrime about 1 year ago
- 1 comment
#18 - Pruning Llama2-7B
Issue -
State: open - Opened by acalatrava about 1 year ago
- 4 comments
#17 - recover training
Issue -
State: open - Opened by xcg940123 about 1 year ago
- 1 comment
#16 - Can prune model convert to llama.cpp ggml?
Issue -
State: open - Opened by shaonianyr about 1 year ago
- 1 comment
#15 - Checking the pruned but uncompressed model
Issue -
State: open - Opened by ZN1010 about 1 year ago
- 10 comments
#14 - Code for evaluation on zero-shot tasks using lm-evaluation-harness
Issue -
State: closed - Opened by jameslahm about 1 year ago
- 2 comments
#13 - Eval Loss NaN on Llama-2
Issue -
State: open - Opened by mmichaelzhang about 1 year ago
- 3 comments
Labels: question
#12 - hi, Does post_training support full parameter fine-tuning of the pruned model?
Issue -
State: open - Opened by StevensPrime about 1 year ago
- 1 comment
#11 - Use LLM-Pruner for Baichuan model
Issue -
State: closed - Opened by Daisy5296 about 1 year ago
- 15 comments
Labels: bug
#10 - Zero-shot Evaluation
Issue -
State: closed - Opened by Duncan1115 about 1 year ago
- 2 comments
#9 - Gain using more data
Issue -
State: closed - Opened by Daisy5296 about 1 year ago
- 5 comments
#8 - When would the code for ChatGLM be released?
Issue -
State: open - Opened by moonlightian about 1 year ago
- 1 comment
#7 - RuntimeError of test_speedup.py
Issue -
State: closed - Opened by Daisy5296 about 1 year ago
- 9 comments
#6 - my process have some problems
Issue -
State: closed - Opened by 18140663659 about 1 year ago
- 1 comment
#5 - this method can be used for bloom?
Issue -
State: open - Opened by 18140663659 about 1 year ago
- 5 comments
Labels: enhancement
#4 - Error when using GPU for pruning
Issue -
State: closed - Opened by Daisy5296 about 1 year ago
- 3 comments
#3 - Reload the prunered Model failed
Issue -
State: closed - Opened by Zhuqln over 1 year ago
- 8 comments
#2 - Update README.md
Pull Request -
State: closed - Opened by eltociear over 1 year ago
#1 - Question on recovery and training data
Issue -
State: closed - Opened by Qubitium over 1 year ago
- 2 comments