Ecosyste.ms: Issues
An open API service for providing issue and pull request metadata for open source projects.
GitHub / juncongmoo/pyllama issues and pull requests
#114 - Question regarding EnergonAI repo
Issue -
State: closed - Opened by philipp-fischer 7 months ago
#113 - torch.distributed.elastic.multiprocessing.errors.ChildFailedError
Issue -
State: open - Opened by sido420 9 months ago
- 1 comment
#112 - Quick Question
Issue -
State: closed - Opened by ArgusK17 9 months ago
#111 - How to run an interactive mode in Jupyter?
Issue -
State: open - Opened by myrainbowandsky 11 months ago
#109 - 12GB card
Issue -
State: open - Opened by arthurwolf about 1 year ago
- 2 comments
#108 - no module named llama
Issue -
State: open - Opened by Cooper-Ji about 1 year ago
- 1 comment
#107 - Added transformers to requirements.txt
Pull Request -
State: open - Opened by HireTheHero over 1 year ago
#106 - NVMLError_NoPermission: Insufficient Permissions
Issue -
State: open - Opened by mzdsk2 over 1 year ago
#105 - evaluating has an extremely large value when quantize to 4bit.
Issue -
State: open - Opened by JiachuanDENG over 1 year ago
- 1 comment
#104 - Download 7B model seems stuck
Issue -
State: open - Opened by guanlinz over 1 year ago
- 9 comments
#103 - Download watchdog kicking in? (M1 mac)
Issue -
State: open - Opened by kryt over 1 year ago
#102 - RuntimeError: Error(s) in loading state_dict for LLaMAForCausalLM: Unexpected key(s) in state_dict:
Issue -
State: open - Opened by ZealHua over 1 year ago
#101 - RecursionError: maximum recursion depth exceeded while calling a Python object
Issue -
State: open - Opened by Vaibhav11002 over 1 year ago
#100 - shape mismatch error
Issue -
State: open - Opened by Celppu over 1 year ago
#99 - an operation was attempted on something that is not a socket
Issue -
State: open - Opened by GameDevKitY over 1 year ago
#98 - parameter inncorrect when I run make command
Issue -
State: open - Opened by GameDevKitY over 1 year ago
#97 - gptq github
Issue -
State: open - Opened by austinmw over 1 year ago
- 4 comments
#96 - Try Modular - Mojo
Issue -
State: open - Opened by eznix86 over 1 year ago
#95 - Randomly get shape mismatch error
Issue -
State: open - Opened by vedantroy over 1 year ago
#94 - Does this include the GPTQ quantization tricks?
Issue -
State: open - Opened by vedantroy over 1 year ago
#93 - Why are params.json empty?
Issue -
State: closed - Opened by ItsCRC over 1 year ago
- 5 comments
#92 - Quantize issue
Issue -
State: open - Opened by ZenekZombie over 1 year ago
#91 - Is that possible to quantize a locally converted model, instead of downloading from hugging face?
Issue -
State: closed - Opened by chigkim over 1 year ago
- 1 comment
#90 - RecursionError running llama.download
Issue -
State: open - Opened by anyangpeng over 1 year ago
- 4 comments
#89 - Adjust watchdog time interval from 30 seconds to 2 minutes.
Pull Request -
State: closed - Opened by Jack-Moo over 1 year ago
- 1 comment
#88 - downloading file to pyllama_data/30B/consolidated.00.pth ...please wait for a few minutes ...
Issue -
State: closed - Opened by Nolyzlel over 1 year ago
- 2 comments
#87 - aria2c 'magnet:?xt=urn:btih:ZXXDAUWYLRUXXBHUYEMS6Q5CE5WA3LVA&dn=LLaMA' not working
Issue -
State: open - Opened by Nolyzlel over 1 year ago
- 2 comments
#86 - Run 'inference.py' and 'model parallel group is not initialized'
Issue -
State: open - Opened by ildartregulov over 1 year ago
- 7 comments
#85 - Apply Delta failed
Issue -
State: open - Opened by majidbhatti over 1 year ago
- 1 comment
#84 - How to run 13B model in a single GPU just by inference.by?
Issue -
State: open - Opened by statyui over 1 year ago
#83 - about rotary embedding in llama
Issue -
State: closed - Opened by irasin over 1 year ago
- 2 comments
#82 - Strange characters
Issue -
State: open - Opened by webpolis over 1 year ago
- 1 comment
#81 - Cannot run on Mac with Python 3.11.3
Issue -
State: open - Opened by kornhill over 1 year ago
- 6 comments
#80 - Inference Error :UnicodeDecodeError: 'utf-8' codec can't decode byte 0xe7 in position 18: invalid continuation byte"
Issue -
State: open - Opened by MaiziXiao over 1 year ago
#79 - docs: reduce bit misspell README
Pull Request -
State: closed - Opened by guspan-tanadi over 1 year ago
#78 - Quantized version link suspect
Issue -
State: open - Opened by thistleknot over 1 year ago
- 1 comment
#77 - Is there a way to skip evaluating after quantizing because it takes forever?
Issue -
State: open - Opened by chigkim over 1 year ago
#76 - Gave written examples to run 7B model on GPUs
Pull Request -
State: closed - Opened by george-adams1 over 1 year ago
#75 - Can't Load Quantized Model with GPTQ-for-LLaMa
Issue -
State: open - Opened by chigkim over 1 year ago
- 2 comments
#74 - a questuon about the single GPU Inference
Issue -
State: open - Opened by zsmmsz99 over 1 year ago
- 1 comment
#73 - quantify llama 7B, the md5 value and the model size does not equals to the value in README
Issue -
State: open - Opened by balcklive over 1 year ago
- 6 comments
#72 - Readme Should Have Inference Command to use for Quantization in Text
Issue -
State: open - Opened by chigkim over 1 year ago
- 1 comment
#71 - rewrite download_community.sh
Pull Request -
State: closed - Opened by llimllib over 1 year ago
- 3 comments
#70 - add a shebang to all shell files
Pull Request -
State: closed - Opened by llimllib over 1 year ago
#69 - Document if it works with CPU / Macos
Issue -
State: open - Opened by ikamensh over 1 year ago
#67 - ModuleNotFoundError: No module named 'transformers'
Issue -
State: open - Opened by tasteitslight over 1 year ago
- 6 comments
#66 - Can't see progress bar
Issue -
State: open - Opened by rahulvigneswaran over 1 year ago
- 1 comment
#65 - Has black formatting been considered?
Issue -
State: open - Opened by tanitna over 1 year ago
#64 - How to run the gradio with 30B model? and what devices are needed? please
Issue -
State: open - Opened by TobiasWYH over 1 year ago
#63 - make download work behind proxy
Pull Request -
State: closed - Opened by wanweilove over 1 year ago
#62 - Killed
Issue -
State: open - Opened by javierp183 over 1 year ago
- 6 comments
#61 - Any way to infer a quantized model on multi GPUs?
Issue -
State: open - Opened by Imagium719 over 1 year ago
- 1 comment
#60 - Quantize Original LLaMA Model Files
Issue -
State: open - Opened by htcml over 1 year ago
- 3 comments
#59 - Let it run under WSL
Pull Request -
State: closed - Opened by daniel-kukiela over 1 year ago
#58 - Quantization with "groupsize" makes the results completely wrong.
Issue -
State: open - Opened by daniel-kukiela over 1 year ago
- 8 comments
#57 - already quantize to 4bit and get the model pyllama-7B4b.pt,but can not run in RTX3080. report torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 86.00 MiB (GPU 0; 10.00 GiB total capacity; 9.24 GiB already allocated;
Issue -
State: open - Opened by elven2016 over 1 year ago
- 2 comments
#56 - Downloading get stuck in infinite loop
Issue -
State: open - Opened by jarimustonen over 1 year ago
- 13 comments
#55 - Error trying Quantize 7B model to 8-bit
Issue -
State: closed - Opened by guoti777 over 1 year ago
- 2 comments
#54 - AttributeError: module 'itree' has no attribute 'Node'
Issue -
State: open - Opened by Tor101 over 1 year ago
- 8 comments
#53 - Docker install
Issue -
State: open - Opened by mgpai22 over 1 year ago
#52 - Meaningless Prediction in 13B 2bit
Issue -
State: open - Opened by axenov over 1 year ago
- 3 comments
#51 - error when installing
Issue -
State: closed - Opened by zzzgit over 1 year ago
- 1 comment
#50 - Error Downloading Models from Community on Winodws
Issue -
State: open - Opened by mmortazavi over 1 year ago
- 5 comments
Labels: bug
#49 - add suggestion for quantization and some bug fixes
Pull Request -
State: closed - Opened by juncongmoo over 1 year ago
#48 - error when running model for inference: ModuleNotFoundError: No module named 'transformers.models.llama'
Issue -
State: closed - Opened by AbFynd over 1 year ago
- 2 comments
#47 - pyllama/downloads returns empty folders
Issue -
State: open - Opened by flyjgh over 1 year ago
- 34 comments
Labels: question
#46 - How can I input prompt when I use multi GPU?
Issue -
State: open - Opened by liydxl over 1 year ago
- 1 comment
#45 - Share your evaluate result
Issue -
State: open - Opened by jeff3071 over 1 year ago
- 3 comments
#44 - fix argument in convert_llama
Pull Request -
State: closed - Opened by a1ex90 over 1 year ago
#43 - AttributeError: module 'numpy' has no attribute 'array'
Issue -
State: open - Opened by jameswan over 1 year ago
#42 - watch downloading speed and restart downloading if it drops to very low
Pull Request -
State: closed - Opened by gmlove over 1 year ago
#41 - Error trying Quantize 7B model to 2-bit
Issue -
State: open - Opened by willintonmb over 1 year ago
- 5 comments
#40 - Quantize 7B model to 8-bit --> "Killed"
Issue -
State: closed - Opened by hex4def6 over 1 year ago
- 1 comment
#39 - "KeyError: 'llama'"
Issue -
State: closed - Opened by DirtyKnightForVi over 1 year ago
#38 - "torch.cuda.OutOfMemoryError: CUDA out of memory" when I'm *not* out of memory
Issue -
State: open - Opened by LoganDark over 1 year ago
- 3 comments
#37 - ModuleNotFoundError: No module named 'quant_cuda'
Issue -
State: open - Opened by AceBeaker2 over 1 year ago
- 15 comments
#36 - Unkown cuda error
Issue -
State: closed - Opened by AceBeaker2 over 1 year ago
- 3 comments
#35 - ValueError: Tokenizer class LLaMATokenizer does not exist or is not currently imported.
Issue -
State: closed - Opened by ilovedbsql over 1 year ago
- 10 comments
#34 - ModuleNotFoundError: No module named 'llama.hf'
Issue -
State: closed - Opened by vetka925 over 1 year ago
- 4 comments
#33 - No module named "transformers" error
Issue -
State: closed - Opened by SimoGiuffrida over 1 year ago
- 1 comment
#32 - example.py FAILED
Issue -
State: closed - Opened by yangzhipeng1108 over 1 year ago
- 1 comment
#31 - Model mismatch for 13B
Issue -
State: open - Opened by BOB603049648 over 1 year ago
- 3 comments
#30 - ModuleNotFoundError: No module named 'quant_cuda'
Issue -
State: closed - Opened by WeissAzura over 1 year ago
- 3 comments
#29 - Download takes forever
Issue -
State: closed - Opened by puyuanliu over 1 year ago
- 2 comments
#28 - Model does not split for 65B
Issue -
State: open - Opened by YixinSong-e over 1 year ago
- 5 comments
#27 - How to run llama_quant without downloading models from huggingface ?
Issue -
State: open - Opened by B2F over 1 year ago
- 1 comment
Labels: enhancement, good first issue
#26 - Error when download models
Issue -
State: open - Opened by paulocoutinhox over 1 year ago
- 5 comments
#25 - world size assertionerror
Issue -
State: closed - Opened by sharlec over 1 year ago
- 6 comments
#24 - M1 inference
Issue -
State: open - Opened by zmactep over 1 year ago
- 1 comment
#23 - multiple GPU support
Pull Request -
State: closed - Opened by mldevorg over 1 year ago
#22 - Execuse me, How to use chat mode?
Issue -
State: closed - Opened by baifachuan over 1 year ago
Labels: invalid
#21 - convert
Pull Request -
State: closed - Opened by mldevorg over 1 year ago
#20 - add simple input loop to inference.py
Pull Request -
State: closed - Opened by lucemia over 1 year ago
#19 - Bug fix3
Pull Request -
State: closed - Opened by juncongmoo over 1 year ago
#18 - fix a bug
Pull Request -
State: closed - Opened by mldevorg over 1 year ago
#17 - fix document
Pull Request -
State: closed - Opened by mldevorg over 1 year ago
#16 - add quant and download info
Pull Request -
State: closed - Opened by juncongmoo over 1 year ago
#15 - Vanilla pytorch LLaMA implementation
Issue -
State: closed - Opened by galatolofederico over 1 year ago
- 3 comments
#14 - Struggle with training LLaMA with a single GPU using both PT v1 and v2
Issue -
State: closed - Opened by linhduongtuan over 1 year ago
- 4 comments
#13 - Docker Playground With LLaMA And PyLLaMA
Issue -
State: closed - Opened by soulteary over 1 year ago
- 1 comment