An open API service for providing issue and pull request metadata for open source projects.

GitHub / google/gemma_pytorch issues and pull requests

#80 - Rename unused for loop variable `_` instead of `i`

Pull Request - State: open - Opened by paruby 3 months ago - 3 comments

#78 - No pytorch versions of gemma 3 models on HF

Issue - State: open - Opened by Ocean-Moist 5 months ago - 1 comment

#77 - Failed to cancel access request

Issue - State: closed - Opened by lucassunalt 7 months ago - 2 comments
Labels: stat:awaiting response

#76 - Add required world_size and rank to GemmaDecodeLayer init

Pull Request - State: closed - Opened by DavidRV00 7 months ago - 2 comments

#73 - Non causal sliding window mask ?

Issue - State: closed - Opened by Optimox 9 months ago - 3 comments

#71 - Inconsistent 'query_pre_attn_scalar' Setting Between 9B and 27B Models

Issue - State: open - Opened by kiddj about 1 year ago - 2 comments
Labels: bug, stat:awaiting response

#70 - Hope to See the Source Code of Gemma2 Version

Issue - State: closed - Opened by thefreeman007 about 1 year ago - 1 comment
Labels: type:support

#69 - Remove unused imports

Pull Request - State: closed - Opened by neurosnap about 1 year ago - 1 comment

#68 - Fix downcasting and upcasting similar to https://github.com/google/ge…

Pull Request - State: closed - Opened by michaelmoynihan about 1 year ago - 1 comment

#67 - Fix downcasting and upcasting

Pull Request - State: closed - Opened by danielhanchen about 1 year ago - 1 comment

#66 - Supporting Gemma V2

Pull Request - State: closed - Opened by michaelmoynihan about 1 year ago - 1 comment

#65 - Update run_xla.py

Pull Request - State: closed - Opened by michaelmoynihan about 1 year ago

#64 - gemma-2b-it-pytorch on tpu v5p

Issue - State: closed - Opened by shungcp about 1 year ago - 1 comment

#63 - Modify SentencePiece function calls.

Pull Request - State: closed - Opened by texasmichelle about 1 year ago - 1 comment

#62 - Change return to raise in `get_model_config`.

Pull Request - State: closed - Opened by texasmichelle about 1 year ago - 1 comment

#61 - when to support RecurrentGemma?

Issue - State: closed - Opened by Mddct about 1 year ago - 1 comment
Labels: enhancement

#60 - Gemma finetuning formatting

Issue - State: closed - Opened by mostafamdy over 1 year ago - 3 comments
Labels: type:support

#59 - fix missing torch in requirment

Pull Request - State: closed - Opened by Mddct over 1 year ago - 1 comment

#58 - Add CodeGemma and HF pointers

Pull Request - State: closed - Opened by osanseviero over 1 year ago - 1 comment

#57 - early stop when all sequence reach EOS

Pull Request - State: open - Opened by je1lee over 1 year ago - 3 comments

#56 - Memory saving loading weight for non-quant models

Pull Request - State: closed - Opened by KaneGreen over 1 year ago - 5 comments

#55 - Prepare model for deployment to Private Vertex AI endpoint

Issue - State: closed - Opened by BriianPowell over 1 year ago - 5 comments
Labels: type:support

#54 - Update xla_model_parallel.py

Pull Request - State: closed - Opened by ya0guang over 1 year ago - 2 comments

#53 - Error when run docker/Dockerfile

Issue - State: closed - Opened by Cguanqin over 1 year ago - 3 comments
Labels: type:support, stat:awaiting response

#52 - How to use gemma for multi-round conversations

Issue - State: closed - Opened by ranck626 over 1 year ago - 4 comments
Labels: type:support, stat:awaiting response

#51 - How to save memory when loading weights?

Issue - State: closed - Opened by KaneGreen over 1 year ago - 9 comments
Labels: bug

#50 - Unable to reproduce MATH resulst

Issue - State: open - Opened by wenhuchen over 1 year ago - 3 comments
Labels: type:support, stat:awaiting internal

#49 - fix: raise Exception

Pull Request - State: closed - Opened by leowzz over 1 year ago - 2 comments

#48 - Is it possible to load 7b-it using quantization config

Issue - State: closed - Opened by aliasneo1 over 1 year ago - 1 comment
Labels: enhancement

#47 - Error when running Gemma inference on GPU

Issue - State: closed - Opened by LarryHawkingYoung over 1 year ago - 3 comments
Labels: type:support, stat:awaiting response

#46 - rm fairescale

Pull Request - State: closed - Opened by Mon-ius over 1 year ago - 7 comments

#45 - I got empty result while using 7b-it model

Issue - State: closed - Opened by egbertwong over 1 year ago - 4 comments
Labels: type:support

#44 - Document the existence of 99 unused tokens in the tokenizer

Pull Request - State: closed - Opened by Qubitium over 1 year ago - 1 comment

#43 - fix(temperature): allow passing 0 or None as the temperature parameter

Pull Request - State: closed - Opened by joselpart over 1 year ago - 3 comments

#42 - Can't disable sampling

Issue - State: closed - Opened by joselpart over 1 year ago
Labels: bug

#41 - Is max_position_embeddings=8096 neccessary in 2b model?

Issue - State: closed - Opened by agiwave over 1 year ago - 5 comments
Labels: type:support

#40 - Auto-labels 'Gemma' on 'gemma' issues/PRs.

Pull Request - State: closed - Opened by shmishra99 over 1 year ago - 1 comment

#39 - Objectivity

Issue - State: closed - Opened by o6uoq over 1 year ago
Labels: type:support

#38 - How to fine-tune Gemma with pytorch?

Issue - State: closed - Opened by solitude-alive over 1 year ago - 2 comments
Labels: duplicate

#37 - Gemma fixes - gelu

Pull Request - State: closed - Opened by danielhanchen over 1 year ago - 4 comments

#36 - Torch implementation now same as JAX

Pull Request - State: closed - Opened by thebraingen over 1 year ago - 1 comment

#35 - Implementation now equals JAX

Pull Request - State: closed - Opened by thebraingen over 1 year ago - 1 comment

#34 - Add instructions to download from Hugging Face Hub

Pull Request - State: closed - Opened by osanseviero over 1 year ago - 1 comment

#33 - Inconsistency between PyTorch and JAX implementation

Issue - State: closed - Opened by aboros98 over 1 year ago - 2 comments
Labels: enhancement

#32 - "--output_len" argument ignored

Pull Request - State: closed - Opened by k-nar over 1 year ago - 1 comment

#31 - not found weight file

Issue - State: closed - Opened by Cguanqin over 1 year ago - 5 comments
Labels: type:support, stat:awaiting response

#30 - is it possible to convert gemma_pytorch to onnx to tflite?

Issue - State: closed - Opened by nyadla-sys over 1 year ago - 4 comments
Labels: type:support, stat:awaiting response

#29 - [Question] Embeddings normalization by sqrt(hidden_size)

Issue - State: closed - Opened by Andrei-Aksionov over 1 year ago - 4 comments
Labels: type:support

#26 - After deplyed google/gemma-7b-it, there always is error response.

Issue - State: closed - Opened by ydh10002023 over 1 year ago - 10 comments
Labels: bug

#25 - Cannot run on v4-16 worker 0 TPU VM: "Failed to get global TPU topology"

Issue - State: closed - Opened by markusheimerl over 1 year ago - 6 comments
Labels: type:support

#24 - always loss nan while finetune a few step, wether fp32 or fp16

Issue - State: closed - Opened by yongzhuo over 1 year ago - 1 comment
Labels: type:support

#23 - keras finetuning and inference examples uploaded

Pull Request - State: closed - Opened by r-gheda over 1 year ago - 2 comments

#22 - H

Issue - State: closed - Opened by ZainBinTariq7 over 1 year ago - 1 comment
Labels: type:support

#21 - Changed <2B or 7B> to <2b or 7b> in README

Pull Request - State: closed - Opened by r-gheda over 1 year ago

#20 - Changes <2B or 7B> option to <2b or 7b> in README

Pull Request - State: closed - Opened by r-gheda over 1 year ago - 1 comment

#19 - Output with higher max_length is repetition of base text

Issue - State: closed - Opened by azrael05 over 1 year ago - 9 comments
Labels: type:support, stat:awaiting response

#18 - Update config.py

Pull Request - State: closed - Opened by Khajaamee455 over 1 year ago - 2 comments

#17 - Updated ClassMD

Pull Request - State: closed - Opened by Masomabayat over 1 year ago - 2 comments

#15 - Update xla_model_parallel.py

Pull Request - State: closed - Opened by eltociear over 1 year ago - 3 comments

#13 - RuntimeError: probability tensor contains either `inf`, `nan` or element < 0

Issue - State: closed - Opened by 2579356425 over 1 year ago - 6 comments
Labels: duplicate

#12 - Are there reserved/unused tokens for developers?

Issue - State: closed - Opened by Qubitium over 1 year ago - 3 comments
Labels: type:support

#11 - MPS (Apple Silicon) Support

Issue - State: closed - Opened by dsanmart over 1 year ago - 4 comments
Labels: enhancement

#10 - why some prompt doesn't work, the hidden_states will be nan after GemmaModel.forward

Issue - State: closed - Opened by vupjing over 1 year ago - 11 comments
Labels: bug, stat:awaiting response

#9 - Loading torch checkpoint with weights_only set to True

Pull Request - State: closed - Opened by michaelmoynihan over 1 year ago

#8 - how to finetune with gemma model?

Issue - State: closed - Opened by runningabcd over 1 year ago - 10 comments
Labels: type:support, stat:awaiting response

#7 - Quantised weights are bfloat16 not int8

Issue - State: closed - Opened by dsanmart over 1 year ago - 3 comments
Labels: type:support

#6 - Add utility to convert string to boolean type to fix quant parse arg

Pull Request - State: closed - Opened by nakkapeddi over 1 year ago - 3 comments

#5 - --quant always returns True

Issue - State: closed - Opened by nakkapeddi over 1 year ago - 5 comments
Labels: bug

#4 - A web runtime supported version of gemma is really needed and high value

Issue - State: closed - Opened by Zwe1 over 1 year ago - 4 comments
Labels: enhancement

#3 - RuntimeError: at::cuda::blas::gemm: not implemented for struct c10::BFloat16

Issue - State: closed - Opened by dhchenx over 1 year ago - 9 comments
Labels: bug

#2 - Inconsistencies in Reported Dimensions and Configuration Files

Issue - State: closed - Opened by fvarno over 1 year ago - 2 comments
Labels: type:support

#1 - `torch.load` without `weights_only` parameter is unsafe

Issue - State: closed - Opened by kit1980 over 1 year ago - 2 comments
Labels: bug