Does not work at all
Output is always empty even when using the official example from here (https://huggingface.co/google/gemma-3-4b-it) when replaced model name with this one.
Yes
Output is always empty even when using the official example from here (https://huggingface.co/google/gemma-3-4b-it) when replaced model name with this one.
Yes
Are you guys still having the issue?
is it good to start finetuning with ? or should i use the main model ?
and also update your colab notebook, code is breaking for dinetuinng
Output is always empty even when using the official example from here (https://huggingface.co/google/gemma-3-4b-it) when replaced model name with this one.
Yes
Are you guys still having the issue?
i'm having the same issue as well.
Dont know why so many people downloaded when it does not work.
Actually seem to work when I use bfloat instead of float:
model = Gemma3ForConditionalGeneration.from_pretrained(
model_id,
cache_dir=cache_dir,
torch_dtype=torch.bfloat16,
).to("cuda").eval()
is it good to start finetuning with ? or should i use the main model ?
and also update your colab notebook, code is breaking for dinetuinng
Output is always empty even when using the official example from here (https://huggingface.co/google/gemma-3-4b-it) when replaced model name with this one.
Yes
Are you guys still having the issue?
i'm having the same issue as well.
Dont know why so many people downloaded when it does not work.
We updated the models btw. If you're still having issues let me know exactly what the error is.
Currently Gemma 3 does not work on GPUs that do not support bf16 and unfortunately this is out of our control. so if your GPU only supports f16, it wont work unless you use Unsloth for finetuning/inference
Sorry I meant Gemma 3 only works on bf16. Not f16
ok, and can i use it on kaggle both the free gpus ? or it will always use 1 gpu ?
I tried with accelerator but it failed