Image-Text-to-Text
Transformers
Safetensors
GGUF
English
qwen2_5_vl
remyx
qwen2.5-vl
spatial-reasoning
multimodal
vlm
vqasynth
thinking
reasoning
test-time-compute
robotics
embodied-ai
quantitative-spatial-reasoning
distance-estimation
visual-question-answering
conversational
Eval Results
text-generation-inference
finetuning
#1
by
Abdorifaat
- opened
can you provide the code for finetuning, I want to try it on qwen 2.5 7b
Certainly,
@Abdorifaat
We've added train.py
to the files and you can run:
python train.py \
--model_id YOUR_ORG/QWEN_2.5_MODEL_ID_VARIANT \
--dataset_id remyxai/SpaceThinker \
--lora_r 64 \
--lora_alpha 128 \
--lora_dropout 0.1 \
--target_modules q_proj,v_proj \
--num_train_epochs 5 \
--train_batch_size 2 \
--eval_batch_size 2 \
--gradient_accumulation_steps 4 \
--learning_rate 1e-5 \
--warmup_ratio 0.1 \
--output_dir my-experiment-output \
--wandb_project my-wandb-project \
--wandb_run_name my-wandb-run
Will follow up by adding this to the documentation.
Hope your training run goes well!
salma-remyx
changed discussion status to
closed