Model Card for Sarashina2-13b-finetuned-v1 🧮📘

A Japanese-English bilingual language model fine-tuned on 100 randomly shuffled samples from math datasets. Based on sbintuitions/sarashina2-13b and trained using unsloth, trl, and sft.

Model Details

Description

🧠 Fine-tuned transformer model for solving math problems and answering technical questions in 🇯🇵 Japanese and 🇺🇸 English.

  • Base Model: sbintuitions/sarashina2-13b
  • License: MIT
  • Languages: Japanese, English
  • Model Type: Transformer (Causal LM)

Limitations

⚠️ Known limitations:

  • May generate incorrect or biased answers
  • May hallucinate explanations

🔍 Always verify important outputs.

Get Started

from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Mori-kamiyama/sarashina-13B-finetuned-v1")
tokenizer = AutoTokenizer.from_pretrained("Mori-kamiyama/sarashina-13B-finetuned-v1")

Training Data

🧮 Fine-tuned using 100 randomly shuffled samples from:

  • elyza/JaMARD (Japanese instruction tuning)
  • EleutherAI/hendrycks_math (math reasoning)

Architecture

  • Transformer (Causal LM)
  • Fine-tuned with TRL + Unsloth
Downloads last month
3
Safetensors
Model size
13.7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Mori-kamiyama/sarashina-13B-finetuned-v1

Finetuned
(7)
this model
Quantizations
2 models

Datasets used to train Mori-kamiyama/sarashina-13B-finetuned-v1