Mori-kamiyama's picture
Update README.md
ffc18f7 verified
metadata
library_name: transformers
tags:
  - unsloth
  - trl
  - sft
license: mit
datasets:
  - elyza/JaMARD
  - EleutherAI/hendrycks_math
language:
  - ja
  - en
base_model:
  - sbintuitions/sarashina2-13b
pipeline_tag: text-generation

Model Card for Sarashina2-13b-finetuned-v1 🧮📘

A Japanese-English bilingual language model fine-tuned on 100 randomly shuffled samples from math datasets. Based on sbintuitions/sarashina2-13b and trained using unsloth, trl, and sft.

Model Details

Description

🧠 Fine-tuned transformer model for solving math problems and answering technical questions in 🇯🇵 Japanese and 🇺🇸 English.

  • Base Model: sbintuitions/sarashina2-13b
  • License: MIT
  • Languages: Japanese, English
  • Model Type: Transformer (Causal LM)

Limitations

⚠️ Known limitations:

  • May generate incorrect or biased answers
  • May hallucinate explanations

🔍 Always verify important outputs.

Get Started

from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Mori-kamiyama/sarashina-13B-finetuned-v1")
tokenizer = AutoTokenizer.from_pretrained("Mori-kamiyama/sarashina-13B-finetuned-v1")

Training Data

🧮 Fine-tuned using 100 randomly shuffled samples from:

  • elyza/JaMARD (Japanese instruction tuning)
  • EleutherAI/hendrycks_math (math reasoning)

Architecture

  • Transformer (Causal LM)
  • Fine-tuned with TRL + Unsloth