A newer version of this model is available: CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct-2412

image/webp

Model Description

Model Details

  • Name: Carrot Llama-3.2 Rabbit Ko
  • Version: 3B Instruct
  • Base Model: CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct
  • Languages: Korean, English
  • Model Type: Large Language Model (Instruction-tuned)

Training Process

본 모델은 다음과 같은 주요 훈련 단계를 거쳤습니다:

  1. SFT (Supervised Fine-Tuning)
    • 고품질 한국어 및 영어 데이터셋을 사용하여 기본 모델을 세부 조정

Limitations

  • 3B 파라미터 규모로 인한 복잡한 작업에서의 제한적 성능
  • 특정 도메인에 대한 깊이 있는 전문성 부족
  • 편향성 및 환각 가능성

Ethics Statement

모델 개발 과정에서 윤리적 고려사항을 최대한 반영하였으나, 사용자는 항상 결과를 비판적으로 검토해야 합니다.

How to Use

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct")
tokenizer = AutoTokenizer.from_pretrained("CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct")

Score

Tasks Version Filter n-shot Metric Value Stderr
gsm8k 3 flexible-extract 5 exact_match 0.6490 ± 0.0131
strict-match 5 exact_match 0.0023 ± 0.0013
gsm8k-ko 3 flexible-extract 5 exact_match 0.3275 ± 0.0134
strict-match 5 exact_match 0.2737 ± 0.0134
ifeval 4 none 5 inst_level_loose_acc 0.8058 ± N/A
none 5 inst_level_strict_acc 0.7686 ± N/A
none 5 prompt_level_loose_acc 0.7320 ± 0.0191
none 5 prompt_level_strict_acc 0.6858 ± 0.0200
Tasks Version Filter n-shot Metric Value Stderr
haerae 1 none acc 0.4180 ± 0.0148
none acc_norm 0.4180 ± 0.0148
- haerae_general_knowledge 1 none 5 acc 0.3125 ± 0.0350
none 5 acc_norm 0.3125 ± 0.0350
- haerae_history 1 none 5 acc 0.3404 ± 0.0347
none 5 acc_norm 0.3404 ± 0.0347
- haerae_loan_word 1 none 5 acc 0.4083 ± 0.0379
none 5 acc_norm 0.4083 ± 0.0379
- haerae_rare_word 1 none 5 acc 0.4815 ± 0.0249
none 5 acc_norm 0.4815 ± 0.0249
- haerae_standard_nomenclature 1 none 5 acc 0.4771 ± 0.0405
none 5 acc_norm 0.4771 ± 0.0405
Tasks Version Filter n-shot Metric Value Stderr
kobest_boolq 1 none 5 acc 0.7664 ± 0.0113
none 5 f1 0.7662 ± N/A
kobest_copa 1 none 5 acc 0.5620 ± 0.0157
none 5 f1 0.5612 ± N/A
kobest_hellaswag 1 none 5 acc 0.3840 ± 0.0218
none 5 acc_norm 0.4900 ± 0.0224
none 5 f1 0.3807 ± N/A
kobest_sentineg 1 none 5 acc 0.5869 ± 0.0247
none 5 f1 0.5545 ± N/A
kobest_wic 1 none 5 acc 0.4952 ± 0.0141
none 5 f1 0.4000 ± N/A
Downloads last month
1,821
Safetensors
Model size
3.21B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct

Finetuned
(406)
this model
Finetunes
1 model
Merges
3 models
Quantizations
4 models

Dataset used to train CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct

Collection including CarrotAI/Llama-3.2-Rabbit-Ko-3B-Instruct