SNUH-HARI's picture
Upload merged fine-tuned model 2nd (Trained with Unsloth)
0a551ad verified
|
raw
history blame
3.25 kB
metadata
license: mit
datasets:
  - SNUH-HARI/MedicalLawQA
language:
  - en
  - ko
metrics:
  - accuracy
  - perplexity
base_model:
  - UNIVA-Bllossom/DeepSeek-llama3.1-Bllossom-8B
library_name: transformers
tags:
  - medical
  - unsloth
  - trl
  - sft

SNUH-HARI/DeepSeek-llama3.1-HARI-8B

Model Description

SNUH-HARI/DeepSeek-llama3.1-HARI-8B is a fine-tuned version of DeepSeek-llama3.1-Blossom with 8 billion parameters, optimized for healthcare, legal, and multilingual applications. Developed by Healthcare AI Research Institute (HARI) at Seoul National University Hospital (SNUH), this model integrates medical law and pseudonymized clinical data to enhance patient safety and responsible AI in medicine.

  • Architecture: Transformer-based large language model (LLM)
  • Languages: English, Korean
  • Primary Domains: Healthcare, Legal, and General NLP
  • Use Cases: Medical and legal question answering, clinical decision support, patient safety applications

Training Details

  • Base Model: DeepSeek-llama3.1
  • Fine-Tuned Datasets:
  • Optimization: Mixed precision (FP16) for efficiency
  • Compute Resources: High-performance GPUs (e.g., NVIDIA H100 clusters)

Intended Use

This model is designed for research, healthcare AI, and legal AI applications. It is particularly suitable for:

  • Medical and legal question answering
  • Clinical decision-making support
  • Healthcare policy and compliance

Limitations & Ethical Considerations

  • Not a replacement for medical professionals: Outputs should be validated by experts.
  • Potential biases: Legal and medical knowledge are jurisdiction-specific; users should verify regional applicability.
  • Privacy compliance: No personally identifiable information was used in training.

Evaluation & Benchmarks

  • Perplexity Score: TBD
  • Accuracy on legal-medical QA tasks: TBD
  • Comparison with baseline models (e.g., Med-PaLM, GPT-4-turbo): TBD

How to Use

You can use the model via Hugging Face Transformers:

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "SNUH-HARI/DeepSeek-llama3.1-HARI-8B"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

input_text = "What are the legal requirements for prescribing narcotics in South Korea?"
input_ids = tokenizer(input_text, return_tensors="pt").input_ids

output = model.generate(input_ids, max_length=1024)
print(tokenizer.decode(output[0], skip_special_tokens=True))

License

This model is released under the MIT License.

Citation

If you use this model in your research, please cite:

@misc{SNUH-HARI-DeepSeek-llama3.1-HARI-8B,
  title={SNUH-HARI/DeepSeek-llama3.1-HARI-8B},
  author={Healthcare AI Research Institute (HARI) at Seoul National University Hospital (SNUH)},
  year={2025},
  publisher={Hugging Face},
  url={https://huggingface.co/Seoul National University Hospital (SNUH)-HARI/DeepSeek-llama3.1-HARI-8B}
}