Mol-Llama-3.1-8B-Instruct

[Project Page] [Paper] [GitHub]

This repo contains the weights of Mol-LLaMA including the LoRA weights and projectors, based on meta-llama/Llama-3.1-8B-Instruct.

Architecture

image.png

  1. Molecular encoders: Pretrained 2D encoder (MoleculeSTM) and 3D encoder (Uni-Mol)
  2. Blending Module: Combining complementary information from 2D and 3D encoders via cross-attention
  3. Q-Former: Embed molecular representations into query tokens based on SciBERT
  4. LoRA: Adapters for fine-tuning LLMs

Training Dataset

Mol-LLaMA is trained on Mol-LLaMA-Instruct, to learn the fundamental characteristics of molecules with the reasoning ability and explanbility.

How to Use

Please check out the exemplar code for inference in the Github repo.

Citation

If you find our model useful, please consider citing our work.

@misc{kim2025molllama,
    title={Mol-LLaMA: Towards General Understanding of Molecules in Large Molecular Language Model},
    author={Dongki Kim and Wonbin Lee and Sung Ju Hwang},
    year={2025},
    eprint={2502.13449},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}

Acknowledgements

We appreciate LLaMA, 3D-MoLM, MoleculeSTM, Uni-Mol and SciBERT for their open-source contributions.

Downloads last month
98
Safetensors
Model size
172M params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DongkiKim/Mol-Llama-3.1-8B-Instruct

Finetuned
(1274)
this model

Dataset used to train DongkiKim/Mol-Llama-3.1-8B-Instruct