This is a MicroBERT model for Ancient Greek.

  • Its suffix is -mx, which means that it was pretrained using supervision from masked language modeling and XPOS tagging.
  • The unlabeled Ancient Greek data was taken from the Diorisis corpus, totaling 9,058,227 tokens.
  • The UD treebank UD_Ancient_Greek-PROEIL, v2.9, totaling 213,999 tokens, was used for labeled data.

Please see the repository and the paper for more details.

Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support