metadata
license: apache-2.0
library_name: transformers
pipeline_tag: text-generation
This repository contains the RADLADS models as presented in the paper RADLADS: Rapid Attention Distillation to Linear Attention Decoders at Scale.
More information can be found at the Github repository: https://github.com/recursal/RADLADS-paper