Text Generation
Transformers

Add model card

#1
by nielsr HF Staff - opened

This PR adds a model card for the paper RADLADS: Rapid Attention Distillation to Linear Attention Decoders at Scale.

It adds the Apache 2.0 license, the Transformers library, the text-generation pipeline tag, a link to the paper, and a link to the code repository.

Please review and merge this PR if everything looks good.

KaraKaraWitch changed pull request status to merged
recursal org

Looks good to me. Merged it in! :)

Sign up or log in to comment