Giannis Daras: Improving sparse transformer models for efficient self-attention (spaCy IRL 2019)

Giannis Daras: Improving sparse transformer models for efficient self-attention (spaCy IRL 2019)

Arxiv 2021: Sparse attention PlanningПодробнее

Arxiv 2021: Sparse attention Planning

Peter Baumgartner: Applied NLP: Lessons from the Field (spaCy IRL 2019)Подробнее

Peter Baumgartner: Applied NLP: Lessons from the Field (spaCy IRL 2019)

Sebastian Ruder: Transfer Learning in Open-Source Natural Language Processing (spaCy IRL 2019)Подробнее

Sebastian Ruder: Transfer Learning in Open-Source Natural Language Processing (spaCy IRL 2019)

Sparse Transformers - Tsvetomila Mihaylova [PyData Sofia April 2020]Подробнее

Sparse Transformers - Tsvetomila Mihaylova [PyData Sofia April 2020]

Giannis Daras - Generative Models and Comp. Imaging: Soft Diffusion and Learning from Corrupted DataПодробнее

Giannis Daras - Generative Models and Comp. Imaging: Soft Diffusion and Learning from Corrupted Data

Yoav Goldberg: The missing elements in NLP (spaCy IRL 2019)Подробнее

Yoav Goldberg: The missing elements in NLP (spaCy IRL 2019)

HPCA' SpAtten: Efficient Sparse Attention Architecture w/ Cascade Token/Head Pruning by Hanrui WangПодробнее

HPCA' SpAtten: Efficient Sparse Attention Architecture w/ Cascade Token/Head Pruning by Hanrui Wang

Adaptive Transformers in NLPПодробнее

Adaptive Transformers in NLP

Efficient TransformersПодробнее

Efficient Transformers

McKenzie Marshall: NLP in Asset Management (spaCy IRL 2019)Подробнее

McKenzie Marshall: NLP in Asset Management (spaCy IRL 2019)

Use Sparse Transfer Learning to Create Sparse Models Fine-Tuned to Your DatasetsПодробнее

Use Sparse Transfer Learning to Create Sparse Models Fine-Tuned to Your Datasets

Big Bird: Transformers for Longer SequencesПодробнее

Big Bird: Transformers for Longer Sequences

Is Sparse Attention more Interpretable?Подробнее

Is Sparse Attention more Interpretable?