Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

Collections

Discover the best community collections!

Collections including paper arxiv:2404.05892
LLMs
Collection by Aug 1, 2024
16
  • Attention Is All You Need

    Paper • 1706.03762 • Published Jun 12, 2017 • 104
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

    Paper • 1810.04805 • Published Oct 11, 2018 • 24
  • RoBERTa: A Robustly Optimized BERT Pretraining Approach

    Paper • 1907.11692 • Published Jul 26, 2019 • 9
  • DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter

    Paper • 1910.01108 • Published Oct 2, 2019 • 21
LLMs
Collection by Aug 1, 2024
16
  • Attention Is All You Need

    Paper • 1706.03762 • Published Jun 12, 2017 • 104
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

    Paper • 1810.04805 • Published Oct 11, 2018 • 24
  • RoBERTa: A Robustly Optimized BERT Pretraining Approach

    Paper • 1907.11692 • Published Jul 26, 2019 • 9
  • DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter

    Paper • 1910.01108 • Published Oct 2, 2019 • 21
  • Previous
  • 1
  • 2
  • Next
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs