You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- Low barrier to entry for educators and practitioners
15
+
16
+
State-of-the-art NLP for everyone
17
+
- Deep learning researchers
18
+
- Hands-on practitioners
19
+
- AI/ML/NLP teachers and educators
20
+
21
+
Lower compute costs, smaller carbon footprint
22
+
- Researchers can share trained models instead of always retraining
23
+
- Practitioners can reduce compute time and production costs
24
+
- 8 architectures with over 30 pretrained models, some in more than 100 languages
25
+
26
+
Choose the right framework for every part of a model's lifetime
27
+
- Train state-of-the-art models in 3 lines of code
28
+
- Deep interoperability between TensorFlow 2.0 and PyTorch models
29
+
- Move a single model between TF2.0/PyTorch frameworks at will
30
+
- Seamlessly pick the right framework for training, evaluation, production
31
+
32
+
Contents
33
+
---------------------------------
34
+
35
+
The library currently contains PyTorch and Tensorflow implementations, pre-trained model weights, usage scripts and conversion utilities for the following models:
7
36
8
37
1. `BERT <https://github.com/google-research/bert>`_ (from Google) released with the paper `BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding <https://arxiv.org/abs/1810.04805>`_ by Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova.
9
38
2. `GPT <https://github.com/openai/finetune-transformer-lm>`_ (from OpenAI) released with the paper `Improving Language Understanding by Generative Pre-Training <https://blog.openai.com/language-unsupervised>`_ by Alec Radford, Karthik Narasimhan, Tim Salimans and Ilya Sutskever.
@@ -14,6 +43,7 @@ The library currently contains PyTorch implementations, pre-trained model weight
14
43
7. `RoBERTa <https://github.com/pytorch/fairseq/tree/master/examples/roberta>`_ (from Facebook), released together with the paper a `Robustly Optimized BERT Pretraining Approach <https://arxiv.org/abs/1907.11692>`_ by Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, Veselin Stoyanov.
15
44
8. `DistilBERT <https://huggingface.co/transformers/model_doc/distilbert.html>`_ (from HuggingFace) released together with the blog post `Smaller, faster, cheaper, lighter: Introducing DistilBERT, a distilled version of BERT <https://medium.com/huggingface/distilbert-8cf3380435b5>`_ by Victor Sanh, Lysandre Debut and Thomas Wolf.
16
45
46
+
17
47
.. toctree::
18
48
:maxdepth:2
19
49
:caption:Notes
@@ -37,6 +67,7 @@ The library currently contains PyTorch implementations, pre-trained model weight
0 commit comments