There is a trend for huge Transformers. We went the other way: decreasing the size! 🤗
— Victor Sanh (@SanhEstPasMoi) August 28, 2019
Introducing DistilBERT: a smaller, faster, cheaper, lighter BERT trained w/ distillation!
95% of BERT's GLUE perf w/ 66M parameters.
📃: https://t.co/4YR4KkSUwQ
💻: https://t.co/vwTPPRsi7r pic.twitter.com/TraVr39J6v
from Twitter https://twitter.com/SanhEstPasMoi
August 28, 2019 at 07:57AM
via IFTTT
No comments:
Post a Comment