Learn how to make BERT smaller and faster

Let's look at compression methods for neural networks, such as quantization and pruning. Then, we apply one to BERT using TensorFlow Lite.

5 mentions: @alanmnichol@Rasa_HQ@rlebron_bioinfo@data4gud@chetanhere
Keywords: bert
Date: 2019/08/08 14:56

Referring Tweets

@alanmnichol Great work by our (undergraduate!) ML intern Sam on compressing big language models like BERT https://t.co/uIKhkSx4y1 If you'd like to intern with us too, just ping me (DMs open) or apply for a research role and mention internships.
@Rasa_HQ Check out our recent blogpost about how to apply compression methods to #BERT using #TensorFlow Lite. #conversationalAI #NLP https://t.co/f1EbKQmvMV
@rlebron_bioinfo Compressing BERT for faster prediction https://t.co/KVR4w6TBqo

Related Entries

Read more GitHub - soskek/bert-chainer: Chainer implementation of "BERT: Pre-training of Deep Bidirectional Tr...
Read more [DL Hacks]BERT: Pre-training of Deep Bidirectional Transformers for L…
Read more [DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Lang…
Read more GitHub - huggingface/pytorch-pretrained-BERT: The Big-&-Extending-Repository-of-Transformers: PyTorc...
Read more The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning) – Jay Alammar – Visualizing ...