GitHub - soskek/bert-chainer: Chainer implementation of "BERT: Pre-training...

Chainer implementation of "BERT: Pre-training of Deep Bidirectional Transformers for Language Unders...

7 users, 0 mentions 2018/12/02 18:01
[DL Hacks]BERT: Pre-training of Deep Bidirectional Transformers for L…

2018/11/29 Deep Learning JP: http://deeplearning.jp/hacks/

4 users, 5 mentions 2018/12/07 04:31
The Annotated Transformer

The Annotated Transformer

0 users, 0 mentions 2018/08/27 01:24
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Lang…

2018/10/19 Deep Learning JP: http://deeplearning.jp/seminar-2/

0 users, 0 mentions 2018/10/20 12:15
GitHub - huggingface/pytorch-pretrained-BERT: The Big-&-Extending-Repositor...

The Big-&-Extending-Repository-of-Transformers: PyTorch pretrained models for Google's BERT, OpenAI ...

1 users, 7 mentions 2019/03/04 21:47
The Annotated Transformer

The Annotated Transformer

0 users, 0 mentions 2018/06/19 22:30
[1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Lang...

Abstract: We introduce a new language representation model called BERT, which stands for Bidirectio...

0 users, 0 mentions 2018/10/12 06:54
Transformers from scratch | Peter Bloem

I will assume a basic understanding of neural networks and backpropagation. If you'd like to brush u...

5 users, 15 mentions 2019/08/22 05:17
Transformers from scratch | Peter Bloem

I will assume a basic understanding of neural networks and backpropagation. If you'd like to brush u...

0 users, 48 mentions 2019/08/30 14:17
[1906.09777] A Tensorized Transformer for Language Modeling

Latest development of neural models has connected the encoder and decoder through a self-attention m...

0 users, 1 mentions 2019/08/14 05:16