19:52
What Is A Language Model? GPT-3: Language Models Are Few-Shot Learners #GPT3 (part 2)
Deep Learning Explainer
Shared 09/08/2020
55:15
Transformer Architecture Explained | Attention Is All You Need | Foundation of BERT, GPT-3, RoBERTa
Deep Learning Explainer
Shared 07/09/2020
50:21
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (Paper Explained)
Deep Learning Explainer
Shared 19/10/2020
28:22
Sandwich Transformer: Improving Transformer Models by Reordering their Sublayers
Deep Learning Explainer
Shared 27/09/2020
24:03
Can Machines Learn Like Humans - In-context Learning\Meta\Zero-shot Learning | #GPT3 (part 3)
Deep Learning Explainer
Shared 15/08/2020
26:55
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning (Paper Explained)
Deep Learning Explainer
Shared 04/08/2020
48:04
Revealing Dark Secrets of BERT (Analysis of BERT's Attention Heads) - Paper Explained
Deep Learning Explainer
Shared 28/06/2020