The “tl;dr” on a few notable transformer papers
# tldr-transformers The tl;dr on a few notable transformer/language model papers + other papers (alignment, memorization, etc). Models: GPT- *, * BERT *, Adapter- *, * T5, etc. Each set of notes includes links to the paper, the original code implementation (if available) and the Huggingface :hugs: implementation. Here is an example: t5. The transformers papers are presented somewhat chronologically below. Go to the “:point_right: Notes :point_left:” column below to find the notes for each paper. This repo also includes […]
Read more