Skip to main content

Transformers: Bad Band-Aids and Missing Memory

·38 words·1 min · Download pdf

Monthly paper reminder that Transformer architecture is still a stop-gap solution. Here authors create tasks to test generalization of formal language and find that positional encoding is a bad band-aid and augmented memory is dearly missing.

https://arxiv.org/abs/2207.02098

Discussion