RESEARCH AREA
Self-attention based sequence models that revolutionized NLP and beyond.
Attention Is All You Need
BERT: Pre-training of Deep Bidirectional Transformers
An Image is Worth 16x16 Words: Transformers for Vision
Ilya Sutskever
Safe Superintelligence Inc.