User profiles for Myle Ott
Myle OttCharacter AI Verified email at character.ai Cited by 46149 |
Roberta: A robustly optimized bert pretraining approach
Language model pretraining has led to significant performance gains but careful comparison
between different approaches is challenging. Training is computationally expensive, often …
between different approaches is challenging. Training is computationally expensive, often …
fairseq: A fast, extensible toolkit for sequence modeling
fairseq is an open-source sequence modeling toolkit that allows researchers and developers
to train custom models for translation, summarization, language modeling, and other text …
to train custom models for translation, summarization, language modeling, and other text …
Finding deceptive opinion spam by any stretch of the imagination
Consumers increasingly rate, review and research products online. Consequently, websites
containing consumer reviews are becoming targets of opinion spam. While recent work has …
containing consumer reviews are becoming targets of opinion spam. While recent work has …
Recipes for building an open-domain chatbot
Building open-domain chatbots is a challenging area for machine learning research. While
prior work has shown that scaling neural models in the number of parameters and the size of …
prior work has shown that scaling neural models in the number of parameters and the size of …
Unsupervised cross-lingual representation learning at scale
This paper shows that pretraining multilingual language models at scale leads to significant
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer-…
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer-…
Opt: Open pre-trained transformer language models
Large language models, which are often trained for hundreds of thousands of compute days,
have shown remarkable capabilities for zero- and few-shot learning. Given their …
have shown remarkable capabilities for zero- and few-shot learning. Given their …
Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences
In the field of artificial intelligence, a combination of scale in data and model capacity
enabled by unsupervised learning has led to major advances in representation learning and …
enabled by unsupervised learning has led to major advances in representation learning and …
Understanding back-translation at scale
An effective method to improve neural machine translation with monolingual data is to
augment the parallel training corpus with back-translations of target language sentences. This …
augment the parallel training corpus with back-translations of target language sentences. This …
Sustainable ai: Environmental implications, challenges and opportunities
This paper explores the environmental impact of the super-linear growth trends for AI from a
holistic perspective, spanning Data, Algorithms, and System Hardware. We characterize the …
holistic perspective, spanning Data, Algorithms, and System Hardware. We characterize the …
Phrase-based & neural unsupervised machine translation
Machine translation systems achieve near human-level performance on some languages, yet
their effectiveness strongly relies on the availability of large amounts of parallel sentences, …
their effectiveness strongly relies on the availability of large amounts of parallel sentences, …