User profiles for Myle Ott

Myle Ott

Character AI
Verified email at character.ai
Cited by 46149

Roberta: A robustly optimized bert pretraining approach

Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen… - arXiv preprint arXiv …, 2019 - arxiv.org
Language model pretraining has led to significant performance gains but careful comparison
between different approaches is challenging. Training is computationally expensive, often …

fairseq: A fast, extensible toolkit for sequence modeling

M Ott, S Edunov, A Baevski, A Fan, S Gross… - arXiv preprint arXiv …, 2019 - arxiv.org
fairseq is an open-source sequence modeling toolkit that allows researchers and developers
to train custom models for translation, summarization, language modeling, and other text …

Finding deceptive opinion spam by any stretch of the imagination

M Ott, Y Choi, C Cardie, JT Hancock - arXiv preprint arXiv:1107.4557, 2011 - arxiv.org
Consumers increasingly rate, review and research products online. Consequently, websites
containing consumer reviews are becoming targets of opinion spam. While recent work has …

Recipes for building an open-domain chatbot

…, N Goyal, D Ju, M Williamson, Y Liu, J Xu, M Ott… - arXiv preprint arXiv …, 2020 - arxiv.org
Building open-domain chatbots is a challenging area for machine learning research. While
prior work has shown that scaling neural models in the number of parameters and the size of …

Unsupervised cross-lingual representation learning at scale

…, G Wenzek, F Guzmán, E Grave, M Ott… - arXiv preprint arXiv …, 2019 - arxiv.org
This paper shows that pretraining multilingual language models at scale leads to significant
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer-…

Opt: Open pre-trained transformer language models

…, M Diab, X Li, XV Lin, T Mihaylov, M Ott… - arXiv preprint arXiv …, 2022 - arxiv.org
Large language models, which are often trained for hundreds of thousands of compute days,
have shown remarkable capabilities for zero- and few-shot learning. Given their …

Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences

…, S Goyal, Z Lin, J Liu, D Guo, M Ott… - Proceedings of the …, 2021 - National Acad Sciences
In the field of artificial intelligence, a combination of scale in data and model capacity
enabled by unsupervised learning has led to major advances in representation learning and …

Understanding back-translation at scale

S Edunov, M Ott, M Auli, D Grangier - arXiv preprint arXiv:1808.09381, 2018 - arxiv.org
An effective method to improve neural machine translation with monolingual data is to
augment the parallel training corpus with back-translations of target language sentences. This …

Sustainable ai: Environmental implications, challenges and opportunities

…, C Bai, M Gschwind, A Gupta, M Ott… - Proceedings of …, 2022 - proceedings.mlsys.org
This paper explores the environmental impact of the super-linear growth trends for AI from a
holistic perspective, spanning Data, Algorithms, and System Hardware. We characterize the …

Phrase-based & neural unsupervised machine translation

G Lample, M Ott, A Conneau, L Denoyer… - arXiv preprint arXiv …, 2018 - arxiv.org
Machine translation systems achieve near human-level performance on some languages, yet
their effectiveness strongly relies on the availability of large amounts of parallel sentences, …