User profiles for Alexander Ku

Alexander Ku

Google DeepMind, Princeton University
Verified email at princeton.edu
Cited by 4350

A universal SNP and small-indel variant caller using deep neural networks

…, D Alexander, S Schwartz, T Colthurst, A Ku… - Nature …, 2018 - nature.com
Despite rapid advances in sequencing technologies, accurately calling genetic variants
present in an individual genome from billions of short, errorful sequence reads remains …

[PDF][PDF] Scaling autoregressive models for content-rich text-to-image generation

…, T Luong, G Baid, Z Wang, V Vasudevan, A Ku… - arXiv preprint arXiv …, 2022 - 3dvar.com
We present the Pathways [1] Autoregressive Text-to-Image (Parti) model, which generates
high-fidelity photorealistic images and supports content-rich synthesis involving complex …

Image transformer

…, L Kaiser, N Shazeer, A Ku… - International …, 2018 - proceedings.mlr.press
Image generation has been successfully cast as an autoregressive sequence generation or
transformation problem. Recent work has shown that self-attention is an effective way of …

Vector-quantized image modeling with improved vqgan

…, X Li, JY Koh, H Zhang, R Pang, J Qin, A Ku… - arXiv preprint arXiv …, 2021 - arxiv.org
Pretraining language models with next-token prediction on massive text corpora has delivered
phenomenal zero-shot, few-shot, transfer learning and multi-tasking capabilities on both …

Room-across-room: Multilingual vision-and-language navigation with dense spatiotemporal grounding

A Ku, P Anderson, R Patel, E Ie, J Baldridge - arXiv preprint arXiv …, 2020 - arxiv.org
We introduce Room-Across-Room (RxR), a new Vision-and-Language Navigation (VLN)
dataset. RxR is multilingual (English, Hindi, and Telugu) and larger (more paths and …

Stay on the path: Instruction fidelity in vision-and-language navigation

V Jain, G Magalhaes, A Ku, A Vaswani, E Ie… - arXiv preprint arXiv …, 2019 - arxiv.org
Advances in learning and representations have reinvigorated work that connects language
to other modalities. A particularly exciting direction is Vision-and-Language Navigation(VLN), …

Transferable representation learning in vision-and-language navigation

H Huang, V Jain, H Mehta, A Ku… - Proceedings of the …, 2019 - openaccess.thecvf.com
Vision-and-Language Navigation (VLN) tasks such as Room-to-Room (R2R) require
machine agents to interpret natural language instructions and learn to act in visually realistic …

General evaluation for instruction conditioned navigation using dynamic time warping

G Ilharco, V Jain, A Ku, E Ie, J Baldridge - arXiv preprint arXiv:1907.05446, 2019 - arxiv.org
In instruction conditioned navigation, agents interpret natural language and their surroundings
to navigate through an environment. Datasets for studying this task typically contain pairs …

A new path: Scaling vision-and-language navigation with synthetic instructions and imitation learning

…, P Anderson, S Wang, JY Koh, A Ku… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recent studies in Vision-and-Language Navigation (VLN) train RL agents to execute natural-language
navigation instructions in photorealistic environments, as a step towards robots …

Gaussian process probes (gpp) for uncertainty-aware probing

Z Wang, A Ku, J Baldridge… - Advances in Neural …, 2024 - proceedings.neurips.cc
Understanding which concepts models can and cannot represent has been fundamental to
many tasks: from effective and responsible use of models to detecting out of distribution data. …