John Nguyen is a Member of Technical Staff at AMI Labs, where he focuses on building visually intelligent systems and advancing research toward unified world models. His work sits at the intersection of multimodal generation, perception, and understanding, with a particular emphasis on open-world intelligence and large-scale system design.
John has contributed to a range of influential research projects exploring the future of AI architectures, including concurrent mixed-modal generation (OneFlow), temporally expansive video generation (Flowception), scaling laws for multimodal pretraining (Beyond Language Modeling), and tokenization-free language modeling (Byte Latent Transformer), which received an ACL Outstanding Paper award.
Alongside his research, he has built widely adopted systems infrastructure. He is the creator of Opacus, a library for differentially private training with over four million downloads, and Papaya, a large-scale asynchronous federated learning platform presented at MLSys 2022 and deployed to millions of users.
Previously a Research Engineer at Meta’s FAIR lab, John worked on large multimodal models and distributed learning systems. He graduated cum laude from the University of California, Davis, earning degrees in Computer Science and Statistics, as well as a Master’s in Computer Science.