Seattle, WA, USA
I am a research scientist at Apple AIML. My research interests include efficient NLP, mobile computing, and ML systems. I have focused on building efficient and practical NLP systems for both edge devices and the cloud, such as on-device (visual) question answering and faster Transformer models.
Previously, I was a postdoc in the UW NLP group at the University of Washington. I hold a Ph.D. degree in computer science at Stony Brook University. I was a recipient of the Catacosinos Fellowship at Stony Brook University and a Rising Star in Data Science at the University of Chicago.
|Glad to be invited to serve as Action Editor / Area Chair for ACL 2024 !
|BTR was accepted to ICLR as a spotlight paper! 🎊
|Gave a talk at the Efficient ML workshop hosted by Google Research.
- APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and InferenceJan 2024
- BTR: Binary Token Representations for Efficient Retrieval Augmented Language ModelsIn The Twelfth International Conference on Learning Representations, Jan 2024
- PuMer: Pruning and Merging Tokens for Efficient Vision Language ModelsIn Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Jul 2023