Welcome to Yu Yu's homepage!
Google Scholar | CV | Contact
I'm a fourth year PhD student advised by Jia Xu at Stevens Institute of Technology.
My research interest is robustness and generalization of Natural language processing (NLP) systems such as Machine Translation and Dialogue Systems.
I served as the review committee member for EMNLP22, NIPS22, EMNLP 2023, ACL 2023, NIPS 2023, ICML 2023, ICML 2024.
Publications
We propose a neural language modeling system based on low-rank adaptation (LoRA) for speech recognition output rescoring.
Conventional data selection methods select training samples based on the test domain knowledge and not on real life data, thus they frequently fail in unknown domains like patent and Twitter. We propose to select training samples that maximize information uncertainty measured by observation entropy like empirical Shannon entropy, and prediction entropy using mutual information, to cover more possible queries that may appear in unknown worlds.
COLING 2022 (Oral)
This paper introduces our Diversity Advanced Actor-Critic reinforcement learning (A2C) framework (DAAC) to improve the generalization and accuracy of Natural Language Processing (NLP). We quantify diversity on a set of samples using the max dispersion, convex hull volume, and graph entropy based on sentence embeddings in high dimensional metric space.
COLING 2022 (Oral)
Unlike previous work that has defined robustness using Minimax to bound worst cases, we measure robustness based on the consistency of cross-domain accuracy and introduce the coefficient of variation and (ε, γ)-Robustness.
Internship
Google DeepMind
Research Intern, Fall 2023
Topics: Large Language Model, Low-resource code generation, In-context Transfer Learning
Host: Sheena Panthaplackel, Christian Walder
Amazon
Applied Scientist Intern, Summer 2023
Topics: Speech Recognition Rescoring, Parameter-efficient Fine-tuning, Low-rank Adaptation
Host: Huck Yang, Ivan Bulyko