About me

Hi! 👋🏻 I’m Yuxuan (Leo) Lu, a Ph.D. student at Northeastern University. Before that, I got my B.E. in Computer Science and Technology and Graduated with honor at Beijing University of Technology. I’m advised by Prof. Dakuo Wang. My research interest includes Human Computer Interaction and Natural Language Processing , especially in training, running and utilizing Large Language Models (LLMs) effiently and effectively. In the past, I’ve worked as Machine Learning Researcher at a joint program between LinkedIn and Microsoft Research Asia. I’ve also worked as an intern research assistant at THUNLP lab, supervised by Prof. Zhiyuan Liu(刘知远).

Picture of me, taken in The Sayram Lake (赛里木湖)

Education

I’m currently persuing my Ph.D. in Computer Science at Khoury College of Computer Sciences, Northeastern University, advised by Prof. Dakuo Wang.

I got my B.E. in Computer Science and Technology and Graduated with honor at Beijing University of Technology. Before that, I’ve finished my junior and senior high at Beijing National Day School (北京市十一学校).

Publications

2024

  1. Professional Network Matters: Connections Empower Person-Job Fit
    Hao ChenLun DuYuxuan Lu, Qiang Fu, Xu Chen, Shi Han, Yanbin Kang, Guangming Lu, and Zi Li
    In Proceedings of the 17th ACM International Conference on Web Search and Data Mining, 2024
  2. In Submission
    Exploring Parent’s Needs for Children-Centered AI to Support Preschoolers’ Storytelling and Reading Activities
    Yuling Sun, Jiali Liu, Bingsheng Yao, Jiaju Chen, Dakuo Wang, Xiaojuan Ma, Yuxuan Lu, Ying Xu, and Liang He
    In Submission, 2024
  3. Rethinking human-ai collaboration in complex medical decision making: A case study in sepsis diagnosis
    Shao Zhang, Jianing Yu, Xuhai Xu, Changchang Yin, Yuxuan LuBingsheng Yao, Melanie Tory, Lace M Padilla, Jeffrey Caterino, Ping Zhang, and  others
    2024

2023

  1. More Samples or More Prompt Inputs? Exploring Effective In-Context Sampling for LLM Few-Shot Prompt Engineering
    arXiv preprint arXiv:2311.09782, 2023
  2. Human Still Wins over LLM: An Empirical Study of Active Learning on Domain-Specific Annotation Tasks
    Yuxuan LuBingsheng YaoShao Zhang, Yun Wang, Peng Zhang, Tun Lu, Toby Jia-Jun Li, and Dakuo Wang
    arXiv preprint arXiv:2311.09825, 2023
  3. FairytaleCQA: Integrating a Commonsense Knowledge Graph into Children’s Storybook Narratives
    Jiaju Chen, Yuxuan LuShao ZhangBingsheng Yao, Yuanzhe Dong, Ying Xu, Yunyao Li, Qianwen Wang, Dakuo Wang, and Yuling Sun
    arXiv preprint arXiv:2311.09756, 2023
  4. Beyond Labels: Empowering Human Annotators with Natural Language Explanations through a Novel Active-Learning Architecture
    Bingsheng YaoIshan JindalLucian PopaYannis Katsis, Sayan Ghosh, Lihong He, Yuxuan Lu, Shashank Srivastava, Yunyao Li, James Hendler, and Dakuo Wang
    In Findings of the Association for Computational Linguistics: EMNLP 2023, Dec 2023
  5. TCBB 2023
    Improving Biomedical Question Answering by Data Augmentation and Model Weighting
    Yongping Du, Jingya Yan, Yuxuan Lu, Yiliang Zhao, and Xingnan Jin
    IEEE/ACM Transactions on Computational Biology and Bioinformatics, Dec 2023

2022

  1. Contextual Embedding and Model Weighting by Fusing Domain Knowledge on Biomedical Question Answering
    Yuxuan Lu, Jingya Yan, Zhixuan Qi, Zhongzheng Ge, and Yongping Du
    In Proceedings of the 13th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics, Dec 2022

2021

  1. BIBM 2021
    Dual Model Weighting Strategy and Data Augmentation in Biomedical Question Answering
    Yongping Du, Jingya Yan, Yiliang Zhao, Yuxuan Lu, and Xingnan Jin
    In 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Dec 2021

Research Experience

My current research fields includes data annotation and optimizing LLMs.

Before that, I’ve worked as Machine Learning Researcher at a joint program between LinkedIn and Microsoft Research Asia where I do study about LinkedIn’s social network data. I’ve also worked as an intern research assistant at THUNLP lab, supervised by Prof. Zhiyuan Liu(刘知远). My research area there includes Knowledge Embedding.

Open source communities

I’ve participated in many open-source communities. I’m the maintainer of the VSCode extension LaTeX-Utilities, and I’m the founder and maintainer of the EduOJ project. Furthermore, I’ve contributed to many open-source projects, like GitLab, UniversalOJ, OI-Wiki, nix and others.

I’ve participated as mentor and community leader in the Open Source Promotion Plan 2021. All my 3 students successfully finished their projects. I’ve participated as a student in the OSPP 2020 in the UniversalOJ community, and successfully finished my project.

Learn more about my open-source experience at here.