Rui Ye / 叶锐

I am a third-year PhD candidate at Shanghai Jiao Tong University (SJTU) in Shanghai, China. Before that, I received my Bachelor degree from SJTU, ranked the first out of 150.

I am currently advised by Prof. Siheng Chen, in the MediaBrain Lab. My research interests are in Trustworthy AI and Collaborative AI. Specifically, I am interested in trustworthy large language models (LLMs), multi-agent system, and federated learning. I have interned at Microsoft Research Asia (MSRA) and Shanghai AI Laboratory.

I am actively seeking collaborations and opportunities as a research intern or visiting student (could be one-year long), please feel free to contact me!!!

Email  /  Google Scholar  /  Github  /  LinkedIn  /  Twitter

profile photo
🔥 News
  • [2024.09] I was awarded the National Scholarship 2024, thanks to the guidance of Prof. Chen.
  • [2024.09] One paper (FedLLM-Bench) is accepted by NeurIPS 2024!
  • [2024.08] One co-first authored paper (FedRSU) is accepted by T-ITS!
  • [2024.07] I will attend ICML 2024 at Vienna, Austria. See you there!
  • [2024.06] We release the first realistic benchmark for FedLLM: FedLLM-Bench!
  • [2024.05] One paper (OpenFedLLM) is accepted by KDD 2024!
  • [2024.05] One co-first authored paper (Reverse Alignment) is accepted by Findings of ACL 2024!
  • [2024.05] One co-first authored paper (MATRIX) is accepted by ICML 2024 (Spotlight)!
  • [2024.02] We release a comprehensive [FL x LLMs] framework OpenFedLLM!!!
  • [2024.01] One paper (FedCOG) is accepted by ICLR 2024! See you at Vienna!
  • [2023.11] Checking out from MSRA in November and actively seeking new collaboration opportunities.
  • [2023.08] One paper (FedFM) is accepted by IEEE Transactions on Signal Processing (T-SP)!
  • [2023.07] Start second internship at Microsoft Research Asia (MSRA), Beijing (on-site).
  • [2023.04] Two papers (FedDisco & pFedGraph) are accepted by ICML 2023!
  • [2022.11] Start internship at Microsoft Research Asia (MSRA), Beijing (remote).
📑 Publications

* denotes equal contribution, denotes corresponding author, see full list in Google Scholar, some are highlighted.

2024
fedllmbench

fedllmbench

fedllmbench

FedLLM-Bench: Realistic Benchmarks for Federated Learning of Large Language Models
Rui Ye*, Rui Ge*, Xinyu Zhu, Jingyi Chai, Yaxin Du, Yang Liu, Yanfeng Wang, Siheng Chen
Conference on Neural Information Processing Systems (NeurIPS), 2024
arXiv / BibTeX / Code

This paper proposes the first realistic benchmark for federated learning of large language models, termed FedLLM-Bench. It encompasses 3 datasets for instruction tuning task and 1 dataset for preference alignment task, which exhibit diversities in language, quality, quantity, instruction, length, embedding, and preference.

fedllmattack

fedllmattack

Emerging Safety Attack and Defense in Federated Instruction Tuning of Large Language Models
Rui Ye*, Jingyi Chai*, Xiangrui Liu, Yaodong Yang, Yanfeng Wang, Siheng Chen
NeurIPS FL@FM Workshop, 2024
arXiv / BibTeX

This paper for the first time reveals the vulnerability of safety alignment during federated instruction tuning by proposing a simple safety attack method. While many existing FL defense methods fail to defend against such attack, we propose a post-hoc defense method that automatically and effectively enhances the safety alignment of LLMs.

openfedllm

openfedllm

OpenFedLLM: Training Large Language Models on Decentralized Private Data via Federated Learning
Rui Ye, Wenhao Wang, Jingyi Chai, Dihan Li, Zexi Li, Yinda Xu, Yaxin Du, Yanfeng Wang, Siheng Chen
Conference on Knowledge Discovery and Data Mining (KDD), 2024
ICLR AGI Workshop and DPFM Workshop, 2024
arXiv / ACM / BibTeX / Code

This paper proposes OpenFedLLM for training large language models on decentralized private data via federated learning, which covers instruction tuning, value alignment, 7 FL algorithms, 8 training datasets, and 30+ evaluation metrics. Based on OpenFedLLM, we conduct a comprehensive empirical study, provide insights, and point out future directions.

matrix

matrix

Self-Alignment of Large Language Models via Monopolylogue-based Social Scene Simulation
Xianghe Pang*, Shuo Tang*, Rui Ye*, Yuxin Xiong, Bolun Zhang, Yanfeng Wang, Siheng Chen
International Conference on Machine Learning (ICML), Spotlight, 2024
ICLR AGI Workshop, Oral, 2024
arXiv / OpenReview / BibTeX / Project / Code

This paper proposes to self-align large language models via social scene simulation, which is powered by our proposed simulator called MATRIX. Human evaluations show that our aligned 13/30B LLMs can outperform GPT-4 on value alignment.

ipfl
Incentivizing Inclusive Data Contributions in Personalized Federated Learning
Enpei Zhang*, Jingyi Chai*, Rui Ye*, Yanfeng Wang, Siheng Chen
ICLR AGI Workshop and DPFM Workshop, 2024
OpenReview / BibTeX

This paper proposes inclusive and incentivized personalized federated learning (iPFL), which incentivizes data holders with diverse purposes to collaboratively train personalized models without revealing raw data.

reverse_alignment
On the Vulnerability of Safety Alignment in Open-Access LLMs
Jingwei Yi*, Rui Ye*, Qisi Chen, Bin Zhu, Siheng Chen, Defu Lian, Guangzhong Sun, Xing Xie, Fangzhao Wu
Findings of the Association for Computational Linguistics (ACL), 2024
Paper / BibTeX

This paper unreveals the vulnerability of value alignment in aligned open-source LLMs by proposing a series of efficient attack methods (i.e., reverse alignment). Experiments show that simple fine-tuning can significantly compromise the alignment of the LLMs.

2023
fedgc
Federated Learning Empowered by Generative Content
Rui Ye, Xinyu Zhu, Jingyi Chai, Siheng Chen, Yanfeng Wang
NeurIPS FL@FM Workshop, 2024
arXiv / BibTeX

This paper for the first time explores how advanced generative models can benefit FL on heterogeneous private data. We show that generative content can not only mitigate data heterogeneity, but also enhance privacy preservation for FL.

fedcog
Fake It Till Make It: Federated Learning with Consensus-Oriented Generation
Rui Ye, Yaxin Du, Zhenyang Ni, Siheng Chen, Yanfeng Wang
International Conference on Learning Representations (ICLR), 2024
Paper / BibTeX / Code

This paper proposes to more fundamentally handle data heterogeneity from the perspective of data, which is achieved by extracting consensus data from the global model to complement clients' heterogeneous data.

feddisco
FedDisco: Federated Learning with Discrepancy-aware Collaboration
Rui Ye, Mingkai Xu, Jianyu Wang, Chenxin Xu, Siheng Chen, Yanfeng Wang
International Conference on Machine Learning (ICML), 2023
arXiv / BibTeX / PMLR / Code

Based on our empirical and theoretical observations, we propose to aggregate models based on both dataset size and a defined discrepancy value.

pfedgraph
Personalized Federated Learning with Inferred Collaboration Graphs
Rui Ye*, Zhenyang Ni*, Fangzhao Wu, Siheng Chen, Yanfeng Wang
International Conference on Machine Learning (ICML), 2023
PMLR / BibTeX / Code

We propose a pFedGraph algorithm to promote more collaboration between clients with more similar data distributions.

fedfm
FedFM: Anchor-based Feature Matching for Data Heterogeneity in Federated Learning
Rui Ye, Zhenyang Ni, Chenxin Xu, Jianyu Wang, Siheng Chen, Yanfeng Wang
IEEE Transactions on Signal Processing (TSP), 2023
Paper / IEEE / BibTeX / Code (PyTorch, PaddlePaddle, MindSpore)

We propose to align category-wise feature spaces of clients in FL, which achieves pleasant performance with theoretical convergence guarantee.

🎓 Educations
sjtu Degree: Bachelor
Period: 2018.09 - 2022.06
Major: Information Engineering (AI Class)
GPA: 3.94/4.3 (ranked 1st out of 150)
🥇 Honors & Awards
  • National Scholarship for PhD Students, 2024
  • National Scholarship for Undergraduates, 2020 (2 out of 150)
  • Shanghai Outstanding Graduates, 2022
  • Samsung Scholarship, 2023 (only one awardee)
  • Mathematical Contest in Modeling, Finalist, 2021 (<1%)
  • Shanghai Jiao Tong University Wenjun Wu AI Class, 2022 (16 are selected)
  • Shanghai Jiao Tong University Xu Zhang Academician Scholarship, 2022 (3 out of 150)
  • Shanghai Jiao Tong University Ceyear Scholarship, 2021
  • Shanghai Jiao Tong University Fujian Alumni Association Scholarship, 2019 (the youngest awardee)
  • Shanghai Jiao Tong University Class B Scholarship, 2019&2020&2021
👀 Misc
Review: NeurIPS (main) 2023, 2024 / NeurIPS (dataset) 2024 / ICLR 2024, 2025 / ICML 2024 / AAAI 2025 / ICASSP 2024.
Life:I love playing basketball / listening rap music / travelling.

Derived from Jon Barron's website.