Zhijun Chen

Ph.D. graduate from Beihang University (Beijing, China)

figure_homepage.jpeg

Email: zhijunchen[at]buaa.edu.cn, zhijunchen.cs[at]gmail.com

I am a PhD graduate from the School of Computer Science at Beihang University, under the supervision of Professor Hailong Sun. I have also had the privilege of working closely with Assistant Professor Jie Yang from Delft University of Technology. Since my graduation in 2024, I have continued my research work (e.g., arXiv 2025) and am currently actively seeking a suitable postdoc position.

Research. My primary research interests lie in the general areas of Machine Learning and Natural Language Processing, with particular focus on Ensemble Learning for LLMs, including: LLM Ensemble (for Inference) (arXiv 2025), Ensemble Fine-Tuning for LLMs, Best-of-N Test-Time Scaling, Multi-Prompt Learning, etc. Looking ahead, I aim to embark on research that is more fundamental, generalizable, and of broader significance.

Ongoing. 1) An extended Journal version of our survey (arXiv 2025) on LLM Ensemble; 2) A new LLM Ensemble method focused on solving inference-intensive tasks; 3) Ensemble fine-tuning for LLMs.

Collaboration. 1) For researchers with relevant experience: If you are interested in discussing ideas, collaborating on papers, or other related opportunities, please feel free to email me or schedule a chat; 2) For highly motivated early-stage researchers: I would be happy to provide guidance and discuss ideas. Feel free to reach out.

Seeking Postdoc opportunities. If you are interested in my research and have relevant opportunities, please feel free to contact me.

Selected Papers

  1. To Be Submitted
    Scoring, Reasoning, and Selecting the Best! Ensembling Large Language Models via a Peer-Review Process
    Zhijun Chen, Zeyu Ji, Qianren Mao, and 11 more authors
    To be submitted to arXiv soon, 2025
  2. arXiv 2025
    Harnessing Multiple Large Language Models: A Survey on LLM Ensemble
    Zhijun Chen, Jingzheng Li, Pengpeng Chen, and 10 more authors
    arXiv preprint arXiv:2502.18036, 2025
  3. AAAI 2025
    Implicit Word Reordering with Knowledge Distillation for Cross-Lingual Dependency Parsing
    Zhuoran Li, Chunming Hu, Junfan Chen, and 2 more authors
    In Proceedings of the AAAI Conference on Artificial Intelligence, 2025
  4. arXiv 2025
    Privacy-Preserving Federated Embedding Learning for Localized Retrieval-Augmented Generation
    Qianren Mao, Qili Zhang, Hanwen Hao, and 8 more authors
    arXiv preprint arXiv:2504.19101, 2025
  5. arXiv 2024
    XRAG: eXamining the Core–Benchmarking Foundational Components in Advanced Retrieval-Augmented Generation
    Qianren Mao, Yangyifei Luo, Qili Zhang, and 8 more authors
    arXiv preprint arXiv:2412.15529, 2024
  6. IJCAI 2024
    Improving Zero-Shot Cross-Lingual Transfer via Progressive Code-Switching
    Zhuoran Li, Chunming Hu, Junfan Chen, and 3 more authors
    arXiv preprint arXiv:2406.13361, 2024
  7. KDD 2023
    Neural-Hidden-CRF: A Robust Weakly-Supervised Sequence Labeler
    Zhijun Chen, Hailong Sun, Wanhao Zhang, and 3 more authors
    In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023
  8. ICDE 2023
    Learning from Noisy Crowd Labels with Logics
    Zhijun Chen, Hailong Sun, Haoqian He, and 1 more author
    In 2023 IEEE 39th International Conference on Data Engineering (ICDE), 2023
  9. IJCAI 2023
    Black-Box Data Poisoning Attacks on Crowdsourcing
    Pengpeng Chen, Yongqiang Yang, Dingqi Yang, and 3 more authors
    In IJCAI, 2023
  10. AAAI 2022
    Adversarial Learning from Crowds
    Pengpeng Chen, Hailong Sun, Yongqiang Yang, and 1 more author
    In Proceedings of the AAAI Conference on Artificial Intelligence, 2022
  11. IJCAI 2020
    Structured Probabilistic End-to-End Learning from Crowds
    Zhijun Chen, Huimin Wang, Hailong Sun, and 4 more authors
    In Proceedings of the twenty-ninth international conference on international joint conferences on artificial intelligence, 2021