I work at ByteDance as a AI researcher now in Shanghai, China.

I am now working on NLP, multi-modality, LLM-related foundational research, particular in LLM speedup, KIE and document AI. If you are seeking any form of academic cooperation, please feel free to email me at lujinghui@bytedance.com. We are hiring interns!

I graduated with a Bachelor’s degree in Software Engineering from Xiamen University (厦门大学) and earned a Master’s degree from the School of Computer Science at University College Dublin (都柏林大学), where I was advised by Prof. Mark Keane, focusing on natural language processing. Then, I completed my PhD at University College Dublin under the supervision of Prof. Brian Mac Namee, focusing on interdisciplinary methods for Deep Learning, NLP and Human-in-the-loop. I also collaborate with Dr. Irene Li from Li Lab, the University of Tokyo closely.

My research interest includes NLP, multi-modality, document AI and human-in-the-loop. I have published 20+ papers at the top international AI and NLP conferences such as NeurIPS, ACL, NAACL, AAAI, COLING.

🔥 News

  • 2024.10:  🎉 One first-author paper has been accepted to NeurIPS 2024.
  • 2024.07:  🎉 We released LayTextLLM, interleaving layout and text information for LLM, achieving SOTA performance across multiple document AI benchmarks | Project | | Hugging Face.
  • 2024.05:  🎉 We released MTVQA, which is the first text-centric multilingual multimodality benchmark for LLMs | Project | | Hugging Face.
  • 2024.05:  🎉 One paper has been accepted to ACL 2024 findings.
  • 2024.03:  🎉 One paper has been accepted to NAACL 2024.
  • 2024.02:  🎉 One co-first author paper has been accepted to COLING-LREC 2024.
  • 2023.08:  🎉 I join ByteDance as a AI researcher in Shanghai.
  • 2023.06:  🎉 One first-author paper has been accepted to AAAI 2023.
  • 2023.05:  🎉 One first-author paper has been accepted to ACL 2023.
  • 2023.05:  🎉 One paper has been accepted to ACL 2023 findings.
  • 2022.03:  🎉 One first-author paper has been accepted to ACL 2022.
  • 2022.01:   I join SenseTime Group Ltd. as a AI researcher for NLP!
  • 2021.07:  🎉 One first-author paper has been accepted to ICML 2021 Workshop.

📝 Selected Publications

NeurIPS 2024

PaDeLLM-NER: Parallel Decoding in Large Language Models for Named Entity Recognition

Jinghui Lu *, Ziwei Yang *, Yanjie Wang *, Xuejing Liu, Brian Mac Namee, Can Huang ✉️

In this study, we aim to reduce generation latency for Named Entity Recognition (NER) with Large Language Models (LLMs). The main cause of high latency in LLMs is the sequential decoding process, which autoregressively generates all labels and mentions for NER, significantly increase the sequence length. To this end, we introduce Parallel Decoding in LLM for NE} (PaDeLLM-NER), a approach that integrates seamlessly into existing generative model frameworks without necessitating additional modules or architectural modifications. PaDeLLM-NER allows for the simultaneous decoding of all mentions, thereby reducing generation latency. Experiments reveal that PaDeLLM-NER significantly increases inference speed that is 1.76 to 10.22 times faster than the autoregressive approach for both English and Chinese. Simultaneously it maintains the quality of predictions as evidenced by the performance that is on par with the state-of-the-art across various datasets.

Paper Code

AAAI 2023

PUnifiedNER: a prompting-based unified NER system for diverse datasets

Jinghui Lu , Rui Zhao, Brian Mac Namee, Fei Tan ✉️

Much of named entity recognition (NER) research focuses on developing dataset-specific models based on data from the domain of interest, and a limited set of related entity types. This is frustrating as each new dataset requires a new model to be trained and stored. In this work, we present a “versatile” model—the Prompting-based Unified NER system (PUnifiedNER)—that works with data from different domains and can recognise up to 37 entity types simultaneously, and theoretically it could be as many as possible. By using prompt learning, PUnifiedNER is a novel approach that is able to jointly train across multiple corpora, implementing intelligent on-demand entity recognition. Experimental results show that PUnifiedNER leads to significant prediction benefits compared to dataset-specific models with impressively reduced model deployment costs. Furthermore, the performance of PUnifiedNER can achieve competitive or even better performance than state-of-the-art domain-specific methods for some datasets. We also perform comprehensive pilot and ablation studies to support in-depth analysis of each component in PUnifiedNER.

Paper Code

ACL 2023

What Makes Pre-trained Language Models Better Zero-shot Learners?

Jinghui Lu , Dongsheng Zhu, Weidong Han, Rui Zhao, Brian Mac Namee, Fei Tan ✉️

Current methods for prompt learning in zero-shot scenarios widely rely on a development set with sufficient human-annotated data to select the best-performing prompt template a posteriori. This is not ideal because in a real-world zero-shot scenario of practical relevance, no labelled data is available. Thus, we propose a simple yet effective method for screening reasonable prompt templates in zero-shot text classification: Perplexity Selection (Perplection). We hypothesize that language discrepancy can be used to measure the efficacy of prompt templates, and thereby develop a substantiated perplexity-based scheme allowing for forecasting the performance of prompt templates in advance. Experiments show that our method leads to improved prediction performance in a realistic zero-shot setting, eliminating the need for any labelled examples.

Paper

ACL 2022

A Rationale-Centric Framework for Human-in-the-loop Machine Learning

Jinghui Lu *, Linyi Yang *, Brian Mac Namee, Yue Zhang ✉️

We present a novel rational-centric framework with human-in-the-loop – Rationales-centric Double-robustness Learning (RDL) – to boost model out-of-distribution performance in few-shot learning scenarios. By using static semi-factual generation and dynamic human-intervened correction, RDL, acting like a sensible “inductive bias”, exploits rationales (i.e. phrases that cause the prediction), human interventions and semi-factual augmentations to decouple spurious associations and bias models towards generally applicable underlying distributions, which enables fast and accurate generalisation. Experimental results show that RDL leads to significant prediction benefits on both in-distribution and out-of-distribution tests, especially for few-shot learning scenarios, compared to many state-of-the-art benchmarks.

Paper Code

🎖 Honors and Awards

  • 2024.06, Award for High-Level Overseas Talents, Shanghai, China
  • 2023.01, Outstanding Project Annual, SCG, SenseTime Group Ltd. China
  • 2016.09, Walsh Fellowship Award, Ireland.
  • 2015.09, Graduate Global Scholarship (<0.5\%), University College Dublin, Ireland

📖 Educations

  • 2016.09 - 2021.01, Phd, University College Dublin, Ireland.
  • 2015.09 - 2016.06, Master, University College Dublin, Ireland.
  • 2011.09 - 2015.06, Undergraudate, Xiamen University, China.
  • 2008.09 - 2011.06, Fuzhou No. 1 Middle School, China.

💬 Invited Talks

  • 2023.01, ChatGPT’s past, present and future (CCB Trust, internal talk, Beijing, China)
  • 2022.11, Diffusion Model and its Applications (SuzumuraLab tutorial, the Univeristy of Tokyo), Slide, Post

💻 Community Service

  • 2020 - Now, serve as a review at ACL, AAAI, ICLR, NAACL, COLING, EMNLP, NeurIPS, etc.