LawShift: Benchmarking Legal Judgment Prediction Under Statute Shifts

Zhuo Han, Yi Feng, Yi Yang, Wanhong Huang, Ding Xuxing, Chuanyi Li, Jidong Ge, and Vincent Ng.
39th Annual Conference on Neural Information Processing Systems, 2025.

Click here for the PDF version.

Abstract

Legal Judgment Prediction (LJP) seeks to predict case outcomes given available case information, offering practical value for both legal professionals and laypersons. However, a key limitation of existing LJP models is their limited adaptability to statutory revisions. Current SOTA models are neither designed nor evaluated for statutory revisions. To bridge this gap, we introduce LawShift, a benchmark dataset for evaluating LJP under statutory revisions. Covering 31 fine-grained change types, LawShift enables systematic assessment of SOTA models' ability to handle legal changes. We evaluate five representative SOTA models on LawShift, uncovering significant limitations in their response to legal updates. Our findings show that model architecture plays a critical role in adaptability, offering actionable insights and guiding future research on LJP in dynamic legal contexts.

BibTeX entry

@InProceedings{Han+etal:25a,
  author = {Zhuo Han and Yi Feng and Yi Yang and Wanhong Huang and Ding Xuxing and Chuanyi Li and Jidong Ge and Vincent Ng},
  title = {LawShift: Benchmarking Legal Judgment Prediction Under Statute Shifts},
  booktitle = {39th Annual Conference on Neural Information Processing Systems},

  year = 2025}