[2410.04949] Leverage Knowledge Graph and Large Language Model for Law Article Recommendation: A Case Study of Chinese Criminal Law

[2410.04949] Leverage Knowledge Graph and Large Language Model for Law Article Recommendation: A Case Study of Chinese Criminal Law

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2410.04949: Leverage Knowledge Graph and Large Language Model for Law Article Recommendation: A Case Study of Chinese Criminal Law

Computer Science > Information Retrieval arXiv:2410.04949 (cs) [Submitted on 7 Oct 2024 (v1), last revised 3 Mar 2026 (this version, v3)] Title:Leverage Knowledge Graph and Large Language Model for Law Article Recommendation: A Case Study of Chinese Criminal Law Authors:Yongming Chen, Miner Chen, Ye Zhu, Juan Pei, Siyu Chen, Yu Zhou, Yi Wang, Yifan Zhou, Hao Li, Songan Zhang View a PDF of the paper titled Leverage Knowledge Graph and Large Language Model for Law Article Recommendation: A Case Study of Chinese Criminal Law, by Yongming Chen and 9 other authors View PDF HTML (experimental) Abstract:Judicial efficiency is critical to social stability. However, in many countries worldwide, grassroots courts face substantial case backlogs, and judicial decisions remain heavily dependent on judges' cognitive efforts, with insufficient intelligent tools to enhance efficiency. To address this issue, we propose a highly efficient law article recommendation approach combining a Knowledge Graph (KG) and a Large Language Model (LLM). First, we construct a Case-Enhanced Law Article Knowledge Graph (CLAKG) to store current law articles, historical case information, and their interconnections, alongside an LLM-based automated construction method. Building on this, we propose a closed-loop law article recommendation framework integrating graph embedding-based retrieval and KG-grounded LLM reasoning. Experiments on judgment documents from China Judgments Online demonstrate that our method ...

Originally published on March 04, 2026. Curated by AI News.

Related Articles

Llms

[R] Depth-first pruning transfers: GPT-2 → TinyLlama with stable gains and minimal loss

TL;DR: Removing the right layers (instead of shrinking all layers) makes transformer models ~8–12% smaller with only ~6–8% quality loss, ...

Reddit - Machine Learning · 1 min ·
Llms

Built a training stability monitor that detects instability before your loss curve shows anything — open sourced the core today

Been working on a weight divergence trajectory curvature approach to detecting neural network training instability. Treats weight updates...

Reddit - Artificial Intelligence · 1 min ·
Llms

This Is Not Hacking. This Is Structured Intelligence.

Watch me demonstrate everything I've been talking about—live, in real time. The Setup: Maestro University AI enrollment system Standard c...

Reddit - Artificial Intelligence · 1 min ·
Llms

[D] Howcome Muon is only being used for Transformers?

Muon has quickly been adopted in LLM training, yet we don't see it being talked about in other contexts. Searches for Muon on ConvNets tu...

Reddit - Machine Learning · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime