OpenNovelty: An LLM-powered Agentic System for Verifiable Scholarly Novelty Assessment

Published in arXiv preprint, 2026

Ming Zhang * , Kexin Tan * , Yueyuan Huang * , Yujiong Shen , Chunchun Ma , Li Ju , Xinran Zhang , Yuhui Wang , Wenqing Jing , Jingyi Deng , Huayu Sha , Binze Hu , Jingqi Tong , Changhao Jiang , Yage Geng , Yuankai Ying , Yue Zhang , Zhangyue Yin , Zhiheng Xi , Shihan Dou , Tao Gui , Qi Zhang , Xuanjing Huang

* 共同一作;† 通讯作者

OpenNovelty focuses on a difficult but essential part of research evaluation: verifiable novelty assessment with explicit evidence traces.

OpenNovelty pipeline

Why this system matters

Novelty review is usually time-constrained, inconsistent across reviewers, and highly dependent on retrieval coverage. OpenNovelty reframes this as a reproducible pipeline problem instead of a one-shot LLM judgment.

Four-phase pipeline

The public repository describes a staged workflow:

  1. Phase I — Information Extraction
    Extract paper text, core task, and contribution claims.
  2. Phase II — Literature Retrieval
    Retrieve related work candidates and build citation indices.
  3. Phase III — Deep Analysis
    Compare claims with retrieved literature and classify novelty evidence.
  4. Phase IV — Report Generation
    Export structured novelty reports (Markdown/PDF) with citations and snippets.

Output artifacts

Typical outputs include:

  • phase1_extracted.json
  • citation_index.json
  • phase3_complete_report.json
  • final novelty report (.md / .pdf)

This makes the full process auditable, with intermediate artifacts available for debugging and review.

Quick-start workflow

The repository provides script entrypoints for each phase:

# Phase 1
python scripts/run_phase1_batch.py --papers "<paper-url>" --out-root output/demo --force-year 2026

# Phase 2
bash scripts/run_phase2_concurrent.sh <paper_id> --base-dir output/demo

# Phase 3
bash scripts/run_phase3_all.sh output/demo/<paper_id>

# Phase 4
bash scripts/run_phase4.sh output/demo/<paper_id>

Engineering notes from the repo

  • Python 3.8+ environment
  • modular scripts for batch and single-paper workflows
  • retrieval, analysis, and rendering decoupled by intermediate JSON artifacts
  • some external service dependencies are marked as evolving in the current release

Practical takeaway

OpenNovelty is useful when you want traceable novelty review, especially for internal pre-review, large-scale triage, or evidence-grounded reviewer assistance where “why this is novel (or not)” must be inspectable.

Citation

@article{abs-2601-01576,
  author       = {Ming Zhang and
                  Kexin Tan and
                  Yueyuan Huang and
                  Yujiong Shen and
                  Chunchun Ma and
                  Li Ju and
                  Xinran Zhang and
                  Yuhui Wang and
                  Wenqing Jing and
                  Jingyi Deng and
                  Huayu Sha and
                  Binze Hu and
                  Jingqi Tong and
                  Changhao Jiang and
                  Yage Geng and
                  Yuankai Ying and
                  Yue Zhang and
                  Zhangyue Yin and
                  Zhiheng Xi and
                  Shihan Dou and
                  Tao Gui and
                  Qi Zhang and
                  Xuanjing Huang},
  title        = {OpenNovelty: An LLM-powered Agentic System for Verifiable Scholarly
                  Novelty Assessment},
  journal      = {CoRR},
  volume       = {abs/2601.01576},
  year         = {2026},
  url          = {https://doi.org/10.48550/arXiv.2601.01576},
  doi          = {10.48550/ARXIV.2601.01576},
  eprinttype   = {arXiv},
  eprint       = {2601.01576},
  biburl       = {https://dblp.org/rec/journals/corr/abs-2601-01576.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}