Crawler Summary

citation-audit answer-first brief

Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consistency, data source citation, checking if a paper is real, 文献审查, 引用审查, 参考文献检查, 伪造文献, 投稿前检查, DOI核对, 软件版本核对, 数据源引用. Applicable file types: .docx, .tex, .bib, .ris, .enl, .nbib, manuscript files. --- name: citation-audit description: > Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consi Published capability contract available. No trust telemetry is available yet. Last updated 3/1/2026.

Freshness

Last checked 3/1/2026

Best For

Contract is available with explicit auth and schema references.

Not Ideal For

citation-audit is not ideal for teams that need stronger public trust telemetry, lower setup complexity, or more explicit contract coverage before production rollout.

Evidence Sources Checked

editorial-content, capability-contract, runtime-metrics, public facts pack

Claim this agent
Agent DossierGitHubSafety: 89/100

citation-audit

Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consistency, data source citation, checking if a paper is real, 文献审查, 引用审查, 参考文献检查, 伪造文献, 投稿前检查, DOI核对, 软件版本核对, 数据源引用. Applicable file types: .docx, .tex, .bib, .ris, .enl, .nbib, manuscript files. --- name: citation-audit description: > Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consi

OpenClawself-declared

Public facts

6

Change events

1

Artifacts

0

Freshness

Mar 1, 2026

Verifiededitorial-contentNo verified compatibility signals

Published capability contract available. No trust telemetry is available yet. Last updated 3/1/2026.

Schema refs publishedTrust evidence available

Trust score

Unknown

Compatibility

OpenClaw

Freshness

Mar 1, 2026

Vendor

Chinelytra

Artifacts

0

Benchmarks

0

Last release

Unpublished

Executive Summary

Key links, install path, and a quick operational read before the deeper crawl record.

Verifiededitorial-content

Summary

Published capability contract available. No trust telemetry is available yet. Last updated 3/1/2026.

Setup snapshot

git clone https://github.com/Chinelytra/academic-citation-audit-skill.git
  1. 1

    Setup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.

  2. 2

    Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.

Evidence Ledger

Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.

Verifiededitorial-content
Vendor (1)

Vendor

Chinelytra

profilemedium
Observed Mar 1, 2026Source linkProvenance
Compatibility (2)

Protocol compatibility

OpenClaw

contractmedium
Observed Feb 24, 2026Source linkProvenance

Auth modes

api_key

contracthigh
Observed Feb 24, 2026Source linkProvenance
Artifact (1)

Machine-readable schemas

OpenAPI or schema references published

contracthigh
Observed Feb 24, 2026Source linkProvenance
Security (1)

Handshake status

UNKNOWN

trustmedium
Observed unknownSource linkProvenance
Integration (1)

Crawlable docs

6 indexed pages on the official domain

search_documentmedium
Observed Apr 15, 2026Source linkProvenance

Release & Crawl Timeline

Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.

Self-declaredagent-index

Artifacts Archive

Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.

Self-declaredGITHUB OPENCLEW

Extracted files

0

Examples

3

Snippets

0

Languages

typescript

Parameters

Executable Examples

r

pkgs <- c('ecospat', 'biomod2', 'terra', 'sf')
for (p in pkgs) cat(sprintf("%-12s %s\n", p, packageVersion(p)))
cat(sprintf("%-12s %s\n", "R", R.version.string))

python

import pkg_resources, sys
for p in ['numpy', 'pandas', 'scikit-learn', 'tensorflow']:
    try: print(f"{p:20s} {pkg_resources.get_distribution(p).version}")
    except: print(f"{p:20s} NOT INSTALLED")
print(f"{'Python':20s} {sys.version.split()[0]}")

markdown

# Citation Audit Report / 参考文献审查报告

## 🔴 Must-fix errors / 必须修正
(Ordered: fabricated > missing > bibliographic)

## 🟡 Recommended improvements / 建议改进
(Appropriateness, formatting)

## ✅ Verified entries / 已验证通过
(Full checklist with per-entry status)

Docs & README

Full documentation captured from public sources, including the complete README when available.

Self-declaredGITHUB OPENCLEW

Docs source

GITHUB OPENCLEW

Editorial quality

ready

Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consistency, data source citation, checking if a paper is real, 文献审查, 引用审查, 参考文献检查, 伪造文献, 投稿前检查, DOI核对, 软件版本核对, 数据源引用. Applicable file types: .docx, .tex, .bib, .ris, .enl, .nbib, manuscript files. --- name: citation-audit description: > Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consi

Full README

name: citation-audit description: > Systematic audit of academic manuscript references: authenticity verification, bibliographic accuracy, citation appropriateness, and software/data version consistency. Triggers on: citation audit, reference check, bibliography verification, fabricated/fake/hallucinated reference detection, DOI verification, pre-submission check, manuscript review, R/Python package version consistency, data source citation, checking if a paper is real, 文献审查, 引用审查, 参考文献检查, 伪造文献, 投稿前检查, DOI核对, 软件版本核对, 数据源引用. Applicable file types: .docx, .tex, .bib, .ris, .enl, .nbib, manuscript files.

Citation Audit / 学术文献审查

Systematic audit of all references in an academic manuscript before submission. 投稿前对学术稿件参考文献进行系统性全面审查。

Audit Dimensions / 审查维度

| Level | Scope / 范围 | Severity / 严重度 | | ----- | ------------ | ----------------- | | L1 | Authenticity — does the paper exist? Is the DOI correct? / 真实性——论文是否存在?DOI 是否正确? | 🔴 Fatal | | L2 | Bibliographic accuracy — authors, year, volume, pages, journal / 书目信息——作者、年份、卷号、页码、期刊 | 🔴 Critical | | L3 | Text–list consistency — every in-text citation has a matching entry and vice versa / 正文与列表一致性 | 🟡 Important | | L4 | Citation appropriateness — each citation supports the claim it is attached to / 引用恰当性 | 🟡 Improvement | | L5 | Formatting & version consistency — style uniformity, software/data versions match actual usage / 格式与版本一致性 | ⚪ Housekeeping |

Workflow / 工作流程

Phase 1: Extract manuscript text / 提取稿件全文

Extract all text with paragraph indices for cross-referencing. See scripts/extract_docx.py.

For .tex files, parse directly. For .docx, use the python-docx library. Separate the reference list from the body text and index each entry.

Phase 2: L1 — Authenticity verification / 真实性验证

[!CAUTION] AI-assisted writing frequently introduces "hallucinated" references — DOIs that resolve to unrelated papers, or entirely fabricated entries. This is the most severe error class.

AI 辅助写作极易引入"幻觉文献"。此类错误一旦发表后果严重。

Method / 方法: CrossRef API + web search dual verification

  1. Run scripts/crossref_batch_check.py to batch-query CrossRef API metadata.
  2. Mandatory web-search re-verification for:
    • Entries where API results mismatch the manuscript
    • Connection errors or timeouts
    • Papers published within the last 1–2 years (CrossRef indexing lag)
    • Any citation that "looks too perfect" but cannot be independently found

Red flags for fabricated references / 伪造文献特征:

  • DOI resolves to an unrelated paper
  • Author + year + journal combination yields zero Google Scholar results
  • Claims to cite a "preprint" but provides a formal journal DOI

Verification chain for suspicious entries / 可疑条目验证链:

  1. Resolve DOI directly → check title and author match
  2. Google Scholar: search author + keywords
  3. Author's personal page / ORCID publication list
  4. Journal website: browse the table of contents for the cited volume/issue

Phase 3: L2 — Bibliographic accuracy / 书目信息核对

Check every entry against its verified source for:

| Field | Common errors / 常见错误 | | ----- | ----------------------- | | Authors | Missing co-authors (especially 4th+), wrong initials (G.H. vs C.H.) / 遗漏合著者、名缩写错误 | | Year | Early Online vs. official publication date confusion / 在线优先与正式出版日期混淆 | | Journal | Abbreviated vs. full name inconsistency / 缩写不统一 | | Volume/Pages | Mismatch with DOI record / 与 DOI 记录不符 | | DOI | Placeholder not replaced (e.g. zenodo.XXXXXXX), points to wrong article / 占位符未替换 |

Phase 4: L3 — Text–list cross-check / 正文-列表交叉核对

  1. Extract all (Author, Year) and (Author et al., Year) citations from the body text.
  2. Match bidirectionally:
    • In text → not in list = missing reference (must add) / 缺失引用
    • In list → not in text = orphan reference (delete or cite) / 幽灵引用
  3. Special attention to data sources, software packages, and datasets that are mentioned in text but absent from the reference list.

Phase 5: L4 — Citation appropriateness / 引用恰当性

Evaluate each citation:

  • Does it directly support the claim it is attached to?
  • Is there a more canonical or more recent alternative?
  • Excessive self-citation or citation stacking?

Phase 6: L5 — Formatting & version consistency / 格式与版本一致性

Style uniformity / 格式统一

  • "et al." usage, punctuation, spacing
  • Author name ordering for multi-work citations

Software & package version verification / 软件版本核对

[!IMPORTANT] The manuscript MUST report the actual software versions used for the analysis, not the latest CRAN/PyPI versions.

R environment:

pkgs <- c('ecospat', 'biomod2', 'terra', 'sf')
for (p in pkgs) cat(sprintf("%-12s %s\n", p, packageVersion(p)))
cat(sprintf("%-12s %s\n", "R", R.version.string))

Python environment:

import pkg_resources, sys
for p in ['numpy', 'pandas', 'scikit-learn', 'tensorflow']:
    try: print(f"{p:20s} {pkg_resources.get_distribution(p).version}")
    except: print(f"{p:20s} NOT INSTALLED")
print(f"{'Python':20s} {sys.version.split()[0]}")

Other environments (Julia, MATLAB, etc.): adapt the pattern to query installed package versions.

Cross-check steps:

  1. Search project scripts for all library() / import / using calls.
  2. Query actual installed versions in the runtime environment.
  3. Compare with versions stated in the manuscript and reference list.
  4. Flag packages mentioned in the manuscript but never called in any script (may indicate a method–code mismatch).

Data source & dataset citation / 数据源引用核对

Verify that every external data source used in the analysis is properly cited:

| Data type | Examples | What to check | | --------- | -------- | ------------- | | Remote sensing | MODIS, Landsat, Sentinel | Product name, version, DOI or data center URL | | Climate data | WorldClim, CHELSA, ERA5 | Version number, resolution, temporal coverage | | Biodiversity records | GBIF, iNaturalist, VertNet | Download DOI, access date, query parameters | | Geospatial layers | Natural Earth, GADM, OpenStreetMap | Version, access date | | Genomic data | GenBank, SRA, ENA | Accession numbers | | Statistical databases | World Bank, UN, national bureaus | Dataset name, access date, URL |

Common issues:

  • Dataset is used in methods but has no reference entry
  • DOI or accession number is a placeholder
  • Version mismatch between what was downloaded and what is cited

Output format / 输出格式

Generate a citation_audit.md report structured as:

# Citation Audit Report / 参考文献审查报告

## 🔴 Must-fix errors / 必须修正
(Ordered: fabricated > missing > bibliographic)

## 🟡 Recommended improvements / 建议改进
(Appropriateness, formatting)

## ✅ Verified entries / 已验证通过
(Full checklist with per-entry status)

Key lessons / 关键经验

  1. Never trust CrossRef alone — its "best match" is frequently wrong for books, chapters, datasets, non-English literature, and same-surname authors. Always web-search verify. CrossRef 返回的"最佳匹配"经常是错误的,必须用 Web 搜索二次验证。

  2. Year discrepancies need judgment — "Early Online" vs. print dates can differ by 1–2 years; both are acceptable. Differences > 2 years likely indicate a real error. 年份差异需判断:Early Online 与正式出版差 1–2 年属正常。

  3. Methods must match code — if the manuscript claims package X was used but the scripts call package Y, this is a reviewable error. Cross-check Methods section against actual scripts line by line. 稿件方法描述必须与代码一致,需逐行比对。

  4. Data sources need citations too — remote sensing products, climate databases, and biodiversity data portals all require proper citation with DOI/version/access date. 数据源也需要规范引用。

Anti-patterns

| Don't / 不要 | Do instead / 应该 | | ------------ | ---------------- | | Trust CrossRef blindly | CrossRef + web search dual verification | | Ignore recent publications | Extra scrutiny for papers < 2 years old | | Assume all DOIs are correct | Resolve every DOI and verify the target | | Only check the reference list | Also cross-check body citations and code | | Report everything at once | Triage by severity: fatal → critical → improvement | | Skip data source citations | Verify every dataset, layer, and product is cited |

Contract & API

Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.

Verifiedcapability-contract

Contract coverage

Status

ready

Auth

api_key

Streaming

No

Data region

global

Protocol support

OpenClaw: self-declared

Requires: openclew, lang:typescript

Forbidden: none

Guardrails

Operational confidence: medium

Contract is available with explicit auth and schema references.
Trust confidence is not low and verification freshness is acceptable.
Invocation examples
curl -s "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/snapshot"
curl -s "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract"
curl -s "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/trust"

Reliability & Benchmarks

Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.

Missingruntime-metrics

Trust signals

Handshake

UNKNOWN

Confidence

unknown

Attempts 30d

unknown

Fallback rate

unknown

Runtime metrics

Observed P50

unknown

Observed P95

unknown

Rate limit

unknown

Estimated cost

unknown

No benchmark suites or observed failure patterns are available.

Media & Demo

Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.

Missingno-media
No screenshots, media assets, or demo links are available.

Related Agents

Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.

Self-declaredprotocol-neighbors
GITHUB_REPOSactivepieces

Rank

70

AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents

Traction

No public download signal

Freshness

Updated 2d ago

OPENCLAW
GITHUB_REPOScherry-studio

Rank

70

AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs

Traction

No public download signal

Freshness

Updated 5d ago

MCPOPENCLAW
GITHUB_REPOSAionUi

Rank

70

Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!

Traction

No public download signal

Freshness

Updated 6d ago

MCPOPENCLAW
GITHUB_REPOSCopilotKit

Rank

70

The Frontend for Agents & Generative UI. React + Angular

Traction

No public download signal

Freshness

Updated 23d ago

OPENCLAW
Machine Appendix

Contract JSON

{
  "contractStatus": "ready",
  "authModes": [
    "api_key"
  ],
  "requires": [
    "openclew",
    "lang:typescript"
  ],
  "forbidden": [],
  "supportsMcp": false,
  "supportsA2a": false,
  "supportsStreaming": false,
  "inputSchemaRef": "https://github.com/Chinelytra/academic-citation-audit-skill#input",
  "outputSchemaRef": "https://github.com/Chinelytra/academic-citation-audit-skill#output",
  "dataRegion": "global",
  "contractUpdatedAt": "2026-02-24T19:45:25.730Z",
  "sourceUpdatedAt": "2026-02-24T19:45:25.730Z",
  "freshnessSeconds": 4421143
}

Invocation Guide

{
  "preferredApi": {
    "snapshotUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/snapshot",
    "contractUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "trustUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/trust"
  },
  "curlExamples": [
    "curl -s \"https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/snapshot\"",
    "curl -s \"https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract\"",
    "curl -s \"https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/trust\""
  ],
  "jsonRequestTemplate": {
    "query": "summarize this repo",
    "constraints": {
      "maxLatencyMs": 2000,
      "protocolPreference": [
        "OPENCLEW"
      ]
    }
  },
  "jsonResponseTemplate": {
    "ok": true,
    "result": {
      "summary": "...",
      "confidence": 0.9
    },
    "meta": {
      "source": "GITHUB_OPENCLEW",
      "generatedAt": "2026-04-16T23:51:08.983Z"
    }
  },
  "retryPolicy": {
    "maxAttempts": 3,
    "backoffMs": [
      500,
      1500,
      3500
    ],
    "retryableConditions": [
      "HTTP_429",
      "HTTP_503",
      "NETWORK_TIMEOUT"
    ]
  }
}

Trust JSON

{
  "status": "unavailable",
  "handshakeStatus": "UNKNOWN",
  "verificationFreshnessHours": null,
  "reputationScore": null,
  "p95LatencyMs": null,
  "successRate30d": null,
  "fallbackRate": null,
  "attempts30d": null,
  "trustUpdatedAt": null,
  "trustConfidence": "unknown",
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Capability Matrix

{
  "rows": [
    {
      "key": "OPENCLEW",
      "type": "protocol",
      "support": "unknown",
      "confidenceSource": "profile",
      "notes": "Listed on profile"
    },
    {
      "key": "differ",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    },
    {
      "key": "the",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    }
  ],
  "flattenedTokens": "protocol:OPENCLEW|unknown|profile capability:differ|supported|profile capability:the|supported|profile"
}

Facts JSON

[
  {
    "factKey": "docs_crawl",
    "category": "integration",
    "label": "Crawlable docs",
    "value": "6 indexed pages on the official domain",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true
  },
  {
    "factKey": "vendor",
    "category": "vendor",
    "label": "Vendor",
    "value": "Chinelytra",
    "href": "https://github.com/Chinelytra/academic-citation-audit-skill",
    "sourceUrl": "https://github.com/Chinelytra/academic-citation-audit-skill",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-03-01T06:05:17.732Z",
    "isPublic": true
  },
  {
    "factKey": "protocols",
    "category": "compatibility",
    "label": "Protocol compatibility",
    "value": "OpenClaw",
    "href": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "sourceUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "sourceType": "contract",
    "confidence": "medium",
    "observedAt": "2026-02-24T19:45:25.730Z",
    "isPublic": true
  },
  {
    "factKey": "auth_modes",
    "category": "compatibility",
    "label": "Auth modes",
    "value": "api_key",
    "href": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "sourceUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "sourceType": "contract",
    "confidence": "high",
    "observedAt": "2026-02-24T19:45:25.730Z",
    "isPublic": true
  },
  {
    "factKey": "schema_refs",
    "category": "artifact",
    "label": "Machine-readable schemas",
    "value": "OpenAPI or schema references published",
    "href": "https://github.com/Chinelytra/academic-citation-audit-skill#input",
    "sourceUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/contract",
    "sourceType": "contract",
    "confidence": "high",
    "observedAt": "2026-02-24T19:45:25.730Z",
    "isPublic": true
  },
  {
    "factKey": "handshake_status",
    "category": "security",
    "label": "Handshake status",
    "value": "UNKNOWN",
    "href": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/trust",
    "sourceUrl": "https://xpersona.co/api/v1/agents/chinelytra-academic-citation-audit-skill/trust",
    "sourceType": "trust",
    "confidence": "medium",
    "observedAt": null,
    "isPublic": true
  }
]

Change Events JSON

[
  {
    "eventType": "docs_update",
    "title": "Docs refreshed: Sign in to GitHub · GitHub",
    "description": "Fresh crawlable documentation was indexed for the official domain.",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true
  }
]

Sponsored

Ads related to citation-audit and adjacent AI workflows.