Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Crawler Summary
Agent-aware cost intelligence for AI. Per-agent, per-task cost attribution, waste detection, and routing recommendations. Works with LiteLLM, LangGraph, CrewAI, Google ADK. AgentLedger **Your agents are burning money. AgentLedger shows you exactly where.** $1 $1 $1 $1 --- **LiteLLM tells you how much you spent. AgentLedger tells you *why*.** Per-agent. Per-task. With waste detection and savings recommendations. --- Table of Contents - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 --- The Problem A typical unoptimized multi-agent system proc Capability contract not published. No trust telemetry is available yet. Last updated 4/15/2026.
Freshness
Last checked 4/15/2026
Best For
AgentLedger is best for crewai, multi-agent workflows where OpenClaw compatibility matters.
Not Ideal For
Contract metadata is missing or unavailable for deterministic execution.
Evidence Sources Checked
editorial-content, GITHUB REPOS, runtime-metrics, public facts pack
Agent-aware cost intelligence for AI. Per-agent, per-task cost attribution, waste detection, and routing recommendations. Works with LiteLLM, LangGraph, CrewAI, Google ADK. AgentLedger **Your agents are burning money. AgentLedger shows you exactly where.** $1 $1 $1 $1 --- **LiteLLM tells you how much you spent. AgentLedger tells you *why*.** Per-agent. Per-task. With waste detection and savings recommendations. --- Table of Contents - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 --- The Problem A typical unoptimized multi-agent system proc
Public facts
4
Change events
1
Artifacts
0
Freshness
Apr 15, 2026
Capability contract not published. No trust telemetry is available yet. Last updated 4/15/2026.
Trust score
Unknown
Compatibility
OpenClaw
Freshness
Apr 15, 2026
Vendor
Nabeelkhial
Artifacts
0
Benchmarks
0
Last release
Unpublished
Key links, install path, and a quick operational read before the deeper crawl record.
Summary
Capability contract not published. No trust telemetry is available yet. Last updated 4/15/2026.
Setup snapshot
Setup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.
Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.
Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.
Vendor
Nabeelkhial
Protocol compatibility
OpenClaw
Handshake status
UNKNOWN
Crawlable docs
6 indexed pages on the official domain
Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.
Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.
Extracted files
0
Examples
6
Snippets
0
Languages
python
text
Your Agent Code AgentLedger SDK AgentLedger Server
┌─────────────────┐ ┌───────────────┐ ┌──────────────────┐
│ │ @track │ │ HTTP POST │ │
│ research-agent │──────────>│ Event Buffer │───────────>│ /api/v1/events │
│ task: search │ │ (batch of 50) │ │ │
│ model: sonnet │ │ │ │ PostgreSQL/SQLite│
│ tokens: 4200 │ │ Falls back to │ │ │
│ cost: $0.03 │ │ local .jsonl │ │ Waste Detector │
│ │ │ if server is │ │ Routing Advisor │
│ writer-agent │──────────>│ unreachable │ │ Budget Monitor │
│ task: draft │ │ │ │ │
│ model: opus │ └───────────────┘ └──────────────────┘
│ cost: $0.21 │ │
└─────────────────┘ v
CLI / Dashboard / APIbash
pip install agentledger
bash
cd agentledger-server pip install fastapi uvicorn sqlalchemy aiosqlite pydantic pydantic-settings greenlet PYTHONPATH=. uvicorn app.main:app --port 8100
bash
docker compose up
bash
pip install agentledger-cli
bash
curl http://localhost:8100/health
Full documentation captured from public sources, including the complete README when available.
Docs source
GITHUB REPOS
Editorial quality
ready
Agent-aware cost intelligence for AI. Per-agent, per-task cost attribution, waste detection, and routing recommendations. Works with LiteLLM, LangGraph, CrewAI, Google ADK. AgentLedger **Your agents are burning money. AgentLedger shows you exactly where.** $1 $1 $1 $1 --- **LiteLLM tells you how much you spent. AgentLedger tells you *why*.** Per-agent. Per-task. With waste detection and savings recommendations. --- Table of Contents - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 - $1 --- The Problem A typical unoptimized multi-agent system proc
Your agents are burning money. AgentLedger shows you exactly where.
LiteLLM tells you how much you spent. AgentLedger tells you why.
Per-agent. Per-task. With waste detection and savings recommendations.
A typical unoptimized multi-agent system processes 10-50x more tokens than it needs to. Developers discover $180/month bills with zero visibility into which agent, which task, or which model drove those costs.
Existing tools (LiteLLM, Helicone, Langfuse) work at the API call level. They see individual requests. They don't understand agents.
| What LiteLLM Sees | What You Need |
|---|---|
| POST /v1/chat, model=claude-sonnet, 4200 tokens, $0.03 | research-agent > competitor-analysis > step 3 of 7, $0.03 |
| Total spend: $487/month | research-agent: $210, writer-agent: $89, code-agent: $188. Retries: $74 wasted |
| Avg cost per request: $0.04 | This task used Opus but Haiku would work. Savings: $540/month |
Your Agent Code AgentLedger SDK AgentLedger Server
┌─────────────────┐ ┌───────────────┐ ┌──────────────────┐
│ │ @track │ │ HTTP POST │ │
│ research-agent │──────────>│ Event Buffer │───────────>│ /api/v1/events │
│ task: search │ │ (batch of 50) │ │ │
│ model: sonnet │ │ │ │ PostgreSQL/SQLite│
│ tokens: 4200 │ │ Falls back to │ │ │
│ cost: $0.03 │ │ local .jsonl │ │ Waste Detector │
│ │ │ if server is │ │ Routing Advisor │
│ writer-agent │──────────>│ unreachable │ │ Budget Monitor │
│ task: draft │ │ │ │ │
│ model: opus │ └───────────────┘ └──────────────────┘
│ cost: $0.21 │ │
└─────────────────┘ v
CLI / Dashboard / API
Key concepts:
@track scopeThe SDK buffers events in memory and ships them in batches (default: every 50 events or 5 seconds). If the server is unreachable, events are written to a local .jsonl file as fallback.
pip install agentledger
Option A: Local dev (SQLite, zero dependencies)
cd agentledger-server
pip install fastapi uvicorn sqlalchemy aiosqlite pydantic pydantic-settings greenlet
PYTHONPATH=. uvicorn app.main:app --port 8100
Option B: Docker Compose (PostgreSQL + Redis)
docker compose up
Option C: Skip the server entirely
The SDK works standalone — events are written to agentledger_events.jsonl as fallback. Useful for local development.
pip install agentledger-cli
curl http://localhost:8100/health
# {"status":"ok","service":"agentledger"}
agentledger status
Best for: Custom agent code, any framework, quick starts.
The simplest integration. Wrap your agent functions with @track and record LLM calls.
from agentledger import track, ledger
ledger.init(
project="my-saas", # Groups agents under one project
server_url="http://localhost:8100", # Your AgentLedger server
# api_key="al_xxx", # Optional: auth key
# fallback_path="events.jsonl", # Optional: custom fallback path
# batch_size=50, # Ship every N events
# flush_interval=5.0, # Or every N seconds
)
@track(agent="research-agent", task="competitor-analysis")
def run_research(query: str) -> str:
"""Everything inside this function is attributed to research-agent."""
# Your existing code — call any LLM however you want
response = call_my_llm(query)
# Manually record the call
from agentledger.tracker import get_current_context
ctx = get_current_context()
ctx.record_call(
model="claude-sonnet-4-6",
tokens_in=1500,
tokens_out=800,
cost_usd=0.0165, # Or let AgentLedger compute from pricing DB
latency_ms=450.0,
prompt_hash="abc123", # Optional: enables retry detection
status="success", # or "error"
)
return response
Instead of calculating cost yourself:
from agentledger.pricing import calculate_cost
cost = calculate_cost("claude-sonnet-4-6", tokens_in=1500, tokens_out=800)
# Returns: 0.0165 (based on pricing/models.json)
For more granular control:
from agentledger import track_context
def run_pipeline():
with track_context(agent="research-agent", task="search") as ctx:
response = llm.call(...)
ctx.record_call(model="gpt-4o", tokens_in=500, tokens_out=200, cost_usd=0.003, latency_ms=300)
with track_context(agent="writer-agent", task="draft") as ctx:
response = llm.call(...)
ctx.record_call(model="claude-opus-4-6", tokens_in=2000, tokens_out=1000, cost_usd=0.105, latency_ms=800)
# At app shutdown — ships any remaining buffered events
ledger.flush()
Best for: Teams already using LiteLLM. Zero manual recording needed.
This is the recommended integration for businesses. If you use litellm.completion(), the callback automatically captures every call's model, tokens, cost, and latency — no manual record_call() needed.
import litellm
from agentledger import track, ledger
from agentledger.integrations.litellm import AgentLedgerCallback
# Initialize
ledger.init(project="my-saas")
litellm.callbacks = [AgentLedgerCallback()] # <-- one line, captures everything
# With @track — agent context comes from the decorator
@track(agent="research-agent", task="competitor-analysis")
def run_research(query: str) -> str:
# Every litellm.completion() call inside here is auto-tracked
response = litellm.completion(
model="claude-sonnet-4-6",
messages=[{"role": "user", "content": query}],
)
return response.choices[0].message.content
# Without @track — pass agent context via metadata
def run_standalone():
response = litellm.completion(
model="gpt-4o",
messages=[{"role": "user", "content": "Summarize revenue"}],
metadata={"agent": "finance-agent", "task": "quarterly-report"},
)
return response.choices[0].message.content
How it works under the hood:
async_log_success_event after every completionkwargs["response_cost"] (LiteLLM pre-calculates this)response_obj.usage.prompt_tokens and completion_tokens@track context is active, it records to that contextmetadata and records directly to the ledgerIf your team shares a LiteLLM Proxy, add AgentLedger to config.yaml:
# litellm_config.yaml
model_list:
- model_name: claude-sonnet
litellm_params:
model: anthropic/claude-sonnet-4-6
api_key: os.environ/ANTHROPIC_API_KEY
- model_name: gpt-4o
litellm_params:
model: openai/gpt-4o
api_key: os.environ/OPENAI_API_KEY
litellm_settings:
callbacks: agentledger.integrations.litellm.callback_instance
Start the proxy:
pip install litellm agentledger
litellm --config litellm_config.yaml
Every request through the proxy is now tracked. Developers tag their agents:
# Any team member's code
response = litellm.completion(
model="claude-sonnet",
messages=[...],
metadata={"agent": "my-agent", "task": "my-task"},
api_base="http://your-proxy:4000",
)
Best for: LangGraph / LangChain workflows.
from agentledger import track, ledger
from agentledger.integrations import langgraph_callback
ledger.init(project="my-saas")
callback = langgraph_callback()
# Compile your graph with the callback
app = workflow.compile(callbacks=[callback])
@track(agent="research-agent", task="web-search")
def run_agent(query: str):
return app.invoke({"query": query})
result = run_agent("Latest AI frameworks")
ledger.flush()
The callback intercepts on_llm_start, on_chat_model_start, on_llm_end, and on_llm_error from LangChain's callback system and extracts model, tokens, cost, and latency automatically.
Best for: CrewAI multi-agent setups.
from crewai import Agent, Task, Crew
from agentledger import track, ledger
from agentledger.integrations import crewai_callback
ledger.init(project="my-saas")
researcher = Agent(role="Researcher", goal="Find data", ...)
writer = Agent(role="Writer", goal="Write reports", ...)
@track(agent="crew-pipeline", task="research-and-write")
def run_crew():
crew = Crew(
agents=[researcher, writer],
tasks=[...],
callbacks=[crewai_callback()],
)
return crew.kickoff()
result = run_crew()
ledger.flush()
Best for: Google Agent Development Kit.
from agentledger import track, ledger
from agentledger.integrations import adk_callback
ledger.init(project="my-saas")
@track(agent="search-agent", task="web-search")
def run_adk_agent(query: str):
agent = Agent(
model="gemini-2.5-pro",
callbacks=[adk_callback()],
)
return agent.run(query)
Best for: Non-Python apps (Node.js, Go, Rust, etc.), or when you want full control.
Skip the SDK entirely and POST events directly.
curl -X POST http://localhost:8100/api/v1/events \
-H "Content-Type: application/json" \
-d '{
"events": [
{
"project": "my-saas",
"type": "llm_call",
"agent_name": "research-agent",
"task_name": "competitor-analysis",
"task_id": "task-001",
"step": 1,
"model": "claude-sonnet-4-6",
"tokens_in": 1500,
"tokens_out": 800,
"cost_usd": 0.0165,
"latency_ms": 450,
"prompt_hash": "a1b2c3d4",
"status": "success"
}
]
}'
# List agents with cost summaries
curl "http://localhost:8100/api/v1/agents?project=my-saas"
# Agent P&L (cost by task, model, and day)
curl "http://localhost:8100/api/v1/agents/research-agent/costs?project=my-saas"
# Waste flags
curl "http://localhost:8100/api/v1/waste?project=my-saas"
# Routing recommendations
curl "http://localhost:8100/api/v1/recommendations?project=my-saas"
# Full dashboard
curl "http://localhost:8100/api/v1/dashboard?project=my-saas"
| Field | Type | Required | Description |
|-------|------|----------|-------------|
| project | string | No (default: "default") | Project identifier |
| type | string | No (default: "llm_call") | llm_call or task_span |
| agent_name | string | Yes | Which agent made this call |
| task_name | string | No | What task this is part of |
| task_id | string | No | Groups calls within a task |
| step | int | No | Step number within a task |
| model | string | No | LLM model used |
| tokens_in | int | No (default: 0) | Input/prompt tokens |
| tokens_out | int | No (default: 0) | Output/completion tokens |
| cost_usd | float | No (default: 0) | Cost in USD (server computes from pricing DB if 0) |
| latency_ms | float | No (default: 0) | Response time in milliseconds |
| prompt_hash | string | No | Hash of prompt (for retry detection) |
| status | string | No (default: "success") | success, error, retry |
| error | string | No | Error message if failed |
| metadata | object | No | Any additional key-value pairs |
Every LLM call tagged to a specific agent and task. No more guessing which part of your pipeline is expensive.
Automatically flags waste patterns — see deep dive below.
Analyzes your actual usage patterns and tells you exactly which tasks can use cheaper models — with estimated monthly savings.
Set per-project or per-agent spending limits with webhook alerts — see setup below.
Not another gateway. AgentLedger is a cost intelligence layer that works WITH whatever you already use:
CustomLogger callback (SDK + Proxy)@track decorator + context managerdocker compose up and you own your data. No vendor lock-in. No usage-based pricing on your observability.
# Project-wide: $100/month, alert at 80%
curl -X POST http://localhost:8100/api/v1/budgets \
-H "Content-Type: application/json" \
-d '{
"project_id": "my-saas",
"limit_usd": 100,
"period": "monthly",
"alert_threshold_pct": 0.8,
"webhook_url": "https://hooks.slack.com/services/YOUR/WEBHOOK/URL"
}'
# Per-agent: research-agent gets $30/month
curl -X POST http://localhost:8100/api/v1/budgets \
-H "Content-Type: application/json" \
-d '{
"project_id": "my-saas",
"agent_name": "research-agent",
"limit_usd": 30,
"period": "monthly",
"alert_threshold_pct": 0.8,
"webhook_url": "https://hooks.slack.com/services/YOUR/WEBHOOK/URL"
}'
The budget monitor runs hourly and checks each budget:
alert_threshold_pct of limit_usd, fires a webhookWebhook payload (works with Slack, PagerDuty, Opsgenie, or any HTTP endpoint):
{
"text": "AgentLedger Budget Alert: research-agent in project 'my-saas' has used 85% of its monthly budget ($25.50 / $30.00)",
"project": "my-saas",
"agent": "research-agent",
"current_spend": 25.50,
"limit": 30.0,
"period": "monthly",
"pct_used": 85.0
}
The waste detector runs hourly and flags three patterns:
retry_loop)prompt_hash appears 3+ times within a single taskover_qualified_model)context_bloat)| Tier | Cost Range | Models | Use For | |------|-----------|--------|---------| | 1 (cheap) | $0.10-$1/M | Haiku, GPT-4o Mini, Gemini Flash, DeepSeek | Classification, extraction, simple Q&A | | 2 (mid) | $1-$5/M | Sonnet, GPT-4o, Gemini Pro | General tasks, summarization, coding | | 3 (expensive) | $10-$75/M | Opus, GPT-4 Turbo, o3 | Complex reasoning, analysis, creative work |
import asyncio
from app.database import async_session
from app.workers.waste_detector import run_waste_detection
from app.workers.routing_advisor import generate_recommendations
async def run_analysis():
async with async_session() as db:
waste_count = await run_waste_detection(db, "my-saas")
print(f"Found {waste_count} waste flags")
rec_count = await generate_recommendations(db, "my-saas")
print(f"Generated {rec_count} routing recommendations")
asyncio.run(run_analysis())
agentledger --project my-saas status # Dashboard overview
agentledger --project my-saas agents # Per-agent cost table
agentledger --project my-saas waste # Waste flags sorted by $ impact
agentledger --project my-saas recommend # Routing recommendations
# Override server URL
agentledger --server http://prod:8100 --project my-saas status
# Environment variables work too
export AGENTLEDGER_SERVER=http://prod:8100
export AGENTLEDGER_PROJECT=my-saas
agentledger status
| Method | Endpoint | Description |
|--------|----------|-------------|
| POST | /api/v1/events | Batch ingest events from SDK |
| GET | /api/v1/agents | List agents with cost summaries |
| GET | /api/v1/agents/{name}/costs | Per-agent P&L by task, model, and time |
| GET | /api/v1/waste | Waste flags sorted by $ impact |
| GET | /api/v1/recommendations | Model routing recommendations with savings |
| POST | /api/v1/budgets | Create/update budget guardrails |
| GET | /api/v1/budgets | List budgets with current spend |
| GET | /api/v1/dashboard | Full dashboard: spend, agents, waste, trends |
| GET | /health | Health check |
Interactive Swagger UI at http://localhost:8100/docs.
export AGENTLEDGER_API_KEY="your-secret-key"
export AGENTLEDGER_DATABASE_URL="postgresql+asyncpg://user:pass@host:5432/agentledger"
docker compose up -d
| Variable | Default | Description |
|----------|---------|-------------|
| AGENTLEDGER_DATABASE_URL | sqlite+aiosqlite:///./agentledger.db | Database connection string |
| AGENTLEDGER_REDIS_URL | redis://localhost:6379/0 | Redis for event queuing |
| AGENTLEDGER_API_KEY | None | Require auth on all API requests |
| AGENTLEDGER_CORS_ORIGINS | ["http://localhost:3000"] | Allowed CORS origins |
| AGENTLEDGER_BATCH_MAX_SIZE | 1000 | Max events per batch POST |
| AGENTLEDGER_LOG_LEVEL | INFO | Logging level |
ledger.init(
project="my-saas",
server_url="https://agentledger.internal.company.com",
api_key="al_production_key_here",
batch_size=100, # Larger batches = fewer HTTP calls
flush_interval=10.0, # Ship every 10 seconds
fallback_path="/var/log/agentledger_fallback.jsonl",
)
1. Your code calls @track(agent="x", task="y")
2. SDK creates a _TrackingContext on a thread-local stack
3. Inside the function, LLM calls are intercepted by framework callbacks
4. Each call → ctx.record_call() → ledger.record() → buffer
5. Buffer fills to batch_size OR flush_interval fires
6. SDK POSTs batch to /api/v1/events
7. Server computes cost (if not provided) from pricing DB
8. Server writes events to PostgreSQL/SQLite
9. Hourly workers analyze events:
- Waste detector flags retry loops, over-qualified models, context bloat
- Routing advisor finds cheaper model alternatives
- Budget monitor checks spend vs limits, fires webhooks
10. Results available via API, CLI, or dashboard
| Table | Purpose |
|-------|---------|
| events | Raw event log (immutable, append-only) |
| agent_summaries | Pre-aggregated cost/token summaries per agent per period |
| waste_flags | Waste detection results with $ estimates and fix suggestions |
| routing_recommendations | Model swap recommendations with monthly savings |
| budgets | Spending limits with webhook alert config |
| model_pricing | Per-model cost rates (synced from pricing/models.json) |
The tracking context uses threading.local() for thread-safe nested scoping:
@track(agent="orchestrator", task="pipeline")
def run_pipeline():
research = run_research("query") # inner @track → research-agent
report = write_report(research) # inner @track → writer-agent
Each call is attributed to the innermost active context.
If the server is unreachable, events are appended to a local JSON-lines file:
{"type":"llm_call","agent_name":"research-agent","model":"claude-sonnet-4-6","cost_usd":0.0165,...}
Replay later:
cat agentledger_events.jsonl | while read line; do
echo "{\"events\":[$line]}" | curl -X POST http://localhost:8100/api/v1/events \
-H "Content-Type: application/json" -d @-
done
AgentLedger/
agentledger-sdk/ # Python SDK (PyPI: agentledger)
agentledger/
integrations/ # LangGraph, CrewAI, ADK, LiteLLM callbacks
tracker.py # @track decorator + context manager
ledger.py # Event buffering + async shipping
pricing.py # Model cost calculator
agentledger-server/ # FastAPI backend
app/
routes/ # API endpoints
workers/ # Waste detector, routing advisor, budget monitor
models/ # SQLAlchemy models + Pydantic schemas
agentledger-cli/ # Terminal dashboard (rich tables)
pricing/ # Community-maintained model pricing DB (25+ models)
examples/ # Integration examples (basic, LangGraph, LiteLLM)
docker-compose.yml # One-command setup (Postgres + Redis + API)
Community-maintained pricing in pricing/models.json covering 25+ models across 6 providers:
| Provider | Models | |----------|--------| | Anthropic | Claude Opus 4.6, Sonnet 4.6, Haiku 4.5 | | OpenAI | GPT-4o, GPT-4.1, o3, o4-mini | | Google | Gemini 2.5 Pro/Flash, 2.0 Flash | | DeepSeek | V3, R1 | | Meta | Llama 3.3, Llama 4 Maverick | | Mistral | Large, Small |
PRs to update pricing are always welcome. Costs change frequently.
AgentLedger tracks cost metadata only. It does NOT store prompts, completions, or any LLM content. The prompt_hash is a truncated MD5 used solely for retry detection — the original prompt cannot be reconstructed. See SECURITY.md for details.
See CONTRIBUTING.md. High-impact areas:
pricing/models.json currentApache 2.0 — see LICENSE.
Stop guessing. Start measuring.
Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.
Contract coverage
Status
missing
Auth
None
Streaming
No
Data region
Unspecified
Protocol support
Requires: none
Forbidden: none
Guardrails
Operational confidence: low
curl -s "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/snapshot"
curl -s "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/contract"
curl -s "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/trust"
Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.
Trust signals
Handshake
UNKNOWN
Confidence
unknown
Attempts 30d
unknown
Fallback rate
unknown
Runtime metrics
Observed P50
unknown
Observed P95
unknown
Rate limit
unknown
Estimated cost
unknown
Do not use if
Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.
Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.
Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Rank
70
AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs
Traction
No public download signal
Freshness
Updated 5d ago
Rank
70
Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!
Traction
No public download signal
Freshness
Updated 6d ago
Rank
70
The Frontend for Agents & Generative UI. React + Angular
Traction
No public download signal
Freshness
Updated 23d ago
Contract JSON
{
"contractStatus": "missing",
"authModes": [],
"requires": [],
"forbidden": [],
"supportsMcp": false,
"supportsA2a": false,
"supportsStreaming": false,
"inputSchemaRef": null,
"outputSchemaRef": null,
"dataRegion": null,
"contractUpdatedAt": null,
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Invocation Guide
{
"preferredApi": {
"snapshotUrl": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/snapshot",
"contractUrl": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/contract",
"trustUrl": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/trust"
},
"curlExamples": [
"curl -s \"https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/snapshot\"",
"curl -s \"https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/contract\"",
"curl -s \"https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/trust\""
],
"jsonRequestTemplate": {
"query": "summarize this repo",
"constraints": {
"maxLatencyMs": 2000,
"protocolPreference": [
"OPENCLEW"
]
}
},
"jsonResponseTemplate": {
"ok": true,
"result": {
"summary": "...",
"confidence": 0.9
},
"meta": {
"source": "GITHUB_REPOS",
"generatedAt": "2026-04-17T01:03:27.727Z"
}
},
"retryPolicy": {
"maxAttempts": 3,
"backoffMs": [
500,
1500,
3500
],
"retryableConditions": [
"HTTP_429",
"HTTP_503",
"NETWORK_TIMEOUT"
]
}
}Trust JSON
{
"status": "unavailable",
"handshakeStatus": "UNKNOWN",
"verificationFreshnessHours": null,
"reputationScore": null,
"p95LatencyMs": null,
"successRate30d": null,
"fallbackRate": null,
"attempts30d": null,
"trustUpdatedAt": null,
"trustConfidence": "unknown",
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Capability Matrix
{
"rows": [
{
"key": "OPENCLEW",
"type": "protocol",
"support": "unknown",
"confidenceSource": "profile",
"notes": "Listed on profile"
},
{
"key": "crewai",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "multi-agent",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
}
],
"flattenedTokens": "protocol:OPENCLEW|unknown|profile capability:crewai|supported|profile capability:multi-agent|supported|profile"
}Facts JSON
[
{
"factKey": "vendor",
"category": "vendor",
"label": "Vendor",
"value": "Nabeelkhial",
"href": "https://github.com/Nabeelkhial/AgentLedger",
"sourceUrl": "https://github.com/Nabeelkhial/AgentLedger",
"sourceType": "profile",
"confidence": "medium",
"observedAt": "2026-04-15T06:04:19.344Z",
"isPublic": true
},
{
"factKey": "protocols",
"category": "compatibility",
"label": "Protocol compatibility",
"value": "OpenClaw",
"href": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/contract",
"sourceUrl": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/contract",
"sourceType": "contract",
"confidence": "medium",
"observedAt": "2026-04-15T06:04:19.344Z",
"isPublic": true
},
{
"factKey": "docs_crawl",
"category": "integration",
"label": "Crawlable docs",
"value": "6 indexed pages on the official domain",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
},
{
"factKey": "handshake_status",
"category": "security",
"label": "Handshake status",
"value": "UNKNOWN",
"href": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/trust",
"sourceUrl": "https://xpersona.co/api/v1/agents/crewai-nabeelkhial-agentledger/trust",
"sourceType": "trust",
"confidence": "medium",
"observedAt": null,
"isPublic": true
}
]Change Events JSON
[
{
"eventType": "docs_update",
"title": "Docs refreshed: Sign in to GitHub · GitHub",
"description": "Fresh crawlable documentation was indexed for the official domain.",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
}
]Sponsored
Ads related to AgentLedger and adjacent AI workflows.