Crawler Summary

crewai-eda-pipeline answer-first brief

Local-first evaluation and analysis pipeline designed to surface regressions, failure modes, and explainable signals in AI-driven workflows. Explainable Multi-Agent System for Automated Data Analysis **EMAS-ADA** transforms raw datasets into comprehensive, insightful reports with explainable visualizations using a team of specialized AI agents. Why this project exists Modern AI systems fail silently. Small changes in prompts, models, preprocessing, or data distributions can cause: - subtle quality regressions - inconsistent outputs across runs - explanati Capability contract not published. No trust telemetry is available yet. Last updated 4/16/2026.

Freshness

Last checked 4/16/2026

Best For

crewai-eda-pipeline is best for crewai, multi-agent workflows where OpenClaw compatibility matters.

Not Ideal For

Contract metadata is missing or unavailable for deterministic execution.

Evidence Sources Checked

editorial-content, GITHUB OPENCLEW, runtime-metrics, public facts pack

Claim this agent
Agent DossierGitHubSafety: 66/100

crewai-eda-pipeline

Local-first evaluation and analysis pipeline designed to surface regressions, failure modes, and explainable signals in AI-driven workflows. Explainable Multi-Agent System for Automated Data Analysis **EMAS-ADA** transforms raw datasets into comprehensive, insightful reports with explainable visualizations using a team of specialized AI agents. Why this project exists Modern AI systems fail silently. Small changes in prompts, models, preprocessing, or data distributions can cause: - subtle quality regressions - inconsistent outputs across runs - explanati

OpenClawself-declared

Public facts

4

Change events

1

Artifacts

0

Freshness

Apr 16, 2026

Verifiededitorial-contentNo verified compatibility signals

Capability contract not published. No trust telemetry is available yet. Last updated 4/16/2026.

Trust evidence available

Trust score

Unknown

Compatibility

OpenClaw

Freshness

Apr 16, 2026

Vendor

Siddharth Narigra

Artifacts

0

Benchmarks

0

Last release

Unpublished

Executive Summary

Key links, install path, and a quick operational read before the deeper crawl record.

Verifiededitorial-content

Summary

Capability contract not published. No trust telemetry is available yet. Last updated 4/16/2026.

Setup snapshot

git clone https://github.com/siddharth-narigra/crewai-eda-pipeline.git
  1. 1

    Setup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.

  2. 2

    Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.

Evidence Ledger

Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.

Verifiededitorial-content
Vendor (1)

Vendor

Siddharth Narigra

profilemedium
Observed Apr 16, 2026Source linkProvenance
Compatibility (1)

Protocol compatibility

OpenClaw

contractmedium
Observed Apr 16, 2026Source linkProvenance
Security (1)

Handshake status

UNKNOWN

trustmedium
Observed unknownSource linkProvenance
Integration (1)

Crawlable docs

6 indexed pages on the official domain

search_documentmedium
Observed Apr 15, 2026Source linkProvenance

Release & Crawl Timeline

Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.

Self-declaredagent-index

Artifacts Archive

Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.

Self-declaredGITHUB OPENCLEW

Extracted files

0

Examples

6

Snippets

0

Languages

python

Executable Examples

mermaid

flowchart LR
    User[User] -->|Upload Dataset| Frontend[Next.js Frontend]
    Frontend -->|REST API| Backend[FastAPI Backend]
    Backend -->|Orchestrate| Crew[CrewAI EDACrew]
  
    subgraph Agents[Specialized Agents]
        direction TB
        A1[Profiler] --> A2[Cleaner]
        A2 --> A3[Statistician]
        A3 --> A4[Visualizer]
        A4 --> A5[XAI Agent]
        A5 --> A6[Reporter]
    end
  
    Crew --> Agents
    Agents -->|Generate| Output[Reports and Charts]
    Output -->|Display| Frontend

mermaid

flowchart TB
    subgraph UserInterface["User Interface Layer"]
        User["User"]
        subgraph NextJS["Next.js Frontend"]
            FileUploader["FileUploader"]
            ProgressBar["ProgressBar"]
            Dashboard["Dashboard Page"]
            ReportViewer["ReportViewer"]
            ChartGallery["ChartGallery"]
            ModelViewer["ModelViewer"]
            BeforeAfter["BeforeAfterPanel"]
        end
    end

    subgraph APILayer["API Layer"]
        subgraph FastAPI["FastAPI Backend"]
            Upload["/api/upload"]
            RunEDA["/api/eda/run"]
            Status["/api/eda/status"]
            Report["/api/report"]
            Charts["/api/charts"]
            Model["/api/model"]
            SHAP["/api/shap"]
        end
        ProgressTracker["ProgressTracker"]
    end

    subgraph OrchestrationLayer["Orchestration Layer"]
        EDACrew["EDACrew Orchestrator"]
        LLM["LLM Provider via OpenRouter"]
        TaskQueue["Task Sequencing"]
        Callbacks["Step and Task Callbacks"]
    end

    subgraph AgentLayer["Agent Layer"]
        direction LR
        Profiler["Data Profiler Agent"]
        Cleaner["Data Cleaner Agent"]
        Statistician["Statistician Agent"]
        Visualizer["Data Visualizer Agent"]
        ModelRec["Model Recommender Agent"]
        XAIAgent["XAI Agent"]
        Reporter["Technical Report Writer"]
    end

    subgraph ToolsLayer["Tools Layer"]
        subgraph DataTools["Data Tools"]
            ProfileDataset["ProfileDatasetTool"]
            DetectOutliers["DetectOutliersTool"]
            CleanMissing["CleanMissingValuesTool"]
            GetColumnInfo["GetColumnInfoTool"]
            GetDataSummary["GetDataSummaryTool"]
        end
        subgraph StatsTools["Statistics Tools"]
            DescriptiveStats["DescriptiveStatsTool"]
            CorrelationAnalysis["CorrelationAnalysisTool"]
            CategoricalAnalysis["CategoricalAnalysisTool"]
            DetectPatterns["DetectPatternsTool"]

bash

git clone https://github.com/siddharth-narigra/crewai-eda-pipeline.git
   cd crewai-eda-pipeline

bash

python -m venv venv
   .\venv\Scripts\activate       # Windows
   # source venv/bin/activate    # Linux/Mac
   pip install -r requirements.txt

bash

cd frontend
   npm install

bash

# From root directory
   uvicorn src.api.main:app --reload

Docs & README

Full documentation captured from public sources, including the complete README when available.

Self-declaredGITHUB OPENCLEW

Docs source

GITHUB OPENCLEW

Editorial quality

ready

Local-first evaluation and analysis pipeline designed to surface regressions, failure modes, and explainable signals in AI-driven workflows. Explainable Multi-Agent System for Automated Data Analysis **EMAS-ADA** transforms raw datasets into comprehensive, insightful reports with explainable visualizations using a team of specialized AI agents. Why this project exists Modern AI systems fail silently. Small changes in prompts, models, preprocessing, or data distributions can cause: - subtle quality regressions - inconsistent outputs across runs - explanati

Full README

Explainable Multi-Agent System for Automated Data Analysis

Python 3.10+ License: MIT Next.js FastAPI CrewAI

EMAS-ADA transforms raw datasets into comprehensive, insightful reports with explainable visualizations using a team of specialized AI agents.

Why this project exists

Modern AI systems fail silently.

Small changes in prompts, models, preprocessing, or data distributions can cause:

  • subtle quality regressions
  • inconsistent outputs across runs
  • explanations that no longer match behavior
  • decisions that are difficult to audit or trust

This project was built to address that problem.

EMAS-ADA is not a dashboard-first analytics tool — it is an evaluation and decision-support pipeline. Its primary goal is to turn raw, messy inputs into reproducible signals that help teams answer:

  • What changed?
  • Why did it change?
  • Can we trust this output?
  • Would we catch this regression again if it happened tomorrow?

To do this, the system enforces:

  • Deterministic, ordered execution (no ad-hoc notebooks)
  • Explicit failure surfaces at each stage of the pipeline
  • Auditable transformations with before/after comparisons
  • Explainability as a first-class output, not an afterthought

While the implementation uses an EDA workflow as the substrate, the design principles map directly to AI evaluation infrastructure — where correctness, traceability, and regression detection matter more than speed or visual polish.

Architecture Diagram

flowchart LR
    User[User] -->|Upload Dataset| Frontend[Next.js Frontend]
    Frontend -->|REST API| Backend[FastAPI Backend]
    Backend -->|Orchestrate| Crew[CrewAI EDACrew]
  
    subgraph Agents[Specialized Agents]
        direction TB
        A1[Profiler] --> A2[Cleaner]
        A2 --> A3[Statistician]
        A3 --> A4[Visualizer]
        A4 --> A5[XAI Agent]
        A5 --> A6[Reporter]
    end
  
    Crew --> Agents
    Agents -->|Generate| Output[Reports and Charts]
    Output -->|Display| Frontend

<sub>Detailed Architecture Diagram (for nerds)</sub>

flowchart TB
    subgraph UserInterface["User Interface Layer"]
        User["User"]
        subgraph NextJS["Next.js Frontend"]
            FileUploader["FileUploader"]
            ProgressBar["ProgressBar"]
            Dashboard["Dashboard Page"]
            ReportViewer["ReportViewer"]
            ChartGallery["ChartGallery"]
            ModelViewer["ModelViewer"]
            BeforeAfter["BeforeAfterPanel"]
        end
    end

    subgraph APILayer["API Layer"]
        subgraph FastAPI["FastAPI Backend"]
            Upload["/api/upload"]
            RunEDA["/api/eda/run"]
            Status["/api/eda/status"]
            Report["/api/report"]
            Charts["/api/charts"]
            Model["/api/model"]
            SHAP["/api/shap"]
        end
        ProgressTracker["ProgressTracker"]
    end

    subgraph OrchestrationLayer["Orchestration Layer"]
        EDACrew["EDACrew Orchestrator"]
        LLM["LLM Provider via OpenRouter"]
        TaskQueue["Task Sequencing"]
        Callbacks["Step and Task Callbacks"]
    end

    subgraph AgentLayer["Agent Layer"]
        direction LR
        Profiler["Data Profiler Agent"]
        Cleaner["Data Cleaner Agent"]
        Statistician["Statistician Agent"]
        Visualizer["Data Visualizer Agent"]
        ModelRec["Model Recommender Agent"]
        XAIAgent["XAI Agent"]
        Reporter["Technical Report Writer"]
    end

    subgraph ToolsLayer["Tools Layer"]
        subgraph DataTools["Data Tools"]
            ProfileDataset["ProfileDatasetTool"]
            DetectOutliers["DetectOutliersTool"]
            CleanMissing["CleanMissingValuesTool"]
            GetColumnInfo["GetColumnInfoTool"]
            GetDataSummary["GetDataSummaryTool"]
        end
        subgraph StatsTools["Statistics Tools"]
            DescriptiveStats["DescriptiveStatsTool"]
            CorrelationAnalysis["CorrelationAnalysisTool"]
            CategoricalAnalysis["CategoricalAnalysisTool"]
            DetectPatterns["DetectPatternsTool"]
            NormalityTest["NormalityTestTool"]
        end
        subgraph VizTools["Visualization Tools"]
            DistributionPlots["DistributionPlotsTool"]
            CorrelationHeatmap["CorrelationHeatmapTool"]
            CategoricalCharts["CategoricalChartsTool"]
            BoxPlots["BoxPlotsTool"]
            CleaningImpact["CleaningImpactPlotTool"]
            DataQualitySummary["DataQualitySummaryTool"]
        end
        subgraph MLTools["ML Tools"]
            SuggestModels["SuggestModelsBasedOnDataTool"]
            TrainModel["TrainSimpleModelTool"]
        end
        subgraph XAITools["XAI Tools"]
            SHAPSummary["GenerateSHAPSummaryTool"]
            LIMEExplanation["GenerateLIMEExplanationTool"]
            FeatureImportance["CompareFeatureImportanceTool"]
        end
    end

    subgraph DataLayer["Data Layer"]
        DataStore["DataStore Singleton"]
        OriginalDF["Original DataFrame"]
        CleanedDF["Cleaned DataFrame"]
        Changelog["Changelog"]
        Metadata["Metadata"]
    end

    subgraph OutputLayer["Output Layer"]
        OutputDir["output/"]
        ChartsDir["output/charts/"]
        ModelsDir["output/models/"]
        ReportMD["report.md + Model Summary"]
        ReportHTML["report.html + Model Summary"]
        CleanedCSV["cleaned_data.csv"]
        PNGCharts["*.png Charts"]
        ModelPKL["trained_model.pkl"]
    end

    User -->|"Upload CSV/Excel"| FileUploader
    FileUploader -->|"POST /api/upload"| Upload
    Dashboard -->|"POST /api/eda/run"| RunEDA
    ProgressBar -->|"GET /api/eda/status"| Status
    ReportViewer -->|"GET /api/report"| Report
    ChartGallery -->|"GET /api/charts"| Charts
    ModelViewer -->|"GET /api/model"| Model

    Upload -->|"Store File"| DataStore
    RunEDA -->|"Trigger Background Task"| EDACrew
    Status -->|"Query Status"| ProgressTracker

    EDACrew -->|"Create LLM"| LLM
    EDACrew -->|"Sequence Tasks"| TaskQueue
    EDACrew -->|"Track Progress"| Callbacks
    Callbacks -->|"Update"| ProgressTracker

    TaskQueue -->|"1. Profile"| Profiler
    TaskQueue -->|"2. Clean"| Cleaner
    TaskQueue -->|"3. Analyze"| Statistician
    TaskQueue -->|"4. Visualize"| Visualizer
    TaskQueue -->|"5. Recommend"| ModelRec
    TaskQueue -->|"6. Explain"| XAIAgent
    TaskQueue -->|"7. Report"| Reporter

    Profiler --> DataTools
    Cleaner --> DataTools
    Statistician --> StatsTools
    Visualizer --> VizTools
    ModelRec --> MLTools
    XAIAgent --> XAITools

    DataTools --> DataStore
    StatsTools --> DataStore
    VizTools --> DataStore
    MLTools --> DataStore
    XAITools --> DataStore

    DataStore --> OriginalDF
    DataStore --> CleanedDF
    DataStore --> Changelog
    DataStore --> Metadata

    VizTools -->ChartsDir
    MLTools --> ModelsDir
    MLTools -->|"Append Model Summary"| ReportMD
    MLTools -->|"Append Model Summary"| ReportHTML
    Reporter --> ReportMD
    Reporter --> ReportHTML
    DataTools --> CleanedCSV

    ChartsDir --> PNGCharts
    ModelsDir --> ModelPKL

Architecture Explanation

  • Data Ingestion Module: src/api/main.py handles file uploads via FastAPI, validating formats (.csv, .xlsx) and storing them safely.
  • Agent Orchestrator: src/crew/eda_crew.py utilizes CrewAI to manage the lifecycle of agents, ensuring tasks are executed in the correct sequential order.
  • Specialized Agents (7 total):
    • Profiler: Audits data quality, structure, and identifies issues.
    • Cleaner: Handles missing values and outlier detection using Pandas.
    • Statistician: Performs statistical analysis (correlations, normality tests, pattern detection).
    • Visualizer: Generates charts using Matplotlib/Seaborn (distributions, heatmaps, box plots).
    • Model Recommender: Suggests and trains ML models using Scikit-learn, capturing detailed training metadata.
    • XAI Agent: Computes SHAP values and LIME explanations for model interpretability.
    • Reporter: Compiles all findings into structured Markdown/HTML reports.
  • Model Summary Integration: After training, detailed model metrics are automatically appended to the report.
  • Frontend UI: A Next.js application that provides a real-time dashboard for monitoring progress, viewing interactive reports, and exploring model training details (confusion matrix, hyperparameters, feature importance).

Key Features

  • Multi-Agent Collaboration: Specialized roles ensure depth in every aspect of analysis (cleaning, stats, viz).
  • Explainable AI (XAI): Integrated SHAP and LIME for transparent model interpretation.
  • Automated Data Cleaning: Intelligent handling of missing values and outliers with audit logs.
  • Interactive Dashboard: Modern UI for easy upload, monitoring, and report consumption.
  • Dual-Format Reporting: Generates both Markdown (for devs) and HTML (for business) reports.
  • Enhanced Model Viewer: Displays training configuration, data split, performance metrics, confusion matrix with interpretation, and hyperparameters.
  • Unified Model Report: Model training summary is automatically appended to the EDA report after training.

Screenshots

<div align="center"> <table> <tr> <th><sub>Upload</sub></th> <th><sub>Charts</sub></th> <th><sub>Report</sub></th> </tr> <tr> <td><img src="screenshots/upload.png" width="350"/></td> <td><img src="screenshots/charts.png" width="350"/></td> <td rowspan="3" align="center"><img src="screenshots/report.png" width="350"/></td> </tr> <tr> <th><sub>Dashboard</sub></th> <th><sub>Model</sub></th> </tr> <tr> <td><img src="screenshots/dashboard.png" width="350"/></td> <td><img src="screenshots/model.png" width="350"/></td> </tr> </table> </div>

Tech Stack (With Purpose)

  • Python 3.10+: The core runtime for backend logic and data science libraries.
  • CrewAI: Orchestrates the multi-agent workflow and task delegation.
  • FastAPI: Provides a high-performance, async REST API for the frontend.
  • Next.js 14: Powers the responsive, server-side rendered user interface.
  • Pandas & NumPy: High-performance data manipulation and analysis.
  • Scikit-learn: Machine learning utilities for predictive modeling and clustering.
  • SHAP & LIME: Explainable AI libraries for model transparency.

System Workflow (Chronological Processing Steps)

  1. Upload: User uploads a dataset via the Web UI.
  2. Profiling: The Profiler agent scans the file for schema, types, and quality issues.
  3. Cleaning: The Cleaner agent executes transformations to fix identified issues.
  4. Statistical Analysis: The Statistician agent computes descriptive stats, correlations, and pattern detection.
  5. Visualization: The Visualizer agent generates distribution plots, heatmaps, and box plots.
  6. Model Training: The Model Recommender agent suggests and trains an appropriate ML model, capturing detailed metrics (accuracy, precision, recall, F1, confusion matrix).
  7. Explanation: The XAI Agent computes SHAP/LIME explanations for model interpretability.
  8. Reporting: The Reporter agent compiles all artifacts into a cohesive narrative.
  9. Model Summary: Training results are automatically appended to the report (training config, data split, performance metrics, hyperparameters).
  10. Review: User views the final report, charts, and model details on the Dashboard.

Input Format

  • Data: .csv, .xlsx, or .xls.
  • Requirements: Tabular data with headers.

Output Format

  • Interactive Report: Displayed directly in the UI with live model metrics.
  • Files:
    • report.md: Comprehensive Markdown report (includes Model Training Summary section).
    • report.html: HTML-formatted report for business stakeholders.
    • cleaned_data.csv: The processed dataset.
    • charts/*.png: Generated visualizations (distributions, correlations, SHAP, LIME).
    • models/trained_model.pkl: Trained ML model with encoders and metadata.

Installation Instructions

  1. Clone the Repository:

    git clone https://github.com/siddharth-narigra/crewai-eda-pipeline.git
    cd crewai-eda-pipeline
    
  2. Backend Setup:

    python -m venv venv
    .\venv\Scripts\activate       # Windows
    # source venv/bin/activate    # Linux/Mac
    pip install -r requirements.txt
    
  3. Frontend Setup:

    cd frontend
    npm install
    
  4. Configuration:

    • Copy .env.example to .env.
    • Add your API Key (e.g., OPENROUTER_API_KEY=sk-...).

How to Run

  1. Start Backend:
    # From root directory
    uvicorn src.api.main:app --reload
    
  2. Start Frontend:
    # From frontend directory
    npm run dev
    
  3. Access App: Open http://localhost:3000.

Folder Structure

crewai-eda-pipeline/
├── src/
│   ├── agents/              # Agent definitions (Profiler, Cleaner, Statistician, etc.)
│   ├── api/                 # FastAPI endpoints and progress tracking
│   ├── crew/                # CrewAI orchestration (eda_crew.py)
│   ├── tools/               # Custom tools (data, stats, viz, ml, xai)
│   └── utils/               # Utility functions (file handling)
├── frontend/                # Next.js Web Application
│   └── src/components/      # React components (FileUploader, ReportViewer, etc.)
├── output/                  # Generated artifacts (Reports, Charts, Models)
├── uploads/                 # Temporary storage for uploaded files
├── sample_data/             # Sample datasets for testing
├── requirements.txt         # Backend Python dependencies
└── README.md                # This document

Limitations

  • LLM Dependency: Requires an active internet connection and API key for the LLM provider.
  • Processing Time: Large datasets may take longer to process due to LLM latency.
  • Token Costs: Extensive analysis of very large files may consume significant API tokens.

License

MIT License

Contract & API

Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.

MissingGITHUB OPENCLEW

Contract coverage

Status

missing

Auth

None

Streaming

No

Data region

Unspecified

Protocol support

OpenClaw: self-declared

Requires: none

Forbidden: none

Guardrails

Operational confidence: low

No positive guardrails captured.
Invocation examples
curl -s "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/snapshot"
curl -s "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/contract"
curl -s "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/trust"

Reliability & Benchmarks

Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.

Missingruntime-metrics

Trust signals

Handshake

UNKNOWN

Confidence

unknown

Attempts 30d

unknown

Fallback rate

unknown

Runtime metrics

Observed P50

unknown

Observed P95

unknown

Rate limit

unknown

Estimated cost

unknown

Do not use if

Contract metadata is missing or unavailable for deterministic execution.
No benchmark suites or observed failure patterns are available.

Media & Demo

Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.

Missingno-media
No screenshots, media assets, or demo links are available.

Related Agents

Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.

Self-declaredprotocol-neighbors
GITHUB_REPOSactivepieces

Rank

70

AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents

Traction

No public download signal

Freshness

Updated 2d ago

OPENCLAW
GITHUB_REPOScherry-studio

Rank

70

AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs

Traction

No public download signal

Freshness

Updated 6d ago

MCPOPENCLAW
GITHUB_REPOSAionUi

Rank

70

Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!

Traction

No public download signal

Freshness

Updated 6d ago

MCPOPENCLAW
GITHUB_REPOSCopilotKit

Rank

70

The Frontend for Agents & Generative UI. React + Angular

Traction

No public download signal

Freshness

Updated 23d ago

OPENCLAW
Machine Appendix

Contract JSON

{
  "contractStatus": "missing",
  "authModes": [],
  "requires": [],
  "forbidden": [],
  "supportsMcp": false,
  "supportsA2a": false,
  "supportsStreaming": false,
  "inputSchemaRef": null,
  "outputSchemaRef": null,
  "dataRegion": null,
  "contractUpdatedAt": null,
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Invocation Guide

{
  "preferredApi": {
    "snapshotUrl": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/snapshot",
    "contractUrl": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/contract",
    "trustUrl": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/trust"
  },
  "curlExamples": [
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/snapshot\"",
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/contract\"",
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/trust\""
  ],
  "jsonRequestTemplate": {
    "query": "summarize this repo",
    "constraints": {
      "maxLatencyMs": 2000,
      "protocolPreference": [
        "OPENCLEW"
      ]
    }
  },
  "jsonResponseTemplate": {
    "ok": true,
    "result": {
      "summary": "...",
      "confidence": 0.9
    },
    "meta": {
      "source": "GITHUB_OPENCLEW",
      "generatedAt": "2026-04-17T02:54:57.120Z"
    }
  },
  "retryPolicy": {
    "maxAttempts": 3,
    "backoffMs": [
      500,
      1500,
      3500
    ],
    "retryableConditions": [
      "HTTP_429",
      "HTTP_503",
      "NETWORK_TIMEOUT"
    ]
  }
}

Trust JSON

{
  "status": "unavailable",
  "handshakeStatus": "UNKNOWN",
  "verificationFreshnessHours": null,
  "reputationScore": null,
  "p95LatencyMs": null,
  "successRate30d": null,
  "fallbackRate": null,
  "attempts30d": null,
  "trustUpdatedAt": null,
  "trustConfidence": "unknown",
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Capability Matrix

{
  "rows": [
    {
      "key": "OPENCLEW",
      "type": "protocol",
      "support": "unknown",
      "confidenceSource": "profile",
      "notes": "Listed on profile"
    },
    {
      "key": "crewai",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    },
    {
      "key": "multi-agent",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    }
  ],
  "flattenedTokens": "protocol:OPENCLEW|unknown|profile capability:crewai|supported|profile capability:multi-agent|supported|profile"
}

Facts JSON

[
  {
    "factKey": "vendor",
    "label": "Vendor",
    "value": "Siddharth Narigra",
    "category": "vendor",
    "href": "https://github.com/siddharth-narigra/crewai-eda-pipeline",
    "sourceUrl": "https://github.com/siddharth-narigra/crewai-eda-pipeline",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-04-16T06:46:43.433Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "protocols",
    "label": "Protocol compatibility",
    "value": "OpenClaw",
    "category": "compatibility",
    "href": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/contract",
    "sourceUrl": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/contract",
    "sourceType": "contract",
    "confidence": "medium",
    "observedAt": "2026-04-16T06:46:43.433Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "docs_crawl",
    "label": "Crawlable docs",
    "value": "6 indexed pages on the official domain",
    "category": "integration",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "handshake_status",
    "label": "Handshake status",
    "value": "UNKNOWN",
    "category": "security",
    "href": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/trust",
    "sourceUrl": "https://xpersona.co/api/v1/agents/crewai-siddharth-narigra-crewai-eda-pipeline/trust",
    "sourceType": "trust",
    "confidence": "medium",
    "observedAt": null,
    "isPublic": true,
    "metadata": {}
  }
]

Change Events JSON

[
  {
    "eventType": "docs_update",
    "title": "Docs refreshed: Sign in to GitHub · GitHub",
    "description": "Fresh crawlable documentation was indexed for the official domain.",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true,
    "metadata": {}
  }
]

Sponsored

Ads related to crewai-eda-pipeline and adjacent AI workflows.