Crawler Summary

Medical_Assistant answer-first brief

AI Clinical Documentation Assistant - Extract structured clinical data from any document format using local LLM, CrewAI multi-agent workflow, AWS Bedrock, FAISS vector search, PII masking & Databricks sync πŸ₯ AI Clinical Documentation Assistant Healthcare-grade AI system for extracting structured clinical data from any document format using multi-agent workflows, local LLM extraction, and enterprise data governance. $1 $1 $1 $1 $1 --- ✨ Features - **πŸ€– 5-Agent CrewAI Workflow** β€” Intelligent case analysis with specialized agents for retrieval, extraction, validation, explanation, and routing - **πŸ“„ Universal File Parse Capability contract not published. No trust telemetry is available yet. 1 GitHub stars reported by the source. Last updated 4/16/2026.

Freshness

Last checked 4/16/2026

Best For

Medical_Assistant is best for crewai, multi-agent workflows where OpenClaw compatibility matters.

Not Ideal For

Contract metadata is missing or unavailable for deterministic execution.

Evidence Sources Checked

editorial-content, GITHUB OPENCLEW, runtime-metrics, public facts pack

Claim this agent
Agent DossierGitHubSafety: 66/100

Medical_Assistant

AI Clinical Documentation Assistant - Extract structured clinical data from any document format using local LLM, CrewAI multi-agent workflow, AWS Bedrock, FAISS vector search, PII masking & Databricks sync πŸ₯ AI Clinical Documentation Assistant Healthcare-grade AI system for extracting structured clinical data from any document format using multi-agent workflows, local LLM extraction, and enterprise data governance. $1 $1 $1 $1 $1 --- ✨ Features - **πŸ€– 5-Agent CrewAI Workflow** β€” Intelligent case analysis with specialized agents for retrieval, extraction, validation, explanation, and routing - **πŸ“„ Universal File Parse

OpenClawself-declared

Public facts

5

Change events

1

Artifacts

0

Freshness

Apr 16, 2026

Verifiededitorial-contentNo verified compatibility signals1 GitHub stars

Capability contract not published. No trust telemetry is available yet. 1 GitHub stars reported by the source. Last updated 4/16/2026.

1 GitHub starsTrust evidence available

Trust score

Unknown

Compatibility

OpenClaw

Freshness

Apr 16, 2026

Vendor

Kazinymul

Artifacts

0

Benchmarks

0

Last release

Unpublished

Executive Summary

Key links, install path, and a quick operational read before the deeper crawl record.

Verifiededitorial-content

Summary

Capability contract not published. No trust telemetry is available yet. 1 GitHub stars reported by the source. Last updated 4/16/2026.

Setup snapshot

git clone https://github.com/kaziNymul/Medical_Assistant.git
  1. 1

    Setup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.

  2. 2

    Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.

Evidence Ledger

Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.

Verifiededitorial-content
Vendor (1)

Vendor

Kazinymul

profilemedium
Observed Apr 16, 2026Source linkProvenance
Compatibility (1)

Protocol compatibility

OpenClaw

contractmedium
Observed Apr 16, 2026Source linkProvenance
Adoption (1)

Adoption signal

1 GitHub stars

profilemedium
Observed Apr 16, 2026Source linkProvenance
Security (1)

Handshake status

UNKNOWN

trustmedium
Observed unknownSource linkProvenance
Integration (1)

Crawlable docs

6 indexed pages on the official domain

search_documentmedium
Observed Apr 15, 2026Source linkProvenance

Release & Crawl Timeline

Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.

Self-declaredagent-index

Artifacts Archive

Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.

Self-declaredGITHUB OPENCLEW

Extracted files

0

Examples

6

Snippets

0

Languages

python

Executable Examples

mermaid

flowchart TB
    subgraph Frontend["πŸ–₯️ Frontend (React + Vite)"]
        Dashboard[πŸ“Š Dashboard]
        Query[πŸ” Query]
        CaseBrief[πŸ“‹ Case Brief]
        SmartExtract[🧠 Smart Extract]
        Documents[πŸ“ Documents]
    end

    subgraph API["⚑ FastAPI Backend"]
        QueryAPI["/api/query"]
        CrewAPI["/api/crew"]
        ExtractAPI["/api/extract"]
        DocsAPI["/api/documents"]
    end

    subgraph Core["πŸ”§ Core Services"]
        RAG["RAG Pipeline"]
        CrewAI["CrewAI Agents"]
        Extractor["LLM Extractor"]
        Masking["PII Masking"]
    end

    subgraph Storage["πŸ’Ύ Data Layer"]
        FAISS[(FAISS\n55,500 docs)]
        SQLite[(SQLite\nPII Store)]
        Vault[(HashiCorp\nVault)]
    end

    subgraph Cloud["☁️ Cloud Services"]
        Bedrock["AWS Bedrock\nClaude 3 Haiku"]
        Titan["Titan Embeddings\nV2 1024-dim"]
        Databricks["Databricks\nUnity Catalog"]
    end

    Frontend --> API
    API --> Core
    Core --> Storage
    Core --> Cloud

    style Frontend fill:#e1f5fe
    style API fill:#fff3e0
    style Core fill:#f3e5f5
    style Storage fill:#e8f5e9
    style Cloud fill:#fce4ec

mermaid

flowchart LR
    subgraph Input["πŸ“₯ Input Sources"]
        PDF[PDF]
        Word[Word]
        HL7[HL7/FHIR]
        CSV[CSV/Excel]
        Images[Images]
    end

    subgraph Parser["πŸ“„ Universal Parser"]
        FileParser[File Parser]
        OCR[Tesseract OCR]
    end

    subgraph Processing["βš™οΈ Processing"]
        Masking[PII Masking]
        Chunking[Text Chunking]
        Embedding[Titan Embeddings]
    end

    subgraph Storage["πŸ’Ύ Storage"]
        FAISS[(FAISS Index)]
        PII[(PII Database)]
        Vault[(Vault Secrets)]
    end

    subgraph AI["πŸ€– AI Layer"]
        LocalLLM[FLAN-T5 Base]
        Bedrock[Claude 3 Haiku]
        CrewAI[CrewAI Agents]
    end

    subgraph Output["πŸ“€ Output"]
        JSON[Structured JSON]
        Databricks[(Databricks)]
        UI[React UI]
    end

    Input --> Parser
    Parser --> Processing
    Processing --> Storage
    Storage --> AI
    AI --> Output

    style Input fill:#ffebee
    style Parser fill:#fff8e1
    style Processing fill:#e8f5e9
    style Storage fill:#e3f2fd
    style AI fill:#f3e5f5
    style Output fill:#e0f2f1

mermaid

flowchart TD
    Start([πŸš€ Clinical Query]) --> Retrieval

    subgraph Agents["CrewAI Agent Workflow"]
        Retrieval["πŸ” Retrieval Agent\n─────────────────\nSearches FAISS index\nFinds relevant documents"]
        Extraction["πŸ“‹ Extraction Agent\n─────────────────\nExtracts structured data\nDiagnoses, medications, labs"]
        Validation["βœ… Validation Agent\n─────────────────\nChecks completeness\nValidates consistency"]
        Explanation["πŸ’¬ Explanation Agent\n─────────────────\nGenerates summaries\nHuman-readable output"]
        Routing["πŸ”€ Routing Agent\n─────────────────\nRoutes to workflows\nPrioritizes cases"]
    end

    Retrieval --> Extraction
    Extraction --> Validation
    Validation --> Explanation
    Explanation --> Routing
    Routing --> End([πŸ“Š Structured Output])

    style Retrieval fill:#bbdefb
    style Extraction fill:#c8e6c9
    style Validation fill:#fff9c4
    style Explanation fill:#f8bbd9
    style Routing fill:#d1c4e9

mermaid

flowchart TB
    subgraph Input["Raw Data Input"]
        RawData[("πŸ“„ Clinical Documents\nwith PHI/PII")]
    end

    subgraph Masking["πŸ”’ PII Detection & Masking"]
        Detect["Detect PHI Entities"]
        Mask["Apply Masking Rules"]
        Store["Store Mapping in Vault"]
    end

    subgraph MaskedEntities["Masked Entities"]
        Name["[PATIENT_NAME]"]
        SSN["[SSN]"]
        DOB["[DOB]"]
        Phone["[PHONE]"]
        Email["[EMAIL]"]
        MRN["[MRN]"]
    end

    subgraph Secure["πŸ” Secure Storage"]
        Vault[("HashiCorp Vault\nSecrets & Mappings")]
        PIIDb[("SQLite\nPII Database")]
    end

    subgraph Output["Safe Output"]
        MaskedDocs["Masked Documents"]
        Databricks["Databricks\nMasked Tables"]
    end

    RawData --> Detect
    Detect --> Mask
    Mask --> MaskedEntities
    Mask --> Store
    Store --> Secure
    MaskedEntities --> Output

    style Input fill:#ffcdd2
    style Masking fill:#fff9c4
    style MaskedEntities fill:#e1f5fe
    style Secure fill:#c8e6c9
    style Output fill:#d1c4e9

mermaid

flowchart LR
    subgraph Source["πŸ“₯ Source Data"]
        Extractions[Clinical Extractions]
        Metadata[Document Metadata]
    end

    subgraph Sync["πŸ”„ Databricks Sync"]
        SyncService["databricks_sync.py"]
    end

    subgraph Unity["πŸ“Š Unity Catalog"]
        subgraph Catalog["medical_ai"]
            subgraph Schema["clinical_data"]
                Unmasked[("clinical_extractions_unmasked\n────────────────\nπŸ”΄ Contains PHI\nπŸ”’ Restricted Access")]
                Masked[("clinical_extractions_masked\n────────────────\n🟒 PHI Removed\nπŸ“Š Analytics Safe")]
            end
        end
    end

    subgraph Access["πŸ‘₯ Access Control"]
        Clinicians["πŸ‘¨β€βš•οΈ Clinicians\nFull Access"]
        Analysts["πŸ“Š Analysts\nMasked Only"]
        AI["πŸ€– AI Systems\nMasked Only"]
    end

    Source --> Sync
    Sync --> Unity
    Unmasked --> Clinicians
    Masked --> Analysts
    Masked --> AI

    style Unmasked fill:#ffcdd2
    style Masked fill:#c8e6c9

mermaid

flowchart TD
    subgraph Upload["πŸ“€ File Upload"]
        File["Any File Format"]
    end

    subgraph Parse["πŸ“„ Universal Parser"]
        Detect["Detect Format"]
        Extract["Extract Text"]
        OCR["OCR if Image"]
    end

    subgraph LLM["🧠 FLAN-T5 Extraction"]
        Prompt["Build Extraction Prompt"]
        Inference["CPU Inference\n~2-5 seconds"]
        Parse2["Parse JSON Output"]
    end

    subgraph Fields["πŸ“‹ Extracted Fields"]
        Demographics["Demographics\nName, DOB, MRN"]
        Diagnoses["Diagnoses\nICD codes, descriptions"]
        Medications["Medications\nDrugs, dosages"]
        Labs["Lab Results\nValues, dates"]
        Vitals["Vital Signs\nBP, HR, Temp"]
        Plan["Assessment & Plan"]
    end

    subgraph Output["πŸ“€ Output"]
        JSON["Structured JSON"]
        Masked["Masked Version"]
        DB["Save to Database"]
    end

    Upload --> Parse
    Parse --> LLM
    LLM --> Fields
    Fields --> Output

    style Upload fill:#e3f2fd
    style Parse fill:#fff8e1
    style LLM fill:#f3e5f5
    style Fields fill:#e8f5e9
    style Output fill:#fce4ec

Docs & README

Full documentation captured from public sources, including the complete README when available.

Self-declaredGITHUB OPENCLEW

Docs source

GITHUB OPENCLEW

Editorial quality

ready

AI Clinical Documentation Assistant - Extract structured clinical data from any document format using local LLM, CrewAI multi-agent workflow, AWS Bedrock, FAISS vector search, PII masking & Databricks sync πŸ₯ AI Clinical Documentation Assistant Healthcare-grade AI system for extracting structured clinical data from any document format using multi-agent workflows, local LLM extraction, and enterprise data governance. $1 $1 $1 $1 $1 --- ✨ Features - **πŸ€– 5-Agent CrewAI Workflow** β€” Intelligent case analysis with specialized agents for retrieval, extraction, validation, explanation, and routing - **πŸ“„ Universal File Parse

Full README

πŸ₯ AI Clinical Documentation Assistant

Healthcare-grade AI system for extracting structured clinical data from any document format using multi-agent workflows, local LLM extraction, and enterprise data governance.

Python 3.11+ FastAPI React CrewAI License: MIT


✨ Features

  • πŸ€– 5-Agent CrewAI Workflow β€” Intelligent case analysis with specialized agents for retrieval, extraction, validation, explanation, and routing
  • πŸ“„ Universal File Parser β€” Supports PDF, Word, Excel, CSV, JSON, XML, HL7, FHIR, CDA, and Images (OCR)
  • 🧠 Local LLM Extraction β€” CPU-optimized FLAN-T5 for clinical data extraction without cloud dependency
  • ☁️ AWS Bedrock Integration β€” Claude 3 Haiku for reasoning + Titan Embeddings V2 for semantic search
  • πŸ” FAISS Vector Store β€” 55,500+ indexed clinical documents with fast similarity search
  • πŸ”’ PHI/PII Masking β€” Automatic detection and masking of protected health information
  • 🏦 HashiCorp Vault β€” Secure secrets management for credentials and API keys
  • πŸ“Š Databricks Sync β€” Unity Catalog integration with masked/unmasked table separation
  • 🎨 Modern React UI β€” Dashboard, Query, Case Brief, Smart Extract, and Document Management

πŸ—οΈ System Architecture

flowchart TB
    subgraph Frontend["πŸ–₯️ Frontend (React + Vite)"]
        Dashboard[πŸ“Š Dashboard]
        Query[πŸ” Query]
        CaseBrief[πŸ“‹ Case Brief]
        SmartExtract[🧠 Smart Extract]
        Documents[πŸ“ Documents]
    end

    subgraph API["⚑ FastAPI Backend"]
        QueryAPI["/api/query"]
        CrewAPI["/api/crew"]
        ExtractAPI["/api/extract"]
        DocsAPI["/api/documents"]
    end

    subgraph Core["πŸ”§ Core Services"]
        RAG["RAG Pipeline"]
        CrewAI["CrewAI Agents"]
        Extractor["LLM Extractor"]
        Masking["PII Masking"]
    end

    subgraph Storage["πŸ’Ύ Data Layer"]
        FAISS[(FAISS\n55,500 docs)]
        SQLite[(SQLite\nPII Store)]
        Vault[(HashiCorp\nVault)]
    end

    subgraph Cloud["☁️ Cloud Services"]
        Bedrock["AWS Bedrock\nClaude 3 Haiku"]
        Titan["Titan Embeddings\nV2 1024-dim"]
        Databricks["Databricks\nUnity Catalog"]
    end

    Frontend --> API
    API --> Core
    Core --> Storage
    Core --> Cloud

    style Frontend fill:#e1f5fe
    style API fill:#fff3e0
    style Core fill:#f3e5f5
    style Storage fill:#e8f5e9
    style Cloud fill:#fce4ec

πŸ”„ Data Flow Architecture

flowchart LR
    subgraph Input["πŸ“₯ Input Sources"]
        PDF[PDF]
        Word[Word]
        HL7[HL7/FHIR]
        CSV[CSV/Excel]
        Images[Images]
    end

    subgraph Parser["πŸ“„ Universal Parser"]
        FileParser[File Parser]
        OCR[Tesseract OCR]
    end

    subgraph Processing["βš™οΈ Processing"]
        Masking[PII Masking]
        Chunking[Text Chunking]
        Embedding[Titan Embeddings]
    end

    subgraph Storage["πŸ’Ύ Storage"]
        FAISS[(FAISS Index)]
        PII[(PII Database)]
        Vault[(Vault Secrets)]
    end

    subgraph AI["πŸ€– AI Layer"]
        LocalLLM[FLAN-T5 Base]
        Bedrock[Claude 3 Haiku]
        CrewAI[CrewAI Agents]
    end

    subgraph Output["πŸ“€ Output"]
        JSON[Structured JSON]
        Databricks[(Databricks)]
        UI[React UI]
    end

    Input --> Parser
    Parser --> Processing
    Processing --> Storage
    Storage --> AI
    AI --> Output

    style Input fill:#ffebee
    style Parser fill:#fff8e1
    style Processing fill:#e8f5e9
    style Storage fill:#e3f2fd
    style AI fill:#f3e5f5
    style Output fill:#e0f2f1

πŸ€– CrewAI Multi-Agent Workflow

flowchart TD
    Start([πŸš€ Clinical Query]) --> Retrieval

    subgraph Agents["CrewAI Agent Workflow"]
        Retrieval["πŸ” Retrieval Agent\n─────────────────\nSearches FAISS index\nFinds relevant documents"]
        Extraction["πŸ“‹ Extraction Agent\n─────────────────\nExtracts structured data\nDiagnoses, medications, labs"]
        Validation["βœ… Validation Agent\n─────────────────\nChecks completeness\nValidates consistency"]
        Explanation["πŸ’¬ Explanation Agent\n─────────────────\nGenerates summaries\nHuman-readable output"]
        Routing["πŸ”€ Routing Agent\n─────────────────\nRoutes to workflows\nPrioritizes cases"]
    end

    Retrieval --> Extraction
    Extraction --> Validation
    Validation --> Explanation
    Explanation --> Routing
    Routing --> End([πŸ“Š Structured Output])

    style Retrieval fill:#bbdefb
    style Extraction fill:#c8e6c9
    style Validation fill:#fff9c4
    style Explanation fill:#f8bbd9
    style Routing fill:#d1c4e9

πŸ” Security Architecture

flowchart TB
    subgraph Input["Raw Data Input"]
        RawData[("πŸ“„ Clinical Documents\nwith PHI/PII")]
    end

    subgraph Masking["πŸ”’ PII Detection & Masking"]
        Detect["Detect PHI Entities"]
        Mask["Apply Masking Rules"]
        Store["Store Mapping in Vault"]
    end

    subgraph MaskedEntities["Masked Entities"]
        Name["[PATIENT_NAME]"]
        SSN["[SSN]"]
        DOB["[DOB]"]
        Phone["[PHONE]"]
        Email["[EMAIL]"]
        MRN["[MRN]"]
    end

    subgraph Secure["πŸ” Secure Storage"]
        Vault[("HashiCorp Vault\nSecrets & Mappings")]
        PIIDb[("SQLite\nPII Database")]
    end

    subgraph Output["Safe Output"]
        MaskedDocs["Masked Documents"]
        Databricks["Databricks\nMasked Tables"]
    end

    RawData --> Detect
    Detect --> Mask
    Mask --> MaskedEntities
    Mask --> Store
    Store --> Secure
    MaskedEntities --> Output

    style Input fill:#ffcdd2
    style Masking fill:#fff9c4
    style MaskedEntities fill:#e1f5fe
    style Secure fill:#c8e6c9
    style Output fill:#d1c4e9

πŸ“Š Databricks Integration

flowchart LR
    subgraph Source["πŸ“₯ Source Data"]
        Extractions[Clinical Extractions]
        Metadata[Document Metadata]
    end

    subgraph Sync["πŸ”„ Databricks Sync"]
        SyncService["databricks_sync.py"]
    end

    subgraph Unity["πŸ“Š Unity Catalog"]
        subgraph Catalog["medical_ai"]
            subgraph Schema["clinical_data"]
                Unmasked[("clinical_extractions_unmasked\n────────────────\nπŸ”΄ Contains PHI\nπŸ”’ Restricted Access")]
                Masked[("clinical_extractions_masked\n────────────────\n🟒 PHI Removed\nπŸ“Š Analytics Safe")]
            end
        end
    end

    subgraph Access["πŸ‘₯ Access Control"]
        Clinicians["πŸ‘¨β€βš•οΈ Clinicians\nFull Access"]
        Analysts["πŸ“Š Analysts\nMasked Only"]
        AI["πŸ€– AI Systems\nMasked Only"]
    end

    Source --> Sync
    Sync --> Unity
    Unmasked --> Clinicians
    Masked --> Analysts
    Masked --> AI

    style Unmasked fill:#ffcdd2
    style Masked fill:#c8e6c9

🧠 Local LLM Extraction Pipeline

flowchart TD
    subgraph Upload["πŸ“€ File Upload"]
        File["Any File Format"]
    end

    subgraph Parse["πŸ“„ Universal Parser"]
        Detect["Detect Format"]
        Extract["Extract Text"]
        OCR["OCR if Image"]
    end

    subgraph LLM["🧠 FLAN-T5 Extraction"]
        Prompt["Build Extraction Prompt"]
        Inference["CPU Inference\n~2-5 seconds"]
        Parse2["Parse JSON Output"]
    end

    subgraph Fields["πŸ“‹ Extracted Fields"]
        Demographics["Demographics\nName, DOB, MRN"]
        Diagnoses["Diagnoses\nICD codes, descriptions"]
        Medications["Medications\nDrugs, dosages"]
        Labs["Lab Results\nValues, dates"]
        Vitals["Vital Signs\nBP, HR, Temp"]
        Plan["Assessment & Plan"]
    end

    subgraph Output["πŸ“€ Output"]
        JSON["Structured JSON"]
        Masked["Masked Version"]
        DB["Save to Database"]
    end

    Upload --> Parse
    Parse --> LLM
    LLM --> Fields
    Fields --> Output

    style Upload fill:#e3f2fd
    style Parse fill:#fff8e1
    style LLM fill:#f3e5f5
    style Fields fill:#e8f5e9
    style Output fill:#fce4ec

πŸ–₯️ Frontend Component Architecture

flowchart TB
    subgraph App["React Application"]
        Router["React Router"]
    end

    subgraph Pages["πŸ“„ Pages"]
        Dashboard["πŸ“Š Dashboard\n────────────\nStats & Metrics\nRecent Activity"]
        Query["πŸ” Query\n────────────\nRAG Search\nBedrock LLM"]
        CaseBrief["πŸ“‹ Case Brief\n────────────\nCrewAI Analysis\nStructured Output"]
        Extract["🧠 Smart Extract\n────────────\nLocal LLM\nFile Upload"]
        Docs["πŸ“ Documents\n────────────\nDocument List\nUpload/Manage"]
        Settings["βš™οΈ Settings\n────────────\nConfiguration\nAPI Keys"]
    end

    subgraph Components["🧩 Components"]
        Layout["Layout"]
        Sidebar["Sidebar"]
        Cards["Cards"]
        Forms["Forms"]
    end

    subgraph State["πŸ“¦ State Management"]
        TanStack["TanStack Query"]
        LocalState["React State"]
    end

    subgraph API["πŸ”Œ API Layer"]
        ApiClient["api.js"]
    end

    Router --> Pages
    Pages --> Components
    Pages --> State
    State --> API

    style App fill:#e3f2fd
    style Pages fill:#fff8e1
    style Components fill:#e8f5e9
    style State fill:#f3e5f5
    style API fill:#fce4ec

πŸ”Œ Infrastructure Deployment

flowchart TB
    subgraph Docker["🐳 Docker Compose"]
        subgraph Backend["Backend Container"]
            FastAPI["FastAPI :8000"]
            Python["Python 3.11"]
        end
        
        subgraph Frontend["Frontend Container"]
            Vite["Vite Dev :5173"]
            React["React 18"]
        end
        
        subgraph Services["Service Containers"]
            VaultC["HashiCorp Vault :8200"]
        end
    end

    subgraph Volumes["πŸ“ Volumes"]
        Data["./data"]
        FAISS["./data/faiss_index"]
        Models["./models"]
    end

    subgraph External["☁️ External Services"]
        AWS["AWS Bedrock"]
        DB["Databricks"]
    end

    Docker --> Volumes
    Backend --> External
    
    style Docker fill:#e3f2fd
    style Volumes fill:#fff8e1
    style External fill:#fce4ec

πŸš€ Quick Start

Prerequisites

  • Python 3.11+
  • Node.js 18+
  • AWS credentials configured (for Bedrock)
  • HashiCorp Vault (optional, for secrets)

Installation

# Clone the repository
git clone git@github.com:kaziNymul/Medical_Assistant.git
cd Medical_Assistant

# Create virtual environment
python -m venv venv
source venv/bin/activate  # Linux/Mac
# or: venv\Scripts\activate  # Windows

# Install Python dependencies
pip install -r requirements.txt

# Install frontend dependencies
cd frontend && npm install && cd ..

# Setup environment
cp .env.example .env
# Edit .env with your credentials

Configuration

Create a .env file with:

# AWS Bedrock
AWS_REGION=us-east-1
AWS_ACCESS_KEY_ID=your_key
AWS_SECRET_ACCESS_KEY=your_secret
BEDROCK_MODEL_ID=anthropic.claude-3-haiku-20240307-v1:0
BEDROCK_EMBEDDING_MODEL=amazon.titan-embed-text-v2:0

# HashiCorp Vault
VAULT_ADDR=http://127.0.0.1:8200
VAULT_TOKEN=your_vault_token

# Databricks (optional)
DATABRICKS_HOST=your_workspace.cloud.databricks.com
DATABRICKS_TOKEN=your_token
DATABRICKS_CATALOG=medical_ai
DATABRICKS_SCHEMA=clinical_data

# Local LLM
EXTRACTION_MODEL=google/flan-t5-base
EXTRACTION_DEVICE=cpu

Running the Application

# Terminal 1: Start the backend
source venv/bin/activate
uvicorn src.api.app:app --reload --host 0.0.0.0 --port 8000

# Terminal 2: Start the frontend
cd frontend
npm run dev

Access the application at http://localhost:5173


πŸ“ Project Structure

medical_assistant/
β”œβ”€β”€ src/
β”‚   β”œβ”€β”€ api/
β”‚   β”‚   β”œβ”€β”€ app.py              # FastAPI application
β”‚   β”‚   └── routes.py           # API endpoints
β”‚   β”œβ”€β”€ crew/
β”‚   β”‚   β”œβ”€β”€ agents.py           # CrewAI agent definitions
β”‚   β”‚   β”œβ”€β”€ tasks.py            # Agent task definitions
β”‚   β”‚   β”œβ”€β”€ tools.py            # Custom agent tools
β”‚   β”‚   └── crew.py             # Crew orchestration
β”‚   β”œβ”€β”€ extraction/
β”‚   β”‚   β”œβ”€β”€ file_parser.py      # Universal file parser
β”‚   β”‚   └── llm_extractor.py    # Local FLAN-T5 extractor
β”‚   β”œβ”€β”€ rag/
β”‚   β”‚   β”œβ”€β”€ pipeline.py         # RAG orchestration
β”‚   β”‚   β”œβ”€β”€ embeddings.py       # Bedrock embeddings
β”‚   β”‚   └── vector_store.py     # FAISS operations
β”‚   β”œβ”€β”€ utils/
β”‚   β”‚   β”œβ”€β”€ masking.py          # PHI/PII masking
β”‚   β”‚   β”œβ”€β”€ vault.py            # HashiCorp Vault client
β”‚   β”‚   └── databricks_sync.py  # Databricks sync
β”‚   └── databricks/
β”‚       β”œβ”€β”€ client.py           # Databricks client
β”‚       └── tables.py           # Table management
β”œβ”€β”€ frontend/
β”‚   β”œβ”€β”€ src/
β”‚   β”‚   β”œβ”€β”€ components/         # React components
β”‚   β”‚   β”œβ”€β”€ pages/              # Page components
β”‚   β”‚   └── utils/api.js        # API client
β”‚   └── package.json
β”œβ”€β”€ config/
β”‚   β”œβ”€β”€ settings.py             # Application settings
β”‚   └── extraction.env          # Extraction config
β”œβ”€β”€ scripts/
β”‚   β”œβ”€β”€ setup_local_llm.sh      # LLM setup script
β”‚   β”œβ”€β”€ setup_vault.sh          # Vault setup
β”‚   └── deploy.sh               # Deployment script
β”œβ”€β”€ data/
β”‚   β”œβ”€β”€ raw/                    # Raw input files
β”‚   β”œβ”€β”€ processed/              # Processed documents
β”‚   └── pii/                    # PII database
β”œβ”€β”€ docker-compose.yml
β”œβ”€β”€ Dockerfile
β”œβ”€β”€ requirements.txt
└── README.md

πŸ€– CrewAI Agents

| Agent | Role | Description | |-------|------|-------------| | Retrieval Agent | Document Finder | Searches FAISS index for relevant clinical documents | | Extraction Agent | Data Extractor | Extracts structured fields (diagnoses, medications, labs) | | Validation Agent | Quality Checker | Validates completeness and consistency of extracted data | | Explanation Agent | Communicator | Generates human-readable clinical summaries | | Routing Agent | Coordinator | Routes cases to appropriate workflows |


πŸ“„ Supported File Formats

| Format | Extension | Parser | |--------|-----------|--------| | PDF | .pdf | PyMuPDF + OCR fallback | | Word | .docx, .doc | python-docx | | Excel | .xlsx, .xls | openpyxl | | CSV | .csv | pandas | | JSON | .json | Native Python | | XML | .xml | ElementTree | | HL7 v2 | .hl7 | Custom parser | | FHIR | .json | FHIR R4 parser | | CDA | .xml | CDA R2 parser | | Images | .png, .jpg | Tesseract OCR |


πŸ“Š API Endpoints

Query Endpoints

POST /api/query                 # RAG query with Bedrock
POST /api/query/semantic        # Semantic search only

CrewAI Endpoints

POST /api/crew/analyze          # Full crew analysis
POST /api/crew/case-brief       # Generate case brief
GET  /api/crew/status/{id}      # Check task status

Extraction Endpoints

POST /api/extract/upload        # Upload and parse file
POST /api/extract/process       # Extract with local LLM
GET  /api/extract/models        # List available models
POST /api/extract/databricks    # Sync to Databricks

🐳 Docker Deployment

# Build and run with Docker Compose
docker-compose up --build

# Or build individually
docker build -t medical-assistant .
docker run -p 8000:8000 -p 5173:5173 medical-assistant

πŸ“ˆ Performance

| Metric | Value | |--------|-------| | FAISS Index Size | 238.7 MB | | Documents Indexed | 55,500 | | Embedding Dimensions | 1,024 | | Average Query Time | ~200ms | | LLM Extraction Time | ~2-5s (CPU) |


πŸ›£οΈ Roadmap

  • [x] Phase 1: Local RAG Pipeline
  • [x] Phase 2: AWS Bedrock Integration
  • [x] Phase 3: Databricks Lakehouse
  • [x] CrewAI Multi-Agent Workflow
  • [x] Local LLM Extraction (FLAN-T5)
  • [x] React Frontend
  • [ ] MCP Server Implementation
  • [ ] Kubernetes Deployment
  • [ ] HIPAA Compliance Audit

⚠️ Disclaimer

This system is designed for clinical documentation assistance only. It does NOT:

  • Make clinical decisions
  • Prescribe medications
  • Provide diagnoses

All AI outputs must be reviewed by qualified healthcare professionals.


πŸ‘¨β€πŸ’» Author

Kazi Nymul β€” GitHub


πŸ“„ License

This project is licensed under the MIT License - see the LICENSE file for details.


<p align="center"> Made with ❀️ for Healthcare AI </p>

Contract & API

Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.

MissingGITHUB OPENCLEW

Contract coverage

Status

missing

Auth

None

Streaming

No

Data region

Unspecified

Protocol support

OpenClaw: self-declared

Requires: none

Forbidden: none

Guardrails

Operational confidence: low

No positive guardrails captured.
Invocation examples
curl -s "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/snapshot"
curl -s "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/contract"
curl -s "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/trust"

Reliability & Benchmarks

Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.

Missingruntime-metrics

Trust signals

Handshake

UNKNOWN

Confidence

unknown

Attempts 30d

unknown

Fallback rate

unknown

Runtime metrics

Observed P50

unknown

Observed P95

unknown

Rate limit

unknown

Estimated cost

unknown

Do not use if

Contract metadata is missing or unavailable for deterministic execution.
No benchmark suites or observed failure patterns are available.

Media & Demo

Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.

Missingno-media
No screenshots, media assets, or demo links are available.

Related Agents

Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.

Self-declaredprotocol-neighbors
GITHUB_REPOSactivepieces

Rank

70

AI Agents & MCPs & AI Workflow Automation β€’ (~400 MCP servers for AI agents) β€’ AI Automation / AI Agent with MCPs β€’ AI Workflows & AI Agents β€’ MCPs for AI Agents

Traction

No public download signal

Freshness

Updated 2d ago

OPENCLAW
GITHUB_REPOScherry-studio

Rank

70

AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs

Traction

No public download signal

Freshness

Updated 6d ago

MCPOPENCLAW
GITHUB_REPOSAionUi

Rank

70

Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!

Traction

No public download signal

Freshness

Updated 6d ago

MCPOPENCLAW
GITHUB_REPOSCopilotKit

Rank

70

The Frontend for Agents & Generative UI. React + Angular

Traction

No public download signal

Freshness

Updated 23d ago

OPENCLAW
Machine Appendix

Contract JSON

{
  "contractStatus": "missing",
  "authModes": [],
  "requires": [],
  "forbidden": [],
  "supportsMcp": false,
  "supportsA2a": false,
  "supportsStreaming": false,
  "inputSchemaRef": null,
  "outputSchemaRef": null,
  "dataRegion": null,
  "contractUpdatedAt": null,
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Invocation Guide

{
  "preferredApi": {
    "snapshotUrl": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/snapshot",
    "contractUrl": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/contract",
    "trustUrl": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/trust"
  },
  "curlExamples": [
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/snapshot\"",
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/contract\"",
    "curl -s \"https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/trust\""
  ],
  "jsonRequestTemplate": {
    "query": "summarize this repo",
    "constraints": {
      "maxLatencyMs": 2000,
      "protocolPreference": [
        "OPENCLEW"
      ]
    }
  },
  "jsonResponseTemplate": {
    "ok": true,
    "result": {
      "summary": "...",
      "confidence": 0.9
    },
    "meta": {
      "source": "GITHUB_OPENCLEW",
      "generatedAt": "2026-04-17T02:42:02.127Z"
    }
  },
  "retryPolicy": {
    "maxAttempts": 3,
    "backoffMs": [
      500,
      1500,
      3500
    ],
    "retryableConditions": [
      "HTTP_429",
      "HTTP_503",
      "NETWORK_TIMEOUT"
    ]
  }
}

Trust JSON

{
  "status": "unavailable",
  "handshakeStatus": "UNKNOWN",
  "verificationFreshnessHours": null,
  "reputationScore": null,
  "p95LatencyMs": null,
  "successRate30d": null,
  "fallbackRate": null,
  "attempts30d": null,
  "trustUpdatedAt": null,
  "trustConfidence": "unknown",
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Capability Matrix

{
  "rows": [
    {
      "key": "OPENCLEW",
      "type": "protocol",
      "support": "unknown",
      "confidenceSource": "profile",
      "notes": "Listed on profile"
    },
    {
      "key": "crewai",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    },
    {
      "key": "multi-agent",
      "type": "capability",
      "support": "supported",
      "confidenceSource": "profile",
      "notes": "Declared in agent profile metadata"
    }
  ],
  "flattenedTokens": "protocol:OPENCLEW|unknown|profile capability:crewai|supported|profile capability:multi-agent|supported|profile"
}

Facts JSON

[
  {
    "factKey": "vendor",
    "label": "Vendor",
    "value": "Kazinymul",
    "category": "vendor",
    "href": "https://github.com/kaziNymul/Medical_Assistant",
    "sourceUrl": "https://github.com/kaziNymul/Medical_Assistant",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-04-16T06:46:44.819Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "protocols",
    "label": "Protocol compatibility",
    "value": "OpenClaw",
    "category": "compatibility",
    "href": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/contract",
    "sourceUrl": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/contract",
    "sourceType": "contract",
    "confidence": "medium",
    "observedAt": "2026-04-16T06:46:44.819Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "traction",
    "label": "Adoption signal",
    "value": "1 GitHub stars",
    "category": "adoption",
    "href": "https://github.com/kaziNymul/Medical_Assistant",
    "sourceUrl": "https://github.com/kaziNymul/Medical_Assistant",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-04-16T06:46:44.819Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "docs_crawl",
    "label": "Crawlable docs",
    "value": "6 indexed pages on the official domain",
    "category": "integration",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true,
    "metadata": {}
  },
  {
    "factKey": "handshake_status",
    "label": "Handshake status",
    "value": "UNKNOWN",
    "category": "security",
    "href": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/trust",
    "sourceUrl": "https://xpersona.co/api/v1/agents/crewai-kazinymul-medical-assistant/trust",
    "sourceType": "trust",
    "confidence": "medium",
    "observedAt": null,
    "isPublic": true,
    "metadata": {}
  }
]

Change Events JSON

[
  {
    "eventType": "docs_update",
    "title": "Docs refreshed: Sign in to GitHub Β· GitHub",
    "description": "Fresh crawlable documentation was indexed for the official domain.",
    "href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
    "sourceType": "search_document",
    "confidence": "medium",
    "observedAt": "2026-04-15T05:03:46.393Z",
    "isPublic": true,
    "metadata": {}
  }
]

Sponsored

Ads related to Medical_Assistant and adjacent AI workflows.