{"id":"f7c15ad0-9e84-4385-a41e-d1714825e9d0","entityType":"agent","slug":"clawhub-jonisjongithub-venice-ai","name":"Venice AI","canonicalUrl":"https://xpersona.co/agent/clawhub-jonisjongithub-venice-ai","canonicalPath":"/agent/clawhub-jonisjongithub-venice-ai","generatedAt":"2026-04-17T05:05:45.913Z","source":"CLAWHUB","claimStatus":"UNCLAIMED","verificationTier":"NONE","summary":{"evidence":{"source":"editorial-content","verified":true,"confidence":"high","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":null},"description":"Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything. Skill: Venice AI Owner: jonisjongithub Summary: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything. Tags: latest:2.0.0 Version history: v2.0.0 | 2026-02-07T19:25:39.635Z | user 🎉 Major update: Merged venice-ai-media into unified skill **New in v2.0.0:** - Complete Veni","descriptionLabel":"Technical summary","evidenceSummary":"Capability contract not published. No trust telemetry is available yet. 2K downloads reported by the source. Last updated 4/15/2026.","installCommand":"clawhub skill install kn7eeaajfmabdgahexes49syrn80f3b9:venice-ai","sourceUrl":"https://clawhub.ai/jonisjongithub/venice-ai","homepage":"https://clawhub.ai/jonisjongithub/venice-ai","primaryLinks":[{"label":"View on ClawHub","url":"https://clawhub.ai/jonisjongithub/venice-ai","kind":"source"}],"safetyScore":84,"overallRank":62,"popularityScore":66,"trustScore":null,"claimedByName":null,"isOwner":false,"seoDescription":"Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private"},"coverage":{"evidence":{"source":"public-profile","verified":false,"confidence":"medium","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":null},"protocols":[{"protocol":"OPENCLEW","label":"OpenClaw","status":"self-declared","notes":"Declared in the public agent profile."}],"capabilities":[],"verifiedCount":0,"selfDeclaredCount":1,"capabilityMatrix":{"rows":[{"key":"OPENCLEW","type":"protocol","support":"unknown","confidenceSource":"profile","notes":"Listed on profile"}],"flattenedTokens":"protocol:OPENCLEW|unknown|profile"}},"adoption":{"evidence":{"source":"CLAWHUB","verified":false,"confidence":"medium","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":null},"stars":null,"forks":null,"downloads":1999,"packageName":null,"latestVersion":"2.0.0","tractionLabel":"2K downloads"},"release":{"evidence":{"source":"CLAWHUB","verified":false,"confidence":"medium","updatedAt":"2026-02-28T19:30:02.219Z","emptyReason":null},"lastUpdatedAt":"2026-04-15T00:45:39.800Z","lastCrawledAt":"2026-02-28T19:30:02.219Z","lastIndexedAt":null,"nextCrawlAt":"2026-03-01T19:30:02.219Z","lastVerifiedAt":null,"highlights":[{"version":"2.0.0","createdAt":"2026-02-07T19:25:39.635Z","changelog":"🎉 Major update: Merged venice-ai-media into unified skill **New in v2.0.0:** - Complete Venice platform in one skill (text, audio, images, video) - Added image generation (Flux, GPT-Image, etc.) - Added video creation (Sora, WAN, Runway) - Added image upscaling with AI enhancement - Added AI image editing - Backward compatible with existing venice-ai-media configs One skill to unlock everything Venice has to offer!","fileCount":9,"zipByteSize":31743},{"version":"1.0.1","createdAt":"2026-02-07T06:21:13.865Z","changelog":"Updated skill","fileCount":4,"zipByteSize":15060},{"version":"1.0.0","createdAt":"2026-02-03T22:50:52.393Z","changelog":"Venice AI 1.0.0 — Initial Release - Introduces the full Venice AI platform skill with private, uncensored AI inference. - Supports model discovery, text generation with web search/citations, embeddings, TTS, speech-to-text, and model comparison. - Compatible with 30+ models, including GPT-5.2 and Claude Opus 4.5; supports advanced features like character personas and reasoning control. - OpenAI-compatible API; easy setup using environment variables or Clawdbot config. - Includes a comprehensive CLI for all workflows with detailed usage instructions.","fileCount":null,"zipByteSize":null}]},"execution":{"evidence":{"source":"CLAWHUB","verified":false,"confidence":"low","updatedAt":null,"emptyReason":"No published capability contract is available yet."},"installCommand":"clawhub skill install kn7eeaajfmabdgahexes49syrn80f3b9:venice-ai","setupComplexity":"low","setupSteps":["Setup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.","Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data."],"contract":{"contractStatus":"missing","authModes":[],"requires":[],"forbidden":[],"supportsMcp":false,"supportsA2a":false,"supportsStreaming":false,"inputSchemaRef":null,"outputSchemaRef":null,"dataRegion":null,"contractUpdatedAt":null,"sourceUpdatedAt":null,"freshnessSeconds":null},"invocationGuide":{"preferredApi":{"snapshotUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/snapshot","contractUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/contract","trustUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/trust"},"curlExamples":["curl -s \"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/snapshot\"","curl -s \"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/contract\"","curl -s \"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/trust\""],"jsonRequestTemplate":{"query":"summarize this repo","constraints":{"maxLatencyMs":2000,"protocolPreference":["OPENCLEW"]}},"jsonResponseTemplate":{"ok":true,"result":{"summary":"...","confidence":0.9},"meta":{"source":"CLAWHUB","generatedAt":"2026-04-17T05:05:45.912Z"}},"retryPolicy":{"maxAttempts":3,"backoffMs":[500,1500,3500],"retryableConditions":["HTTP_429","HTTP_503","NETWORK_TIMEOUT"]}},"endpoints":{"dossierUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/dossier","snapshotUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/snapshot","contractUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/contract","trustUrl":"https://xpersona.co/api/v1/agents/clawhub-jonisjongithub-venice-ai/trust"}},"reliability":{"evidence":{"source":"runtime-metrics","verified":false,"confidence":"low","updatedAt":null,"emptyReason":"No trust, reliability, or runtime telemetry is available."},"trust":{"status":"unavailable","handshakeStatus":"UNKNOWN","verificationFreshnessHours":null,"reputationScore":null,"p95LatencyMs":null,"successRate30d":null,"fallbackRate":null,"attempts30d":null,"trustUpdatedAt":null,"trustConfidence":"unknown","sourceUpdatedAt":null,"freshnessSeconds":null},"decisionGuardrails":{"doNotUseIf":["Contract metadata is missing or unavailable for deterministic execution."],"safeUseWhen":[],"riskFlags":["missing_or_unavailable_contract","trust_data_unavailable","schema_references_missing"],"operationalConfidence":"low"},"executionMetrics":{"observedLatencyMsP50":null,"observedLatencyMsP95":null,"estimatedCostUsd":null,"uptime30d":null,"rateLimitRpm":null,"rateLimitBurst":null,"lastVerifiedAt":null,"verificationSource":null},"runtimeMetrics":{"successRate":null,"avgLatencyMs":null,"avgCostUsd":null,"hallucinationRate":null,"retryRate":null,"disputeRate":null,"p50Latency":null,"p95Latency":null,"lastUpdated":null}},"benchmarks":{"evidence":{"source":"no-benchmark-data","verified":false,"confidence":"low","updatedAt":null,"emptyReason":"No benchmark suites or observed failure patterns are available."},"suites":[],"failurePatterns":[]},"artifacts":{"evidence":{"source":"CLAWHUB","verified":false,"confidence":"high","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":null},"readme":"Skill: Venice AI\n\nOwner: jonisjongithub\n\nSummary: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything.\n\nTags: latest:2.0.0\n\nVersion history:\n\nv2.0.0 | 2026-02-07T19:25:39.635Z | user\n\n🎉 Major update: Merged venice-ai-media into unified skill\n\n**New in v2.0.0:**\n- Complete Venice platform in one skill (text, audio, images, video)\n- Added image generation (Flux, GPT-Image, etc.)\n- Added video creation (Sora, WAN, Runway)\n- Added image upscaling with AI enhancement\n- Added AI image editing\n- Backward compatible with existing venice-ai-media configs\n\nOne skill to unlock everything Venice has to offer!\n\nv1.0.1 | 2026-02-07T06:21:13.865Z | user\n\nUpdated skill\n\nv1.0.0 | 2026-02-03T22:50:52.393Z | auto\n\nVenice AI 1.0.0 — Initial Release\n\n- Introduces the full Venice AI platform skill with private, uncensored AI inference.\n- Supports model discovery, text generation with web search/citations, embeddings, TTS, speech-to-text, and model comparison.\n- Compatible with 30+ models, including GPT-5.2 and Claude Opus 4.5; supports advanced features like character personas and reasoning control.\n- OpenAI-compatible API; easy setup using environment variables or Clawdbot config.\n- Includes a comprehensive CLI for all workflows with detailed usage instructions.\n\nArchive index:\n\nArchive v2.0.0: 9 files, 31743 bytes\n\nFiles: references/api.md (7106b), scripts/venice_common.py (7463b), scripts/venice-edit.py (6527b), scripts/venice-image.py (14862b), scripts/venice-upscale.py (9651b), scripts/venice-video.py (16858b), scripts/venice.py (22146b), SKILL.md (14084b), _meta.json (128b)\n\nFile v2.0.0:SKILL.md\n\n---\nname: venice-ai\ndescription: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything.\nhomepage: https://venice.ai\nmetadata:\n  {\n    \"clawdbot\":\n      {\n        \"emoji\": \"🏛️\",\n        \"requires\": { \"bins\": [\"python3\"], \"env\": [\"VENICE_API_KEY\"] },\n        \"primaryEnv\": \"VENICE_API_KEY\",\n        \"notes\": \"Full Venice AI platform — text, images, video, audio, embeddings. One skill for everything.\",\n        \"install\":\n          [\n            {\n              \"id\": \"python-brew\",\n              \"kind\": \"brew\",\n              \"formula\": \"python\",\n              \"bins\": [\"python3\"],\n              \"label\": \"Install Python (brew)\",\n            },\n          ],\n      },\n  }\n---\n\n# Venice AI — Complete Platform Skill\n\n## What is Venice AI?\n\n**Venice AI is the AI platform for people who don't want Big Tech watching over their shoulder.** Unlike OpenAI, Anthropic, and Google — where every prompt is logged, analyzed, and potentially used to train future models — Venice offers **true privacy** with zero data retention on private models. Your conversations stay yours.\n\nVenice is also **uncensored**: no content filters, no refusals, no \"I can't help with that.\" The AI answers what you actually ask, whether that's creative fiction, security research, controversial topics, or anything else mainstream AI platforms won't touch.\n\n**Why agents should care:**\n- **Private inference** — Models marked \"Private\" have zero data retention. Process sensitive data without it leaking into training sets.\n- **Uncensored responses** — No guardrails blocking legitimate use cases. Generate any content your workflow needs.\n- **OpenAI-compatible API** — Drop-in replacement. Same API format, just change the base URL.\n- **30+ models** — From tiny efficient models to Claude Opus 4.5, GPT-5.2, and Venice's own uncensored models.\n- **Built-in web search** — LLMs can search the web and cite sources in a single API call.\n- **Image & video generation** — Flux, Sora, Runway, WAN models for visual content.\n\nThis skill gives you the **complete Venice platform**: text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing.\n\n> **⚠️ API changes:** If something doesn't work as expected, check [docs.venice.ai](https://docs.venice.ai) — the API specs may have been updated since this skill was written.\n\n## Prerequisites\n\n- **Python 3.10+**\n- **Venice API key** (free tier available at [venice.ai/settings/api](https://venice.ai/settings/api))\n\n## Setup\n\n### Get Your API Key\n\n1. Create account at [venice.ai](https://venice.ai)\n2. Go to [venice.ai/settings/api](https://venice.ai/settings/api)\n3. Click \"Create API Key\" → copy the key (starts with `vn_...`)\n\n### Configure\n\n**Option A: Environment variable**\n```bash\nexport VENICE_API_KEY=\"vn_your_key_here\"\n```\n\n**Option B: Clawdbot config** (recommended)\n```json5\n// ~/.clawdbot/clawdbot.json\n{\n  skills: {\n    entries: {\n      \"venice-ai\": {\n        env: { VENICE_API_KEY: \"vn_your_key_here\" }\n      }\n    }\n  }\n}\n```\n\n### Verify\n```bash\npython3 {baseDir}/scripts/venice.py models --type text\n```\n\n## Scripts Overview\n\n| Script | Purpose |\n|--------|---------|\n| `venice.py` | Text generation, models, embeddings, TTS, transcription |\n| `venice-image.py` | Image generation (Flux, etc.) |\n| `venice-video.py` | Video generation (Sora, WAN, Runway) |\n| `venice-upscale.py` | Image upscaling |\n| `venice-edit.py` | AI image editing |\n\n---\n\n# Part 1: Text & Audio\n\n## Model Discovery & Selection\n\nVenice has a huge model catalog spanning text, image, video, audio, and embeddings.\n\n### Browse Models\n```bash\n# List all text models\npython3 {baseDir}/scripts/venice.py models --type text\n\n# List image models\npython3 {baseDir}/scripts/venice.py models --type image\n\n# List all model types\npython3 {baseDir}/scripts/venice.py models --type text,image,video,audio,embedding\n\n# Get details on a specific model\npython3 {baseDir}/scripts/venice.py models --filter llama\n```\n\n### Model Selection Guide\n\n| Need | Recommended Model | Why |\n|------|------------------|-----|\n| **Cheapest text** | `qwen3-4b` ($0.05/M in) | Tiny, fast, efficient |\n| **Best uncensored** | `venice-uncensored` ($0.20/M in) | Venice's own uncensored model |\n| **Best private + smart** | `deepseek-v3.2` ($0.40/M in) | Great reasoning, efficient |\n| **Vision/multimodal** | `qwen3-vl-235b-a22b` ($0.25/M in) | Sees images |\n| **Best coding** | `qwen3-coder-480b-a35b-instruct` ($0.75/M in) | Massive coder model |\n| **Frontier (budget)** | `grok-41-fast` ($0.50/M in) | Fast, 262K context |\n| **Frontier (max quality)** | `claude-opus-4-6` ($6/M in) | Best overall quality |\n| **Reasoning** | `kimi-k2-5` ($0.75/M in) | Strong chain-of-thought |\n| **Web search** | Any model + `enable_web_search` | Built-in web search |\n\n---\n\n## Text Generation (Chat Completions)\n\n### Basic Generation\n```bash\n# Simple prompt\npython3 {baseDir}/scripts/venice.py chat \"What is the meaning of life?\"\n\n# Choose a model\npython3 {baseDir}/scripts/venice.py chat \"Explain quantum computing\" --model deepseek-v3.2\n\n# System prompt\npython3 {baseDir}/scripts/venice.py chat \"Review this code\" --system \"You are a senior engineer.\"\n\n# Read from stdin\necho \"Summarize this\" | python3 {baseDir}/scripts/venice.py chat --model qwen3-4b\n\n# Stream output\npython3 {baseDir}/scripts/venice.py chat \"Write a story\" --stream\n```\n\n### Web Search Integration\n```bash\n# Auto web search (model decides when to search)\npython3 {baseDir}/scripts/venice.py chat \"What happened in tech news today?\" --web-search auto\n\n# Force web search with citations\npython3 {baseDir}/scripts/venice.py chat \"Current Bitcoin price\" --web-search on --web-citations\n\n# Web scraping (extracts content from URLs in prompt)\npython3 {baseDir}/scripts/venice.py chat \"Summarize: https://example.com/article\" --web-scrape\n```\n\n### Uncensored Mode\n```bash\n# Use Venice's own uncensored model\npython3 {baseDir}/scripts/venice.py chat \"Your question\" --model venice-uncensored\n\n# Disable Venice system prompts for raw model output\npython3 {baseDir}/scripts/venice.py chat \"Your prompt\" --no-venice-system-prompt\n```\n\n### Reasoning Models\n```bash\n# Use a reasoning model with effort control\npython3 {baseDir}/scripts/venice.py chat \"Solve this math problem...\" --model kimi-k2-5 --reasoning-effort high\n\n# Strip thinking from output\npython3 {baseDir}/scripts/venice.py chat \"Debug this code\" --model qwen3-4b --strip-thinking\n```\n\n### Advanced Options\n```bash\n# Temperature and token control\npython3 {baseDir}/scripts/venice.py chat \"Be creative\" --temperature 1.2 --max-tokens 4000\n\n# JSON output mode\npython3 {baseDir}/scripts/venice.py chat \"List 5 colors as JSON\" --json\n\n# Prompt caching (for repeated context)\npython3 {baseDir}/scripts/venice.py chat \"Question\" --cache-key my-session-123\n\n# Show usage stats\npython3 {baseDir}/scripts/venice.py chat \"Hello\" --show-usage\n```\n\n---\n\n## Embeddings\n\nGenerate vector embeddings for semantic search, RAG, and recommendations:\n\n```bash\n# Single text\npython3 {baseDir}/scripts/venice.py embed \"Venice is a private AI platform\"\n\n# Multiple texts (batch)\npython3 {baseDir}/scripts/venice.py embed \"first text\" \"second text\" \"third text\"\n\n# From file (one text per line)\npython3 {baseDir}/scripts/venice.py embed --file texts.txt\n\n# Output as JSON\npython3 {baseDir}/scripts/venice.py embed \"some text\" --output json\n```\n\nModel: `text-embedding-bge-m3` (private, $0.15/M tokens)\n\n---\n\n## Text-to-Speech (TTS)\n\nConvert text to speech with 60+ multilingual voices:\n\n```bash\n# Default voice\npython3 {baseDir}/scripts/venice.py tts \"Hello, welcome to Venice AI\"\n\n# Choose a voice\npython3 {baseDir}/scripts/venice.py tts \"Exciting news!\" --voice af_nova\n\n# List available voices\npython3 {baseDir}/scripts/venice.py tts --list-voices\n\n# Custom output path\npython3 {baseDir}/scripts/venice.py tts \"Some text\" --output /tmp/speech.mp3\n\n# Adjust speed\npython3 {baseDir}/scripts/venice.py tts \"Speaking slowly\" --speed 0.8\n```\n\n**Popular voices:** `af_sky`, `af_nova`, `am_liam`, `bf_emma`, `zf_xiaobei` (Chinese), `jm_kumo` (Japanese)\n\nModel: `tts-kokoro` (private, $3.50/M characters)\n\n---\n\n## Speech-to-Text (Transcription)\n\nTranscribe audio files to text:\n\n```bash\n# Transcribe a file\npython3 {baseDir}/scripts/venice.py transcribe audio.wav\n\n# With timestamps\npython3 {baseDir}/scripts/venice.py transcribe recording.mp3 --timestamps\n\n# From URL\npython3 {baseDir}/scripts/venice.py transcribe --url https://example.com/audio.wav\n```\n\nSupported formats: WAV, FLAC, MP3, M4A, AAC, MP4\n\nModel: `nvidia/parakeet-tdt-0.6b-v3` (private, $0.0001/audio second)\n\n---\n\n## Check Balance\n\n```bash\npython3 {baseDir}/scripts/venice.py balance\n```\n\n---\n\n# Part 2: Images & Video\n\n## Pricing Overview\n\n| Feature | Cost |\n|---------|------|\n| Image generation | ~$0.01-0.03 per image |\n| Image upscale | ~$0.02-0.04 |\n| Image edit | $0.04 |\n| Video (WAN) | ~$0.10-0.50 |\n| Video (Sora) | ~$0.50-2.00 |\n| Video (Runway) | ~$0.20-1.00 |\n\nUse `--quote` with video commands to check pricing before generation.\n\n---\n\n## Image Generation\n\n```bash\n# Basic generation\npython3 {baseDir}/scripts/venice-image.py --prompt \"a serene canal in Venice at sunset\"\n\n# Multiple images\npython3 {baseDir}/scripts/venice-image.py --prompt \"cyberpunk city\" --count 4\n\n# Custom dimensions\npython3 {baseDir}/scripts/venice-image.py --prompt \"portrait\" --width 768 --height 1024\n\n# List available models and styles\npython3 {baseDir}/scripts/venice-image.py --list-models\npython3 {baseDir}/scripts/venice-image.py --list-styles\n\n# Use specific model and style\npython3 {baseDir}/scripts/venice-image.py --prompt \"fantasy\" --model flux-2-pro --style-preset \"Cinematic\"\n\n# Reproducible results with seed\npython3 {baseDir}/scripts/venice-image.py --prompt \"abstract\" --seed 12345\n```\n\n**Key flags:** `--prompt`, `--model` (default: flux-2-max), `--count`, `--width`, `--height`, `--format` (webp/png/jpeg), `--resolution` (1K/2K/4K), `--aspect-ratio`, `--negative-prompt`, `--style-preset`, `--cfg-scale` (0-20), `--seed`, `--safe-mode`, `--hide-watermark`, `--embed-exif`\n\n---\n\n## Image Upscale\n\n```bash\n# 2x upscale\npython3 {baseDir}/scripts/venice-upscale.py photo.jpg --scale 2\n\n# 4x with AI enhancement\npython3 {baseDir}/scripts/venice-upscale.py photo.jpg --scale 4 --enhance\n\n# Enhanced with custom prompt\npython3 {baseDir}/scripts/venice-upscale.py photo.jpg --enhance --enhance-prompt \"sharpen details\"\n\n# From URL\npython3 {baseDir}/scripts/venice-upscale.py --url \"https://example.com/image.jpg\" --scale 2\n```\n\n**Key flags:** `--scale` (1-4, default: 2), `--enhance` (AI enhancement), `--enhance-prompt`, `--enhance-creativity` (0.0-1.0), `--url`, `--output`\n\n---\n\n## Image Edit\n\nAI-powered image editing:\n\n```bash\n# Add elements\npython3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt \"add sunglasses\"\n\n# Modify scene\npython3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt \"change the sky to sunset\"\n\n# Remove objects\npython3 {baseDir}/scripts/venice-edit.py photo.jpg --prompt \"remove the person in background\"\n\n# From URL\npython3 {baseDir}/scripts/venice-edit.py --url \"https://example.com/image.jpg\" --prompt \"colorize\"\n```\n\n**Note:** The edit endpoint uses Qwen-Image which has some content restrictions.\n\n---\n\n## Video Generation\n\n```bash\n# Get price quote first\npython3 {baseDir}/scripts/venice-video.py --quote --model wan-2.6-image-to-video --duration 10s\n\n# Image-to-video (WAN - default)\npython3 {baseDir}/scripts/venice-video.py --image photo.jpg --prompt \"camera pans slowly\" --duration 10s\n\n# Image-to-video (Sora)\npython3 {baseDir}/scripts/venice-video.py --image photo.jpg --prompt \"cinematic\" \\\n  --model sora-2-image-to-video --duration 8s --aspect-ratio 16:9 --skip-audio-param\n\n# Video-to-video (Runway Gen4)\npython3 {baseDir}/scripts/venice-video.py --video input.mp4 --prompt \"anime style\" \\\n  --model runway-gen4-turbo-v2v\n\n# List models with available durations\npython3 {baseDir}/scripts/venice-video.py --list-models\n```\n\n**Key flags:** `--image` or `--video`, `--prompt`, `--model` (default: wan-2.6-image-to-video), `--duration`, `--resolution` (480p/720p/1080p), `--aspect-ratio`, `--audio`/`--no-audio`, `--quote`, `--timeout`\n\n**Models:**\n- **WAN** — Image-to-video, configurable audio, 5s-21s\n- **Sora** — Requires `--aspect-ratio`, use `--skip-audio-param`\n- **Runway** — Video-to-video transformation\n\n---\n\n# Tips & Ideas\n\n### 🔍 Web Search + LLM = Research Assistant\nUse `--web-search on --web-citations` to build a research workflow. Venice searches the web, synthesizes results, and cites sources — all in one API call.\n\n### 🔓 Uncensored Creative Content\nVenice's uncensored models work for both text AND images. No guardrails blocking legitimate creative use cases.\n\n### 🎯 Prompt Caching for Agents\nIf you're running an agent loop that sends the same system prompt repeatedly, use `--cache-key` to get up to 90% cost savings.\n\n### 🎤 Audio Pipeline\nCombine TTS and transcription: generate spoken content with `tts`, process audio with `transcribe`. Both are private inference.\n\n### 🎬 Video Workflow\n1. Generate or find a base image\n2. Use `--quote` to estimate video cost\n3. Generate with appropriate duration/model\n4. Videos take 1-5 minutes depending on settings\n\n---\n\n## Troubleshooting\n\n| Problem | Solution |\n|---------|----------|\n| `VENICE_API_KEY not set` | Set env var or configure in `~/.clawdbot/clawdbot.json` |\n| `Invalid API key` | Verify at [venice.ai/settings/api](https://venice.ai/settings/api) |\n| `Model not found` | Run `--list-models` to see available; use `--no-validate` for new models |\n| Rate limited | Check `--show-usage` output |\n| Video stuck | Videos can take 1-5 min; use `--timeout 600` for long ones |\n\n## Resources\n\n- **API Docs**: [docs.venice.ai](https://docs.venice.ai)\n- **Status**: [veniceai-status.com](https://veniceai-status.com)\n- **Discord**: [discord.gg/askvenice](https://discord.gg/askvenice)\n\nFile v2.0.0:_meta.json\n\n{\n  \"ownerId\": \"kn7eeaajfmabdgahexes49syrn80f3b9\",\n  \"slug\": \"venice-ai\",\n  \"version\": \"2.0.0\",\n  \"publishedAt\": 1770492339635\n}\n\nFile v2.0.0:references/api.md\n\n# Venice AI API Reference\n\n**Base URL:** `https://api.venice.ai/api/v1`\n\n**Authentication:** All requests require `Authorization: Bearer <VENICE_API_KEY>`\n\nVenice implements the **OpenAI API specification** — any OpenAI-compatible client works by changing the base URL.\n\n---\n\n## Chat Completions\n\n### Create Chat Completion\n```\nPOST /chat/completions\n```\n\n**Request Body:**\n```json\n{\n  \"model\": \"deepseek-v3.2\",\n  \"messages\": [\n    {\"role\": \"system\", \"content\": \"You are a helpful assistant.\"},\n    {\"role\": \"user\", \"content\": \"Hello!\"}\n  ],\n  \"temperature\": 0.7,\n  \"max_tokens\": 4096,\n  \"stream\": false,\n  \"response_format\": {\"type\": \"json_object\"},\n  \"reasoning_effort\": \"medium\",\n  \"prompt_cache_key\": \"session-123\",\n  \"venice_parameters\": {\n    \"enable_web_search\": \"auto\",\n    \"enable_web_citations\": true,\n    \"enable_web_scraping\": false,\n    \"include_venice_system_prompt\": true,\n    \"character_slug\": \"coder-dan\",\n    \"strip_thinking_response\": false,\n    \"disable_thinking\": false,\n    \"include_search_results_in_stream\": false,\n    \"return_search_results_as_documents\": false\n  }\n}\n```\n\n**Venice Parameters (unique to Venice):**\n\n| Parameter | Type | Default | Description |\n|-----------|------|---------|-------------|\n| `enable_web_search` | `\"off\"/\"on\"/\"auto\"` | `\"off\"` | LLM-integrated web search. \"auto\" lets the model decide. |\n| `enable_web_citations` | bool | false | Request `[REF]0[/REF]` citation format in responses |\n| `enable_web_scraping` | bool | false | Scrape URLs found in user messages to augment context |\n| `include_venice_system_prompt` | bool | true | Include Venice's default uncensored system prompt |\n| `character_slug` | string | — | Use a Venice public character persona |\n| `strip_thinking_response` | bool | false | Strip `<think>` tags server-side |\n| `disable_thinking` | bool | false | Disable reasoning entirely |\n| `include_search_results_in_stream` | bool | false | Emit search results as first SSE chunk |\n| `return_search_results_as_documents` | bool | false | Return search results as OpenAI-compatible tool call |\n\n**Response:**\n```json\n{\n  \"id\": \"chatcmpl-...\",\n  \"object\": \"chat.completion\",\n  \"model\": \"deepseek-v3.2\",\n  \"choices\": [{\n    \"index\": 0,\n    \"message\": {\n      \"role\": \"assistant\",\n      \"content\": \"Hello! How can I help?\",\n      \"reasoning_content\": \"The user said hello...\"\n    },\n    \"finish_reason\": \"stop\"\n  }],\n  \"usage\": {\n    \"prompt_tokens\": 15,\n    \"completion_tokens\": 8,\n    \"total_tokens\": 23,\n    \"prompt_tokens_details\": {\n      \"cached_tokens\": 0,\n      \"cache_creation_input_tokens\": 0\n    }\n  }\n}\n```\n\n**Model Feature Suffixes:** Append parameters to model name: `qwen3-4b:strip_thinking_response=true:disable_thinking=true`\n\n### Reasoning Models\n\nSupported: `claude-opus-4-6`, `grok-41-fast`, `kimi-k2-5`, `gemini-3-pro-preview`, `qwen3-235b-a22b-thinking-2507`, `qwen3-4b`, `deepseek-ai-DeepSeek-R1`\n\nControl via `reasoning_effort`: `low` | `medium` | `high`\n\n### Prompt Caching\n\nAutomatic for most models (>1024 tokens). Use `prompt_cache_key` for routing affinity.\nClaude requires explicit `cache_control: {\"type\": \"ephemeral\"}` markers.\n\n| Model | Min Tokens | Cache Life | Read Discount |\n|-------|-----------|------------|---------------|\n| Claude Opus 4.6 | ~4,000 | 5 min | 90% |\n| GPT-5.2 | 1,024 | 5-10 min | 90% |\n| Gemini | ~1,024 | 1 hour | 75-90% |\n| DeepSeek | ~1,024 | 5 min | 50% |\n\n---\n\n## Models\n\n### List Models\n```\nGET /models?type={text|image|video|audio|embedding}\n```\n\n**Response:**\n```json\n{\n  \"data\": [{\n    \"id\": \"deepseek-v3.2\",\n    \"type\": \"text\",\n    \"model_spec\": {\n      \"description\": \"DeepSeek V3.2\",\n      \"offline\": false,\n      \"beta\": false,\n      \"availableForPrivateInference\": true,\n      \"deprecation\": {\"date\": null},\n      \"constraints\": {\n        \"max_context_length\": 164000\n      }\n    }\n  }]\n}\n```\n\n---\n\n## Embeddings\n\n### Generate Embeddings\n```\nPOST /embeddings\n```\n\n**Request:**\n```json\n{\n  \"model\": \"text-embedding-bge-m3\",\n  \"input\": \"Your text here\"\n}\n```\n\nOr batch: `\"input\": [\"text1\", \"text2\", \"text3\"]`\n\n**Response:**\n```json\n{\n  \"data\": [{\n    \"index\": 0,\n    \"embedding\": [0.123, -0.456, ...],\n    \"object\": \"embedding\"\n  }],\n  \"usage\": {\"prompt_tokens\": 5, \"total_tokens\": 5}\n}\n```\n\n---\n\n## Audio\n\n### Text-to-Speech\n```\nPOST /audio/speech\n```\n\n**Request:**\n```json\n{\n  \"model\": \"tts-kokoro\",\n  \"input\": \"Hello world\",\n  \"voice\": \"af_sky\",\n  \"speed\": 1.0\n}\n```\n\n**Response:** Audio bytes (MP3). Content-Type: audio/mpeg\n\n**Available voices (60+):**\n- English US: `af_sky`, `af_nova`, `af_bella`, `am_adam`, `am_liam`\n- English UK: `bf_emma`, `bf_isabella`, `bm_daniel`, `bm_george`\n- Chinese: `zf_xiaobei`, `zf_xiaoni`, `zm_yunjian`\n- Japanese: `jf_alpha`, `jm_kumo`\n- And many more (French, Hindi, Italian, Portuguese, Spanish)\n\nPrefix: a=American, b=British, z=Chinese, j=Japanese; f=female, m=male\n\n### Speech-to-Text (Transcription)\n```\nPOST /audio/transcriptions\n```\n\n**Request:** Multipart form data\n- `file`: Audio file (WAV, FLAC, MP3, M4A, AAC, MP4)\n- `model`: `nvidia/parakeet-tdt-0.6b-v3`\n- `timestamps`: `true` (optional, word-level timing)\n\n**Response:**\n```json\n{\n  \"text\": \"Transcribed text here...\"\n}\n```\n\n---\n\n## Images\n\n### Generate Image\n```\nPOST /images/generations\n```\n\n### Edit Image\n```\nPOST /images/edits\n```\n\n### Upscale Image\n```\nPOST /images/upscale\n```\n\n> See `venice-image.py`, `venice-upscale.py`, and `venice-edit.py` in the scripts folder for CLI usage.\n\n---\n\n## Video\n\n### Generate Video\n```\nPOST /video/generate\n```\n\n### Get Video Quote\n```\nPOST /video/generate/quote\n```\n\n> See `venice-video.py` in the scripts folder for CLI usage.\n\n---\n\n## Response Headers\n\nAll authenticated requests include useful headers:\n\n| Header | Description |\n|--------|-------------|\n| `x-venice-balance-usd` | USD credit balance |\n| `x-venice-balance-diem` | DIEM token balance |\n| `x-venice-balance-vcu` | Venice Compute Units |\n| `x-venice-model-id` | Model used for inference |\n| `x-ratelimit-remaining-requests` | Remaining request quota |\n| `x-ratelimit-remaining-tokens` | Remaining token quota |\n| `CF-RAY` | Request ID (for support) |\n\n---\n\n## Pricing Quick Reference\n\n### Text (per 1M tokens)\n\n| Model | Input | Output | Privacy |\n|-------|-------|--------|---------|\n| qwen3-4b | $0.05 | $0.15 | Private |\n| venice-uncensored | $0.20 | $0.90 | Private |\n| deepseek-v3.2 | $0.40 | $1.00 | Private |\n| mistral-31-24b | $0.50 | $2.00 | Private |\n| llama-3.3-70b | $0.70 | $2.80 | Private |\n| grok-41-fast | $0.50 | $1.25 | Anonymized |\n| openai-gpt-52 | $2.19 | $17.50 | Anonymized |\n| claude-opus-4-6 | $6.00 | $30.00 | Anonymized |\n\n### Other\n\n| Feature | Cost |\n|---------|------|\n| Embeddings (BGE-M3) | $0.15/M tokens input |\n| TTS (Kokoro) | $3.50/M characters |\n| Speech-to-Text (Parakeet) | $0.0001/audio second |\n| Web Search | $10/1K calls |\n| Web Scraping | $10/1K calls |\n| Images | $0.01-$0.23/image |\n| Video | Variable (use quote API) |\n\n---\n\n## Resources\n\n- **Docs:** https://docs.venice.ai\n- **Status:** https://veniceai-status.com\n- **Discord:** https://discord.gg/askvenice\n- **Twitter:** https://x.com/AskVenice\n\nArchive v1.0.1: 4 files, 15060 bytes\n\nFiles: references/api.md (7105b), scripts/venice.py (22146b), SKILL.md (12623b), _meta.json (128b)\n\nFile v1.0.1:SKILL.md\n\n---\nname: venice-ai\ndescription: Full-platform Venice AI skill — model discovery, text generation with web search, embeddings, TTS, speech-to-text, and model comparison. Private, uncensored AI inference. Use when you need to generate text with specific models, search the web through an LLM, transcribe audio, create embeddings, or explore what models are available.\nhomepage: https://venice.ai\nmetadata:\n  {\n    \"clawdbot\":\n      {\n        \"emoji\": \"🏛️\",\n        \"requires\": { \"bins\": [\"python3\"], \"env\": [\"VENICE_API_KEY\"] },\n        \"primaryEnv\": \"VENICE_API_KEY\",\n        \"notes\": \"Full Venice AI platform access. For image/video generation, see venice-ai-media skill.\",\n        \"install\":\n          [\n            {\n              \"id\": \"python-brew\",\n              \"kind\": \"brew\",\n              \"formula\": \"python\",\n              \"bins\": [\"python3\"],\n              \"label\": \"Install Python (brew)\",\n            },\n          ],\n      },\n  }\n---\n\n# Venice AI — Full Platform Skill\n\n## What is Venice AI?\n\n**Venice AI is the AI platform for people who don't want Big Tech watching over their shoulder.** Unlike OpenAI, Anthropic, and Google — where every prompt is logged, analyzed, and potentially used to train future models — Venice offers **true privacy** with zero data retention on private models. Your conversations stay yours.\n\nVenice is also **uncensored**: no content filters, no refusals, no \"I can't help with that.\" The AI answers what you actually ask, whether that's creative fiction, security research, controversial topics, or anything else mainstream AI platforms won't touch.\n\n**Why agents should care:**\n- **Private inference** — Models marked \"Private\" have zero data retention. Process sensitive data without it leaking into training sets.\n- **Uncensored responses** — No guardrails blocking legitimate use cases. Generate any content your workflow needs.\n- **OpenAI-compatible API** — Drop-in replacement. Same API format, just change the base URL.\n- **30+ models** — From tiny efficient models to Claude Opus 4.5, GPT-5.2, and Venice's own uncensored models.\n- **Built-in web search** — LLMs can search the web and cite sources in a single API call.\n\nThis skill gives you the **full Venice platform**: model discovery, text generation with Venice-specific superpowers (web search, uncensored mode, character personas, reasoning control), embeddings, TTS, speech-to-text, and intelligent model selection.\n\n> **For image & video generation**, use the companion `venice-ai-media` skill which has dedicated tools for those workflows.\n>\n> **⚠️ API changes:** If something doesn't work as expected, check [docs.venice.ai](https://docs.venice.ai) — the API specs may have been updated since this skill was written.\n\n## Prerequisites\n\n- **Python 3.10+**\n- **Venice API key** (free tier available at [venice.ai/settings/api](https://venice.ai/settings/api))\n\n## Setup\n\n### Get Your API Key\n\n1. Create account at [venice.ai](https://venice.ai)\n2. Go to [venice.ai/settings/api](https://venice.ai/settings/api)\n3. Click \"Create API Key\" → copy the key (starts with `vn_...`)\n\n### Configure\n\n**Option A: Environment variable**\n```bash\nexport VENICE_API_KEY=\"vn_your_key_here\"\n```\n\n**Option B: Clawdbot config** (recommended)\n```json5\n// ~/.clawdbot/clawdbot.json\n{\n  skills: {\n    entries: {\n      \"venice-ai\": {\n        env: { VENICE_API_KEY: \"vn_your_key_here\" }\n      }\n    }\n  }\n}\n```\n\n### Verify\n```bash\npython3 {baseDir}/scripts/venice.py models --type text\n```\n\n## Scripts\n\nAll operations go through a single CLI tool:\n\n```bash\npython3 {baseDir}/scripts/venice.py [command] [options]\n```\n\n---\n\n## Model Discovery & Selection\n\nVenice has a huge model catalog spanning text, image, video, audio, and embeddings. The right model for a task depends on your needs: cost, speed, privacy, context length, and capabilities.\n\n### Browse Models\n```bash\n# List all text models\npython3 {baseDir}/scripts/venice.py models --type text\n\n# List image models\npython3 {baseDir}/scripts/venice.py models --type image\n\n# List all model types\npython3 {baseDir}/scripts/venice.py models --type text,image,video,audio,embedding\n\n# Get details on a specific model\npython3 {baseDir}/scripts/venice.py models --filter llama\n```\n\n### Model Selection Guide\n\n| Need | Recommended Model | Why |\n|------|------------------|-----|\n| **Cheapest text** | `qwen3-4b` ($0.05/M in) | Tiny, fast, efficient |\n| **Best uncensored** | `venice-uncensored` ($0.20/M in) | Venice's own uncensored model |\n| **Best private + smart** | `deepseek-v3.2` ($0.40/M in) | Great reasoning, efficient |\n| **Vision/multimodal** | `qwen3-vl-235b-a22b` ($0.25/M in) | Sees images |\n| **Best coding** | `qwen3-coder-480b-a35b-instruct` ($0.75/M in) | Massive coder model |\n| **Frontier (budget)** | `grok-41-fast` ($0.50/M in) | Fast, 262K context |\n| **Frontier (max quality)** | `claude-opus-4-6` ($6/M in) | Best overall quality (latest Opus) |\n| **Reasoning** | `kimi-k2-5` ($0.75/M in) | Strong chain-of-thought (K2.5) |\n| **Web search** | Any model + `enable_web_search` | Built-in web search |\n\n> **Privacy tiers:** \"Private\" = zero data retention. \"Anonymized\" = logs stripped of identity but may be retained.\n\n---\n\n## Text Generation (Chat Completions)\n\nVenice implements the OpenAI chat completions API with extra superpowers.\n\n### Basic Generation\n```bash\n# Simple prompt\npython3 {baseDir}/scripts/venice.py chat \"What is the meaning of life?\"\n\n# Choose a model\npython3 {baseDir}/scripts/venice.py chat \"Explain quantum computing\" --model deepseek-v3.2\n\n# System prompt\npython3 {baseDir}/scripts/venice.py chat \"Review this code\" --system \"You are a senior engineer. Be direct and critical.\"\n\n# Read from stdin (pipe content in)\necho \"Summarize this\" | python3 {baseDir}/scripts/venice.py chat --model qwen3-4b\n\n# Stream output\npython3 {baseDir}/scripts/venice.py chat \"Write a story\" --stream\n```\n\n### Web Search Integration\nVenice can search the web before answering — no external tools needed:\n```bash\n# Auto web search (model decides when to search)\npython3 {baseDir}/scripts/venice.py chat \"What happened in tech news today?\" --web-search auto\n\n# Force web search\npython3 {baseDir}/scripts/venice.py chat \"Current Bitcoin price\" --web-search on\n\n# Web search with citations\npython3 {baseDir}/scripts/venice.py chat \"Latest AI research papers\" --web-search on --web-citations\n\n# Web scraping (extracts content from URLs in prompt)\npython3 {baseDir}/scripts/venice.py chat \"Summarize this article: https://example.com/article\" --web-scrape\n```\n\n### Uncensored Mode\n```bash\n# Use Venice's own uncensored model\npython3 {baseDir}/scripts/venice.py chat \"Your uncensored question\" --model venice-uncensored\n\n# Disable Venice system prompts for raw model output\npython3 {baseDir}/scripts/venice.py chat \"Your prompt\" --no-venice-system-prompt\n```\n\n### Reasoning Models\n```bash\n# Use a reasoning model with effort control\npython3 {baseDir}/scripts/venice.py chat \"Solve this math problem...\" --model kimi-k2-5 --reasoning-effort high\n\n# Strip thinking from output\npython3 {baseDir}/scripts/venice.py chat \"Debug this code\" --model qwen3-4b --strip-thinking\n\n# Disable thinking entirely (faster, cheaper)\npython3 {baseDir}/scripts/venice.py chat \"Simple question\" --model qwen3-4b --disable-thinking\n```\n\n### Character Personas\nVenice has public character personas that customize model behavior:\n```bash\n# Use a Venice character\npython3 {baseDir}/scripts/venice.py chat \"Tell me a story\" --character coder-dan\n```\n\n### Advanced Options\n```bash\n# Temperature and token control\npython3 {baseDir}/scripts/venice.py chat \"Be creative\" --temperature 1.2 --max-tokens 4000\n\n# JSON output mode\npython3 {baseDir}/scripts/venice.py chat \"List 5 colors as JSON\" --json\n\n# Prompt caching (for multi-turn or repeated context)\npython3 {baseDir}/scripts/venice.py chat \"Question about the doc\" --cache-key my-session-123\n\n# Show usage stats (tokens, cost, cache hits)\npython3 {baseDir}/scripts/venice.py chat \"Hello\" --show-usage\n```\n\n---\n\n## Embeddings\n\nGenerate vector embeddings for semantic search, RAG, and recommendations:\n\n```bash\n# Single text\npython3 {baseDir}/scripts/venice.py embed \"Venice is a private AI platform\"\n\n# Multiple texts (batch)\npython3 {baseDir}/scripts/venice.py embed \"first text\" \"second text\" \"third text\"\n\n# From file (one text per line)\npython3 {baseDir}/scripts/venice.py embed --file texts.txt\n\n# Output as JSON\npython3 {baseDir}/scripts/venice.py embed \"some text\" --output json\n```\n\nModel: `text-embedding-bge-m3` (private, $0.15/M tokens input)\n\n---\n\n## Text-to-Speech (TTS)\n\nConvert text to speech with 60+ multilingual voices:\n\n```bash\n# Default voice\npython3 {baseDir}/scripts/venice.py tts \"Hello, welcome to Venice AI\"\n\n# Choose a voice\npython3 {baseDir}/scripts/venice.py tts \"Exciting news!\" --voice af_nova\n\n# List available voices\npython3 {baseDir}/scripts/venice.py tts --list-voices\n\n# Custom output path\npython3 {baseDir}/scripts/venice.py tts \"Some text\" --output /tmp/speech.mp3\n\n# Adjust speed\npython3 {baseDir}/scripts/venice.py tts \"Speaking slowly\" --speed 0.8\n```\n\n**Popular voices:** `af_sky`, `af_nova`, `am_liam`, `bf_emma`, `zf_xiaobei` (Chinese), `jm_kumo` (Japanese)\n\nModel: `tts-kokoro` (private, $3.50/M characters)\n\n---\n\n## Speech-to-Text (Transcription)\n\nTranscribe audio files to text:\n\n```bash\n# Transcribe a file\npython3 {baseDir}/scripts/venice.py transcribe audio.wav\n\n# With timestamps\npython3 {baseDir}/scripts/venice.py transcribe recording.mp3 --timestamps\n\n# From URL\npython3 {baseDir}/scripts/venice.py transcribe --url https://example.com/audio.wav\n```\n\nSupported formats: WAV, FLAC, MP3, M4A, AAC, MP4\n\nModel: `nvidia/parakeet-tdt-0.6b-v3` (private, $0.0001/audio second — essentially free)\n\n---\n\n## Check Balance\n\n```bash\npython3 {baseDir}/scripts/venice.py balance\n```\n\nShows your Diem, USD, and VCU balances.\n\n---\n\n## Tips & Ideas to Try\n\n### 🔍 Web Search + LLM = Research Assistant\nUse `--web-search on --web-citations` to build a research workflow. Venice searches the web, synthesizes results, and cites sources — all in one API call. Try different models to see which gives the best summaries.\n\n### 🔓 Uncensored Creative Writing\nVenice's uncensored models don't have the guardrails that restrict other AI platforms. Great for fiction, roleplay scenarios, security research, or any topic other AIs refuse to engage with.\n\n### 🧠 Model A/B Testing\nNot sure which model is best for your task? Use the `chat` command with different `--model` flags and compare. Smaller models are surprisingly capable and much cheaper.\n\n### 🔒 Privacy-First Workflows\nIf you're processing sensitive data, stick to \"Private\" models (shown in `models` output). Zero data retention means your prompts literally can't leak.\n\n### 🎯 Prompt Caching for Agents\nIf you're running an agent loop that sends the same system prompt repeatedly, use `--cache-key` to get up to 90% cost savings on the cached portion.\n\n### 🎤 Audio Pipeline\nCombine TTS and transcription for audio workflows: generate spoken content with `tts`, process audio with `transcribe`. Both are private inference.\n\n### 💡 Share What You Build\nCreated something cool with Venice? The community at [discord.gg/askvenice](https://discord.gg/askvenice) loves seeing creative uses. Venice's Twitter [@AskVenice](https://x.com/AskVenice) also showcases community projects.\n\n---\n\n## Model Feature Suffixes\n\nVenice supports inline model configuration via suffixes — append parameters directly to the model name:\n\n```\nmodel_name:param1=value1:param2=value2\n```\n\nExamples:\n```bash\n# Strip thinking tags server-side\n--model \"qwen3-4b:strip_thinking_response=true\"\n\n# Disable thinking entirely\n--model \"qwen3-4b:disable_thinking=true\"\n```\n\nUseful when you can't pass `venice_parameters` directly (e.g., through OpenAI-compatible clients).\n\n---\n\n## Troubleshooting\n\n| Problem | Solution |\n|---------|----------|\n| `VENICE_API_KEY not set` | Set env var or configure in `~/.clawdbot/clawdbot.json` |\n| `Invalid API key` | Verify at [venice.ai/settings/api](https://venice.ai/settings/api) — keys start with `vn_` |\n| `Model not found` | Run `models --type text` to see available models |\n| Rate limited | Check `--show-usage` output for rate limit info |\n| Slow responses | Try a smaller/faster model, or reduce `--max-tokens` |\n\n## Resources\n\n- **API Docs**: [docs.venice.ai](https://docs.venice.ai)\n- **Status**: [veniceai-status.com](https://veniceai-status.com)\n- **Discord**: [discord.gg/askvenice](https://discord.gg/askvenice)\n- **Full API reference**: See `references/api.md` in this skill\n\nFile v1.0.1:_meta.json\n\n{\n  \"ownerId\": \"kn7eeaajfmabdgahexes49syrn80f3b9\",\n  \"slug\": \"venice-ai\",\n  \"version\": \"1.0.1\",\n  \"publishedAt\": 1770445273865\n}\n\nFile v1.0.1:references/api.md\n\n# Venice AI API Reference\n\n**Base URL:** `https://api.venice.ai/api/v1`\n\n**Authentication:** All requests require `Authorization: Bearer <VENICE_API_KEY>`\n\nVenice implements the **OpenAI API specification** — any OpenAI-compatible client works by changing the base URL.\n\n---\n\n## Chat Completions\n\n### Create Chat Completion\n```\nPOST /chat/completions\n```\n\n**Request Body:**\n```json\n{\n  \"model\": \"deepseek-v3.2\",\n  \"messages\": [\n    {\"role\": \"system\", \"content\": \"You are a helpful assistant.\"},\n    {\"role\": \"user\", \"content\": \"Hello!\"}\n  ],\n  \"temperature\": 0.7,\n  \"max_tokens\": 4096,\n  \"stream\": false,\n  \"response_format\": {\"type\": \"json_object\"},\n  \"reasoning_effort\": \"medium\",\n  \"prompt_cache_key\": \"session-123\",\n  \"venice_parameters\": {\n    \"enable_web_search\": \"auto\",\n    \"enable_web_citations\": true,\n    \"enable_web_scraping\": false,\n    \"include_venice_system_prompt\": true,\n    \"character_slug\": \"coder-dan\",\n    \"strip_thinking_response\": false,\n    \"disable_thinking\": false,\n    \"include_search_results_in_stream\": false,\n    \"return_search_results_as_documents\": false\n  }\n}\n```\n\n**Venice Parameters (unique to Venice):**\n\n| Parameter | Type | Default | Description |\n|-----------|------|---------|-------------|\n| `enable_web_search` | `\"off\"/\"on\"/\"auto\"` | `\"off\"` | LLM-integrated web search. \"auto\" lets the model decide. |\n| `enable_web_citations` | bool | false | Request `[REF]0[/REF]` citation format in responses |\n| `enable_web_scraping` | bool | false | Scrape URLs found in user messages to augment context |\n| `include_venice_system_prompt` | bool | true | Include Venice's default uncensored system prompt |\n| `character_slug` | string | — | Use a Venice public character persona |\n| `strip_thinking_response` | bool | false | Strip `<think>` tags server-side |\n| `disable_thinking` | bool | false | Disable reasoning entirely |\n| `include_search_results_in_stream` | bool | false | Emit search results as first SSE chunk |\n| `return_search_results_as_documents` | bool | false | Return search results as OpenAI-compatible tool call |\n\n**Response:**\n```json\n{\n  \"id\": \"chatcmpl-...\",\n  \"object\": \"chat.completion\",\n  \"model\": \"deepseek-v3.2\",\n  \"choices\": [{\n    \"index\": 0,\n    \"message\": {\n      \"role\": \"assistant\",\n      \"content\": \"Hello! How can I help?\",\n      \"reasoning_content\": \"The user said hello...\"\n    },\n    \"finish_reason\": \"stop\"\n  }],\n  \"usage\": {\n    \"prompt_tokens\": 15,\n    \"completion_tokens\": 8,\n    \"total_tokens\": 23,\n    \"prompt_tokens_details\": {\n      \"cached_tokens\": 0,\n      \"cache_creation_input_tokens\": 0\n    }\n  }\n}\n```\n\n**Model Feature Suffixes:** Append parameters to model name: `qwen3-4b:strip_thinking_response=true:disable_thinking=true`\n\n### Reasoning Models\n\nSupported: `claude-opus-4-6`, `grok-41-fast`, `kimi-k2-5`, `gemini-3-pro-preview`, `qwen3-235b-a22b-thinking-2507`, `qwen3-4b`, `deepseek-ai-DeepSeek-R1`\n\nControl via `reasoning_effort`: `low` | `medium` | `high`\n\n### Prompt Caching\n\nAutomatic for most models (>1024 tokens). Use `prompt_cache_key` for routing affinity.\nClaude requires explicit `cache_control: {\"type\": \"ephemeral\"}` markers.\n\n| Model | Min Tokens | Cache Life | Read Discount |\n|-------|-----------|------------|---------------|\n| Claude Opus 4.6 | ~4,000 | 5 min | 90% |\n| GPT-5.2 | 1,024 | 5-10 min | 90% |\n| Gemini | ~1,024 | 1 hour | 75-90% |\n| DeepSeek | ~1,024 | 5 min | 50% |\n\n---\n\n## Models\n\n### List Models\n```\nGET /models?type={text|image|video|audio|embedding}\n```\n\n**Response:**\n```json\n{\n  \"data\": [{\n    \"id\": \"deepseek-v3.2\",\n    \"type\": \"text\",\n    \"model_spec\": {\n      \"description\": \"DeepSeek V3.2\",\n      \"offline\": false,\n      \"beta\": false,\n      \"availableForPrivateInference\": true,\n      \"deprecation\": {\"date\": null},\n      \"constraints\": {\n        \"max_context_length\": 164000\n      }\n    }\n  }]\n}\n```\n\n---\n\n## Embeddings\n\n### Generate Embeddings\n```\nPOST /embeddings\n```\n\n**Request:**\n```json\n{\n  \"model\": \"text-embedding-bge-m3\",\n  \"input\": \"Your text here\"\n}\n```\n\nOr batch: `\"input\": [\"text1\", \"text2\", \"text3\"]`\n\n**Response:**\n```json\n{\n  \"data\": [{\n    \"index\": 0,\n    \"embedding\": [0.123, -0.456, ...],\n    \"object\": \"embedding\"\n  }],\n  \"usage\": {\"prompt_tokens\": 5, \"total_tokens\": 5}\n}\n```\n\n---\n\n## Audio\n\n### Text-to-Speech\n```\nPOST /audio/speech\n```\n\n**Request:**\n```json\n{\n  \"model\": \"tts-kokoro\",\n  \"input\": \"Hello world\",\n  \"voice\": \"af_sky\",\n  \"speed\": 1.0\n}\n```\n\n**Response:** Audio bytes (MP3). Content-Type: audio/mpeg\n\n**Available voices (60+):**\n- English US: `af_sky`, `af_nova`, `af_bella`, `am_adam`, `am_liam`\n- English UK: `bf_emma`, `bf_isabella`, `bm_daniel`, `bm_george`\n- Chinese: `zf_xiaobei`, `zf_xiaoni`, `zm_yunjian`\n- Japanese: `jf_alpha`, `jm_kumo`\n- And many more (French, Hindi, Italian, Portuguese, Spanish)\n\nPrefix: a=American, b=British, z=Chinese, j=Japanese; f=female, m=male\n\n### Speech-to-Text (Transcription)\n```\nPOST /audio/transcriptions\n```\n\n**Request:** Multipart form data\n- `file`: Audio file (WAV, FLAC, MP3, M4A, AAC, MP4)\n- `model`: `nvidia/parakeet-tdt-0.6b-v3`\n- `timestamps`: `true` (optional, word-level timing)\n\n**Response:**\n```json\n{\n  \"text\": \"Transcribed text here...\"\n}\n```\n\n---\n\n## Images\n\n### Generate Image\n```\nPOST /images/generations\n```\n\n### Edit Image\n```\nPOST /images/edits\n```\n\n### Upscale Image\n```\nPOST /images/upscale\n```\n\n> For detailed image/video API usage, see the `venice-ai-media` skill which has dedicated scripts.\n\n---\n\n## Video\n\n### Generate Video\n```\nPOST /video/generate\n```\n\n### Get Video Quote\n```\nPOST /video/generate/quote\n```\n\n> For detailed video API usage, see the `venice-ai-media` skill.\n\n---\n\n## Response Headers\n\nAll authenticated requests include useful headers:\n\n| Header | Description |\n|--------|-------------|\n| `x-venice-balance-usd` | USD credit balance |\n| `x-venice-balance-diem` | DIEM token balance |\n| `x-venice-balance-vcu` | Venice Compute Units |\n| `x-venice-model-id` | Model used for inference |\n| `x-ratelimit-remaining-requests` | Remaining request quota |\n| `x-ratelimit-remaining-tokens` | Remaining token quota |\n| `CF-RAY` | Request ID (for support) |\n\n---\n\n## Pricing Quick Reference\n\n### Text (per 1M tokens)\n\n| Model | Input | Output | Privacy |\n|-------|-------|--------|---------|\n| qwen3-4b | $0.05 | $0.15 | Private |\n| venice-uncensored | $0.20 | $0.90 | Private |\n| deepseek-v3.2 | $0.40 | $1.00 | Private |\n| mistral-31-24b | $0.50 | $2.00 | Private |\n| llama-3.3-70b | $0.70 | $2.80 | Private |\n| grok-41-fast | $0.50 | $1.25 | Anonymized |\n| openai-gpt-52 | $2.19 | $17.50 | Anonymized |\n| claude-opus-4-6 | $6.00 | $30.00 | Anonymized |\n\n### Other\n\n| Feature | Cost |\n|---------|------|\n| Embeddings (BGE-M3) | $0.15/M tokens input |\n| TTS (Kokoro) | $3.50/M characters |\n| Speech-to-Text (Parakeet) | $0.0001/audio second |\n| Web Search | $10/1K calls |\n| Web Scraping | $10/1K calls |\n| Images | $0.01-$0.23/image |\n| Video | Variable (use quote API) |\n\n---\n\n## Resources\n\n- **Docs:** https://docs.venice.ai\n- **Status:** https://veniceai-status.com\n- **Discord:** https://discord.gg/askvenice\n- **Twitter:** https://x.com/AskVenice","readmeExcerpt":"Skill: Venice AI Owner: jonisjongithub Summary: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything. Tags: latest:2.0.0 Version history: v2.0.0 | 2026-02-07T19:25:39.635Z | user 🎉 Major update: Merged venice-ai-media into unified skill **New in v2.0.0:** - Complete Veni","codeSnippets":[],"executableExamples":[{"language":"bash","snippet":"export VENICE_API_KEY=\"vn_your_key_here\""},{"language":"json5","snippet":"// ~/.clawdbot/clawdbot.json\n{\n  skills: {\n    entries: {\n      \"venice-ai\": {\n        env: { VENICE_API_KEY: \"vn_your_key_here\" }\n      }\n    }\n  }\n}"},{"language":"bash","snippet":"python3 {baseDir}/scripts/venice.py models --type text"},{"language":"bash","snippet":"# List all text models\npython3 {baseDir}/scripts/venice.py models --type text\n\n# List image models\npython3 {baseDir}/scripts/venice.py models --type image\n\n# List all model types\npython3 {baseDir}/scripts/venice.py models --type text,image,video,audio,embedding\n\n# Get details on a specific model\npython3 {baseDir}/scripts/venice.py models --filter llama"},{"language":"bash","snippet":"# Simple prompt\npython3 {baseDir}/scripts/venice.py chat \"What is the meaning of life?\"\n\n# Choose a model\npython3 {baseDir}/scripts/venice.py chat \"Explain quantum computing\" --model deepseek-v3.2\n\n# System prompt\npython3 {baseDir}/scripts/venice.py chat \"Review this code\" --system \"You are a senior engineer.\"\n\n# Read from stdin\necho \"Summarize this\" | python3 {baseDir}/scripts/venice.py chat --model qwen3-4b\n\n# Stream output\npython3 {baseDir}/scripts/venice.py chat \"Write a story\" --stream"},{"language":"bash","snippet":"# Auto web search (model decides when to search)\npython3 {baseDir}/scripts/venice.py chat \"What happened in tech news today?\" --web-search auto\n\n# Force web search with citations\npython3 {baseDir}/scripts/venice.py chat \"Current Bitcoin price\" --web-search on --web-citations\n\n# Web scraping (extracts content from URLs in prompt)\npython3 {baseDir}/scripts/venice.py chat \"Summarize: https://example.com/article\" --web-scrape"}],"parameters":null,"dependencies":[],"permissions":[],"extractedFiles":[{"path":"SKILL.md","content":"---\nname: venice-ai\ndescription: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything.\nhomepage: https://venice.ai\nmetadata:\n  {\n    \"clawdbot\":\n      {\n        \"emoji\": \"🏛️\",\n        \"requires\": { \"bins\": [\"python3\"], \"env\": [\"VENICE_API_KEY\"] },\n        \"primaryEnv\": \"VENICE_API_KEY\",\n        \"notes\": \"Full Venice AI platform — text, images, video, audio, embeddings. One skill for everything.\",\n        \"install\":\n          [\n            {\n              \"id\": \"python-brew\",\n              \"kind\": \"brew\",\n              \"formula\": \"python\",\n              \"bins\": [\"python3\"],\n              \"label\": \"Install Python (brew)\",\n            },\n          ],\n      },\n  }\n---\n\n# Venice AI — Complete Platform Skill\n\n## What is Venice AI?\n\n**Venice AI is the AI platform for people who don't want Big Tech watching over their shoulder.** Unlike OpenAI, Anthropic, and Google — where every prompt is logged, analyzed, and potentially used to train future models — Venice offers **true privacy** with zero data retention on private models. Your conversations stay yours.\n\nVenice is also **uncensored**: no content filters, no refusals, no \"I can't help with that.\" The AI answers what you actually ask, whether that's creative fiction, security research, controversial topics, or anything else mainstream AI platforms won't touch.\n\n**Why agents should care:**\n- **Private inference** — Models marked \"Private\" have zero data retention. Process sensitive data without it leaking into training sets.\n- **Uncensored responses** — No guardrails blocking legitimate use cases. Generate any content your workflow needs.\n- **OpenAI-compatible API** — Drop-in replacement. Same API format, just change the base URL.\n- **30+ models** — From tiny efficient models to Claude Opus 4.5, GPT-5.2, and Venice's own uncensored models.\n- **Built-in web search** — LLMs can search the web and cite sources in a single API call.\n- **Image & video generation** — Flux, Sora, Runway, WAN models for visual content.\n\nThis skill gives you the **complete Venice platform**: text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing.\n\n> **⚠️ API changes:** If something doesn't work as expected, check [docs.venice.ai](https://docs.venice.ai) — the API specs may have been updated since this skill was written.\n\n## Prerequisites\n\n- **Python 3.10+**\n- **Venice API key** (free tier available at [venice.ai/settings/api](https://venice.ai/settings/api))\n\n## Setup\n\n### Get Your API Key\n\n1. Create account at [venice.ai](https://venice.ai)\n2. Go to [venice.ai/settings/api](https://venice.ai/settings/api)\n3. Click \"Create API Key\" → copy the key (starts with `vn_...`)\n\n### Configure\n\n**Option A: Environment variable**\n```bash\nexport VENICE_API_KEY=\"vn_your_key_here\"\n```\n\n**Option B: Clawdbot config** (recommended)\n``"},{"path":"_meta.json","content":"{\n  \"ownerId\": \"kn7eeaajfmabdgahexes49syrn80f3b9\",\n  \"slug\": \"venice-ai\",\n  \"version\": \"2.0.0\",\n  \"publishedAt\": 1770492339635\n}"},{"path":"references/api.md","content":"# Venice AI API Reference\n\n**Base URL:** `https://api.venice.ai/api/v1`\n\n**Authentication:** All requests require `Authorization: Bearer <VENICE_API_KEY>`\n\nVenice implements the **OpenAI API specification** — any OpenAI-compatible client works by changing the base URL.\n\n---\n\n## Chat Completions\n\n### Create Chat Completion\n```\nPOST /chat/completions\n```\n\n**Request Body:**\n```json\n{\n  \"model\": \"deepseek-v3.2\",\n  \"messages\": [\n    {\"role\": \"system\", \"content\": \"You are a helpful assistant.\"},\n    {\"role\": \"user\", \"content\": \"Hello!\"}\n  ],\n  \"temperature\": 0.7,\n  \"max_tokens\": 4096,\n  \"stream\": false,\n  \"response_format\": {\"type\": \"json_object\"},\n  \"reasoning_effort\": \"medium\",\n  \"prompt_cache_key\": \"session-123\",\n  \"venice_parameters\": {\n    \"enable_web_search\": \"auto\",\n    \"enable_web_citations\": true,\n    \"enable_web_scraping\": false,\n    \"include_venice_system_prompt\": true,\n    \"character_slug\": \"coder-dan\",\n    \"strip_thinking_response\": false,\n    \"disable_thinking\": false,\n    \"include_search_results_in_stream\": false,\n    \"return_search_results_as_documents\": false\n  }\n}\n```\n\n**Venice Parameters (unique to Venice):**\n\n| Parameter | Type | Default | Description |\n|-----------|------|---------|-------------|\n| `enable_web_search` | `\"off\"/\"on\"/\"auto\"` | `\"off\"` | LLM-integrated web search. \"auto\" lets the model decide. |\n| `enable_web_citations` | bool | false | Request `[REF]0[/REF]` citation format in responses |\n| `enable_web_scraping` | bool | false | Scrape URLs found in user messages to augment context |\n| `include_venice_system_prompt` | bool | true | Include Venice's default uncensored system prompt |\n| `character_slug` | string | — | Use a Venice public character persona |\n| `strip_thinking_response` | bool | false | Strip `<think>` tags server-side |\n| `disable_thinking` | bool | false | Disable reasoning entirely |\n| `include_search_results_in_stream` | bool | false | Emit search results as first SSE chunk |\n| `return_search_results_as_documents` | bool | false | Return search results as OpenAI-compatible tool call |\n\n**Response:**\n```json\n{\n  \"id\": \"chatcmpl-...\",\n  \"object\": \"chat.completion\",\n  \"model\": \"deepseek-v3.2\",\n  \"choices\": [{\n    \"index\": 0,\n    \"message\": {\n      \"role\": \"assistant\",\n      \"content\": \"Hello! How can I help?\",\n      \"reasoning_content\": \"The user said hello...\"\n    },\n    \"finish_reason\": \"stop\"\n  }],\n  \"usage\": {\n    \"prompt_tokens\": 15,\n    \"completion_tokens\": 8,\n    \"total_tokens\": 23,\n    \"prompt_tokens_details\": {\n      \"cached_tokens\": 0,\n      \"cache_creation_input_tokens\": 0\n    }\n  }\n}\n```\n\n**Model Feature Suffixes:** Append parameters to model name: `qwen3-4b:strip_thinking_response=true:disable_thinking=true`\n\n### Reasoning Models\n\nSupported: `claude-opus-4-6`, `grok-41-fast`, `kimi-k2-5`, `gemini-3-pro-preview`, `qwen3-235b-a22b-thinking-2507`, `qwen3-4b`, `deepseek-ai-DeepSeek-R1`\n\nControl via `reasoning_effort`: `low` | `medium` | `high`\n\n### Prompt Caching\n\nAutomatic for most models (>10"}],"languages":[],"docsSourceLabel":"CLAWHUB","editorialOverview":"Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything. Skill: Venice AI Owner: jonisjongithub Summary: Complete Venice AI platform — text generation, web search, embeddings, TTS, speech-to-text, image generation, video creation, upscaling, and AI editing. Private, uncensored AI inference for everything. Tags: latest:2.0.0 Version history: v2.0.0 | 2026-02-07T19:25:39.635Z | user 🎉 Major update: Merged venice-ai-media into unified skill **New in v2.0.0:** - Complete Veni","editorialQuality":{"score":100,"threshold":65,"status":"ready","wordCount":1113,"uniquenessScore":53,"reasons":[]}},"media":{"evidence":{"source":"no-media","verified":false,"confidence":"low","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":"No screenshots, media assets, or demo links are available."},"primaryImageUrl":null,"mediaAssetCount":0,"assets":[],"demoUrl":null},"ownerResources":{"evidence":{"source":"unclaimed","verified":false,"confidence":"low","updatedAt":"2026-04-15T00:45:39.800Z","emptyReason":"This page has not been claimed by the agent owner."},"hasCustomPage":false,"customPageUpdatedAt":null,"customLinks":[],"structuredLinks":{"docsUrl":null,"demoUrl":null,"supportUrl":null,"pricingUrl":null,"statusUrl":null},"customPage":null},"relatedAgents":{"evidence":{"source":"protocol-neighbors","verified":false,"confidence":"medium","updatedAt":"2026-04-17T05:05:45.913Z","emptyReason":null},"items":[{"id":"b917f68a-ebff-438e-84f8-3f4b2494c0bc","entityType":"agent","canonicalPath":"/agent/activepieces-activepieces","slug":"activepieces-activepieces","name":"activepieces","description":"AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents","url":"https://github.com/activepieces/activepieces","homepage":"https://www.activepieces.com","source":"GITHUB_REPOS","protocols":["OPENCLAW"],"capabilities":[],"safetyScore":100,"overallRank":70,"updatedAt":"2026-04-15T02:22:12.426Z","createdAt":"2026-02-25T03:38:12.412Z","downloads":null},{"id":"5cb26759-3a39-483f-94cf-276a98c13bb8","entityType":"agent","canonicalPath":"/agent/cherryhq-cherry-studio","slug":"cherryhq-cherry-studio","name":"cherry-studio","description":"AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs","url":"https://github.com/CherryHQ/cherry-studio","homepage":"https://cherry-ai.com","source":"GITHUB_REPOS","protocols":["MCP","OPENCLAW"],"capabilities":[],"safetyScore":100,"overallRank":70,"updatedAt":"2026-04-11T14:38:40.986Z","createdAt":"2026-02-25T03:38:19.379Z","downloads":null},{"id":"8ebccd8e-3863-4187-8355-c3f14e1f9edf","entityType":"agent","canonicalPath":"/agent/iofficeai-aionui","slug":"iofficeai-aionui","name":"AionUi","description":"Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!","url":"https://github.com/iOfficeAI/AionUi","homepage":"https://www.aionui.com","source":"GITHUB_REPOS","protocols":["MCP","OPENCLAW"],"capabilities":[],"safetyScore":100,"overallRank":70,"updatedAt":"2026-04-10T18:48:31.762Z","createdAt":"2026-02-25T03:38:16.584Z","downloads":null},{"id":"6f6582d0-5d76-4f0f-b81d-86520247950b","entityType":"agent","canonicalPath":"/agent/copilotkit-copilotkit","slug":"copilotkit-copilotkit","name":"CopilotKit","description":"The Frontend for Agents & Generative UI. React + Angular","url":"https://github.com/CopilotKit/CopilotKit","homepage":"https://docs.copilotkit.ai","source":"GITHUB_REPOS","protocols":["OPENCLAW"],"capabilities":[],"safetyScore":100,"overallRank":70,"updatedAt":"2026-03-25T09:50:57.846Z","createdAt":"2026-02-25T03:39:14.617Z","downloads":null}],"links":{"hub":"/agent","source":"/agent/source/clawhub","protocols":[{"label":"OpenClaw","href":"/agent/protocol/openclew"}]}}}