Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Crawler Summary
High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). --- name: qwen3-tts description: > High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). version: "1.0" author: daMustermann repository: https://githu Capability contract not published. No trust telemetry is available yet. Last updated 2/25/2026.
Freshness
Last checked 2/25/2026
Best For
qwen3-tts is best for generate, discover, also workflows where OpenClaw compatibility matters.
Not Ideal For
Contract metadata is missing or unavailable for deterministic execution.
Evidence Sources Checked
editorial-content, GITHUB OPENCLEW, runtime-metrics, public facts pack
High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). --- name: qwen3-tts description: > High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). version: "1.0" author: daMustermann repository: https://githu
Public facts
4
Change events
1
Artifacts
0
Freshness
Feb 25, 2026
Capability contract not published. No trust telemetry is available yet. Last updated 2/25/2026.
Trust score
Unknown
Compatibility
OpenClaw
Freshness
Feb 25, 2026
Vendor
Damustermann
Artifacts
0
Benchmarks
0
Last release
Unpublished
Key links, install path, and a quick operational read before the deeper crawl record.
Summary
Capability contract not published. No trust telemetry is available yet. Last updated 2/25/2026.
Setup snapshot
git clone https://github.com/daMustermann/claw-qwen3-tts.gitSetup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.
Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.
Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.
Vendor
Damustermann
Protocol compatibility
OpenClaw
Handshake status
UNKNOWN
Crawlable docs
6 indexed pages on the official domain
Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.
Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.
Extracted files
0
Examples
6
Snippets
0
Languages
typescript
Parameters
bash
bash <(curl -fsSL https://raw.githubusercontent.com/daMustermann/claw-qwen3-tts/main/install.sh)
bash
bash ~/clawd/skills/qwen3-tts/install.sh
bash
# Start (idempotent — won't restart if already running) bash ~/clawd/skills/qwen3-tts/scripts/start_server.sh # Check health bash ~/clawd/skills/qwen3-tts/scripts/health_check.sh # Stop (when done) bash ~/clawd/skills/qwen3-tts/scripts/stop_server.sh
bash
curl -X POST http://localhost:8880/v1/audio/speech \
-H "Content-Type: application/json" \
-d '{bash
curl -X POST http://localhost:8880/v1/audio/speech \
-H "Content-Type: application/json" \
-d '{
"model": "custom-voice-1.7b",
"input": "TEXT_HERE",
"voice": "default",
"speaker": "Chelsie",
"language": "en",
"instruct": "",
"response_format": "wav"
}' \
--output ~/clawd/skills/qwen3-tts/output/speech.wavbash
curl -X POST http://localhost:8880/v1/audio/voice-design \
-H "Content-Type: application/json" \
-d '{Full documentation captured from public sources, including the complete README when available.
Docs source
GITHUB OPENCLEW
Editorial quality
ready
High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). --- name: qwen3-tts description: > High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). version: "1.0" author: daMustermann repository: https://githu
name: qwen3-tts description: > High-quality text-to-speech using Qwen3-TTS. 10 built-in speakers with emotional instruct control, voice cloning (3s of audio), natural-language voice design, 10+ languages, persistent named voices, and delivering audio via Telegram/WhatsApp as native voice messages. Auto-detects GPU hardware (CUDA, ROCm, Intel XPU, CPU). version: "1.0" author: daMustermann repository: https://github.com/daMustermann/claw-qwen3-tts license: MIT requires:
You have access to a powerful text-to-speech system that can generate human-quality speech with 10 built-in speakers, design new voices from descriptions, clone existing voices from audio samples, and send audio via Telegram/WhatsApp as native voice messages.
If the skill is not yet installed (no ~/clawd/skills/qwen3-tts directory), run:
bash <(curl -fsSL https://raw.githubusercontent.com/daMustermann/claw-qwen3-tts/main/install.sh)
Or if already cloned but not set up (no .venv/ directory):
bash ~/clawd/skills/qwen3-tts/install.sh
This auto-detects the GPU (CUDA, ROCm, Intel XPU, or CPU-only), creates a Python venv, and installs all dependencies. It takes 5–15 minutes on first run.
Before any TTS operation, ensure the server is running:
# Start (idempotent — won't restart if already running)
bash ~/clawd/skills/qwen3-tts/scripts/start_server.sh
# Check health
bash ~/clawd/skills/qwen3-tts/scripts/health_check.sh
# Stop (when done)
bash ~/clawd/skills/qwen3-tts/scripts/stop_server.sh
The server runs at http://localhost:8880.
| Model ID | Use Case | Notes |
|----------|----------|-------|
| custom-voice-1.7b | High-quality TTS with built-in speakers — default | Best quality, ~5 GB VRAM |
| custom-voice-0.6b | Fast TTS with built-in speakers | Lightweight, ~2 GB VRAM |
| voice-design | Design new voices from natural language descriptions | Uses VoiceDesign model |
| base-1.7b | Basic TTS (auto-corrected to custom-voice-1.7b) | Use custom-voice-* instead |
| base-0.6b | Basic TTS (auto-corrected to custom-voice-0.6b) | Use custom-voice-* instead |
Important: On the
/v1/audio/speechendpoint,base-*andvoice-designmodels are automatically corrected to the correspondingcustom-voice-*model. Always prefercustom-voice-1.7borcustom-voice-0.6bfor speech generation.
The custom-voice-* models include 10 built-in voices:
Chelsie · Ethan · Aidan · Serena · Ryan · Vivian · Claire · Lucas · Eleanor · Benjamin
You can discover speakers dynamically: curl http://localhost:8880/v1/speakers
When to use: User asks to speak text, read something aloud, generate audio, do a voiceover, narrate, or say something.
curl -X POST http://localhost:8880/v1/audio/speech \
-H "Content-Type: application/json" \
-d '{
"model": "custom-voice-1.7b",
"input": "TEXT_HERE",
"voice": "default",
"speaker": "Chelsie",
"language": "en",
"instruct": "",
"response_format": "wav"
}' \
--output ~/clawd/skills/qwen3-tts/output/speech.wav
Parameters:
| Parameter | Required | Default | Description |
|-----------|----------|---------|-------------|
| model | no | custom-voice-1.7b | TTS model to use |
| input | yes | — | The text to synthesize |
| voice | no | default | "default" for built-in speakers, or a saved voice name (e.g. "Angie") |
| speaker | no | Chelsie | Built-in speaker name (only when voice is "default") |
| language | no | en | Language code: en, zh, ja, ko, de, fr, ru, pt, es, it |
| instruct | no | "" | Emotional/style instruction (see below) |
| response_format | no | wav | Output format: wav, mp3, ogg, flac |
| speed | no | 1.0 | Speech speed multiplier |
Language codes: en, zh, ja, ko, de, fr, ru, pt, es, it — or full names like English, Chinese, German, etc.
Instruct examples (controls tone, emotion, and style):
"Speak happily and with excitement""Whisper softly, as if telling a secret""Read this in a calm, professional news anchor tone""用愤怒的语气" (Speak angrily — works in target language too)"" (empty string = neutral default)When voice is a saved name: If you pass "voice": "Angie" and a voice named "Angie" exists, the server uses voice cloning with the saved reference audio instead of a built-in speaker. The speaker field is ignored in this case.
When to use: User wants to create a custom voice, describe how a character should sound, design a persona's voice.
curl -X POST http://localhost:8880/v1/audio/voice-design \
-H "Content-Type: application/json" \
-d '{
"model": "voice-design",
"input": "TEXT_TO_SPEAK",
"voice_description": "DESCRIBE THE VOICE IN NATURAL LANGUAGE",
"language": "en",
"response_format": "wav"
}' \
--output ~/clawd/skills/qwen3-tts/output/designed.wav
Parameters:
| Parameter | Required | Default | Description |
|-----------|----------|---------|-------------|
| model | no | voice-design | Must be voice-design |
| input | yes | — | Text to synthesize with the designed voice |
| voice_description | yes | — | Natural language description of the desired voice |
| language | no | en | Target language |
| response_format | no | wav | Output format |
Example descriptions:
"A warm, deep male voice with a slight British accent, calm and authoritative, like a BBC presenter in his 40s""A young, energetic female voice, bright and cheerful, with a slight rasp""An old wizard with a slow, mysterious, gravelly voice"The response includes a X-Voice-Id header — capture it to save the voice (see §4).
When to use: User provides a reference audio clip and wants to generate new speech in that voice.
curl -X POST http://localhost:8880/v1/audio/voice-clone \
-F "reference_audio=@/path/to/reference.wav" \
-F "reference_text=Transcript of the reference audio" \
-F "input=New text to speak in the cloned voice" \
-F "language=en" \
-F "response_format=wav" \
--output ~/clawd/skills/qwen3-tts/output/cloned.wav
Parameters:
| Parameter | Required | Default | Description |
|-----------|----------|---------|-------------|
| reference_audio | yes | — | Audio file to clone the voice from |
| input | yes | — | New text to synthesize in the cloned voice |
| reference_text | no | "" | Transcription of the reference audio (improves quality) |
| language | no | en | Target language |
| response_format | no | wav | Output format |
Guidelines:
reference_text transcription significantly improves resultsreference_text is empty, uses x-vector-only mode (audio features only)The response includes a X-Voice-Id header — capture it to save the voice (see §4).
YOU MUST FOLLOW THESE RULES:
After EVERY voice-design or voice-clone request, ask the user:
"Would you like to save this voice for future use? What name should I give it?"
If the user says yes, capture the X-Voice-Id from the response headers and save it:
curl -X POST http://localhost:8880/v1/voices \
-H "Content-Type: application/json" \
-d '{
"name": "USER_CHOSEN_NAME",
"source_voice_id": "VOICE_ID_FROM_X_VOICE_ID_HEADER",
"description": "Description of the voice",
"tags": ["tag1", "tag2"],
"language": "en"
}'
When user requests TTS with a voice name (e.g. "say this with Angie"):
"voice": "Angie" in the /v1/audio/speech requestWhen user asks to list voices:
curl http://localhost:8880/v1/voices
Present the results as a formatted list with name, description, source, language, tags, and usage count. Voices are sorted by usage count (most used first).
When user asks to delete a voice: Confirm with the user first, then:
curl -X DELETE http://localhost:8880/v1/voices/VOICE_NAME
When user asks to rename a voice:
curl -X PATCH http://localhost:8880/v1/voices/OLD_NAME \
-H "Content-Type: application/json" \
-d '{"name": "NEW_NAME"}'
When user asks to update a voice's metadata (description, tags, language):
curl -X PATCH http://localhost:8880/v1/voices/VOICE_NAME \
-H "Content-Type: application/json" \
-d '{"description": "Updated description", "tags": ["new", "tags"]}'
Voice names are case-insensitive but stored in the casing the user provided.
No duplicate names allowed. If a name already exists, the save will fail (409). Ask the user for a different name or offer to delete the existing one first.
Voice profiles are stored locally in ~/clawd/skills/qwen3-tts/voices/ and persist across server restarts. Each voice consists of:
<name>.json — metadata<name>.pt — embedding tensor<name>_sample.wav — reference audio sample (used for re-cloning)When to use: User needs audio in a specific format, or you need to prepare audio for messaging.
curl -X POST http://localhost:8880/v1/audio/convert \
-F "audio=@input.wav" \
-F "target_format=mp3" \
--output output.mp3
Supported formats: wav, mp3, ogg (Opus), flac
You can also use the shell script directly:
bash ~/clawd/skills/qwen3-tts/scripts/convert_to_ogg_opus.sh input.wav output.ogg
When to use: User is interacting via Telegram, or explicitly asks to send audio to a Telegram chat.
curl -X POST http://localhost:8880/v1/audio/send/telegram \
-H "Content-Type: application/json" \
-d '{
"audio_file": "/path/to/audio.wav",
"chat_id": "CHAT_ID",
"bot_token": "BOT_TOKEN",
"caption": "Optional caption"
}'
bot_token is optional if already configured in config.jsonsendVoice APIWhen to use: User is interacting via WhatsApp, or explicitly asks to send audio there.
curl -X POST http://localhost:8880/v1/audio/send/whatsapp \
-H "Content-Type: application/json" \
-d '{
"audio_file": "/path/to/audio.wav",
"phone_number_id": "PHONE_ID",
"recipient": "+14155551234",
"access_token": "ACCESS_TOKEN"
}'
phone_number_id and access_token are optional if already configured in config.jsonUse these to dynamically discover available models and speakers:
# List all available TTS models
curl http://localhost:8880/v1/models
# List built-in speakers
curl http://localhost:8880/v1/speakers
# Server health check (device info, voice count, version)
curl http://localhost:8880/health
After generating speech:
After saving a voice:
voice: Captain Hook."After sending via Telegram/WhatsApp:
When choosing a speaker: If the user doesn't specify, default to "Chelsie". If they describe the kind of voice they want (but not a full voice-design request), pick the most fitting built-in speaker.
When choosing a model: Default to custom-voice-1.7b. Only use custom-voice-0.6b if the user asks for speed, or if the system has limited VRAM/memory.
The agent can update ~/clawd/skills/qwen3-tts/config.json to set:
custom-voice-1.7b or custom-voice-0.6bIf config.json doesn't exist, copy the template:
cp ~/clawd/skills/qwen3-tts/config.json.template ~/clawd/skills/qwen3-tts/config.json
Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.
Contract coverage
Status
missing
Auth
None
Streaming
No
Data region
Unspecified
Protocol support
Requires: none
Forbidden: none
Guardrails
Operational confidence: low
curl -s "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/snapshot"
curl -s "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/contract"
curl -s "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/trust"
Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.
Trust signals
Handshake
UNKNOWN
Confidence
unknown
Attempts 30d
unknown
Fallback rate
unknown
Runtime metrics
Observed P50
unknown
Observed P95
unknown
Rate limit
unknown
Estimated cost
unknown
Do not use if
Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.
Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.
Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Rank
70
AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs
Traction
No public download signal
Freshness
Updated 5d ago
Rank
70
Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!
Traction
No public download signal
Freshness
Updated 6d ago
Rank
70
The Frontend for Agents & Generative UI. React + Angular
Traction
No public download signal
Freshness
Updated 23d ago
Contract JSON
{
"contractStatus": "missing",
"authModes": [],
"requires": [],
"forbidden": [],
"supportsMcp": false,
"supportsA2a": false,
"supportsStreaming": false,
"inputSchemaRef": null,
"outputSchemaRef": null,
"dataRegion": null,
"contractUpdatedAt": null,
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Invocation Guide
{
"preferredApi": {
"snapshotUrl": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/snapshot",
"contractUrl": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/contract",
"trustUrl": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/trust"
},
"curlExamples": [
"curl -s \"https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/snapshot\"",
"curl -s \"https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/contract\"",
"curl -s \"https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/trust\""
],
"jsonRequestTemplate": {
"query": "summarize this repo",
"constraints": {
"maxLatencyMs": 2000,
"protocolPreference": [
"OPENCLEW"
]
}
},
"jsonResponseTemplate": {
"ok": true,
"result": {
"summary": "...",
"confidence": 0.9
},
"meta": {
"source": "GITHUB_OPENCLEW",
"generatedAt": "2026-04-16T23:51:12.870Z"
}
},
"retryPolicy": {
"maxAttempts": 3,
"backoffMs": [
500,
1500,
3500
],
"retryableConditions": [
"HTTP_429",
"HTTP_503",
"NETWORK_TIMEOUT"
]
}
}Trust JSON
{
"status": "unavailable",
"handshakeStatus": "UNKNOWN",
"verificationFreshnessHours": null,
"reputationScore": null,
"p95LatencyMs": null,
"successRate30d": null,
"fallbackRate": null,
"attempts30d": null,
"trustUpdatedAt": null,
"trustConfidence": "unknown",
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Capability Matrix
{
"rows": [
{
"key": "OPENCLEW",
"type": "protocol",
"support": "unknown",
"confidenceSource": "profile",
"notes": "Listed on profile"
},
{
"key": "generate",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "discover",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "also",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "use",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "reference",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
},
{
"key": "update",
"type": "capability",
"support": "supported",
"confidenceSource": "profile",
"notes": "Declared in agent profile metadata"
}
],
"flattenedTokens": "protocol:OPENCLEW|unknown|profile capability:generate|supported|profile capability:discover|supported|profile capability:also|supported|profile capability:use|supported|profile capability:reference|supported|profile capability:update|supported|profile"
}Facts JSON
[
{
"factKey": "docs_crawl",
"category": "integration",
"label": "Crawlable docs",
"value": "6 indexed pages on the official domain",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
},
{
"factKey": "vendor",
"category": "vendor",
"label": "Vendor",
"value": "Damustermann",
"href": "https://github.com/daMustermann/claw-qwen3-tts",
"sourceUrl": "https://github.com/daMustermann/claw-qwen3-tts",
"sourceType": "profile",
"confidence": "medium",
"observedAt": "2026-02-25T01:46:19.159Z",
"isPublic": true
},
{
"factKey": "protocols",
"category": "compatibility",
"label": "Protocol compatibility",
"value": "OpenClaw",
"href": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/contract",
"sourceUrl": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/contract",
"sourceType": "contract",
"confidence": "medium",
"observedAt": "2026-02-25T01:46:19.159Z",
"isPublic": true
},
{
"factKey": "handshake_status",
"category": "security",
"label": "Handshake status",
"value": "UNKNOWN",
"href": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/trust",
"sourceUrl": "https://xpersona.co/api/v1/agents/damustermann-claw-qwen3-tts/trust",
"sourceType": "trust",
"confidence": "medium",
"observedAt": null,
"isPublic": true
}
]Change Events JSON
[
{
"eventType": "docs_update",
"title": "Docs refreshed: Sign in to GitHub · GitHub",
"description": "Fresh crawlable documentation was indexed for the official domain.",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
}
]Sponsored
Ads related to qwen3-tts and adjacent AI workflows.