Claim this agent
Agent DossierCLAWHUBSafety 84/100

Xpersona Agent

Tech News Digest

Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Skill: Tech News Digest Owner: dinstein Summary: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Tags: latest:3.11.0 Version history: v3.11.0 | 2026-02-28T16:22:08.196Z | user Tavily backend, quality scores, domain limit fix, tests, CI v3.10.3 | 2026-02-27T15:46:08.147Z | user Docs alignment + config namin

OpenClaw · self-declared
2K downloadsTrust evidence available
clawhub skill install kn74589cx1nbhnc3x0f3nwre39814699:tech-news-digest

Overall rank

#62

Adoption

2K downloads

Trust

Unknown

Freshness

Feb 28, 2026

Freshness

Last checked Feb 28, 2026

Best For

Tech News Digest is best for general automation workflows where OpenClaw compatibility matters.

Not Ideal For

Contract metadata is missing or unavailable for deterministic execution.

Evidence Sources Checked

editorial-content, CLAWHUB, runtime-metrics, public facts pack

Overview

Key links, install path, reliability highlights, and the shortest practical read before diving into the crawl record.

Verifiededitorial-content

Overview

Executive Summary

Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Skill: Tech News Digest Owner: dinstein Summary: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Tags: latest:3.11.0 Version history: v3.11.0 | 2026-02-28T16:22:08.196Z | user Tavily backend, quality scores, domain limit fix, tests, CI v3.10.3 | 2026-02-27T15:46:08.147Z | user Docs alignment + config namin Capability contract not published. No trust telemetry is available yet. 2K downloads reported by the source. Last updated 4/15/2026.

No verified compatibility signals2K downloads

Trust score

Unknown

Compatibility

OpenClaw

Freshness

Feb 28, 2026

Vendor

Clawhub

Artifacts

0

Benchmarks

0

Last release

3.11.0

Install & run

Setup Snapshot

clawhub skill install kn74589cx1nbhnc3x0f3nwre39814699:tech-news-digest
  1. 1

    Setup complexity is classified as HIGH. You must provision dedicated cloud infrastructure or an isolated VM. Do not run this directly on your local workstation.

  2. 2

    Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.

Evidence & Timeline

Public facts grouped by evidence type, plus release and crawl events with provenance and freshness.

Verifiededitorial-content

Public facts

Evidence Ledger

Vendor (1)

Vendor

Clawhub

profilemedium
Observed Apr 15, 2026Source linkProvenance
Compatibility (1)

Protocol compatibility

OpenClaw

contractmedium
Observed Apr 15, 2026Source linkProvenance
Release (1)

Latest release

3.11.0

releasemedium
Observed Feb 28, 2026Source linkProvenance
Adoption (1)

Adoption signal

2K downloads

profilemedium
Observed Apr 15, 2026Source linkProvenance
Security (1)

Handshake status

UNKNOWN

trustmedium
Observed unknownSource linkProvenance

Artifacts & Docs

Parameters, dependencies, examples, extracted files, editorial overview, and the complete README when available.

Self-declaredCLAWHUB

Captured outputs

Artifacts Archive

Extracted files

5

Examples

6

Snippets

0

Languages

Unknown

Executable Examples

bash

mkdir -p workspace/config
   cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
   cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json

bash

# Unified pipeline (recommended) — runs all 5 sources in parallel + merge
   python3 scripts/run-pipeline.py \
     --defaults config/defaults \
     --config workspace/config \
     --hours 48 --freshness pd \
     --archive-dir workspace/archive/tech-news-digest/ \
     --output /tmp/td-merged.json --verbose --force

json

{
  "sources": [
    {
      "id": "openai-rss",
      "type": "rss",
      "name": "OpenAI Blog",
      "url": "https://openai.com/blog/rss.xml",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "ai-agent"],
      "note": "Official OpenAI updates"
    },
    {
      "id": "sama-twitter",
      "type": "twitter", 
      "name": "Sam Altman",
      "handle": "sama",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "frontier-tech"],
      "note": "OpenAI CEO"
    }
  ]
}

json

{
  "topics": [
    {
      "id": "llm",
      "emoji": "🧠",
      "label": "LLM / Large Models",
      "description": "Large Language Models, foundation models, breakthroughs",
      "search": {
        "queries": ["LLM latest news", "large language model breakthroughs"],
        "must_include": ["LLM", "large language model", "foundation model"],
        "exclude": ["tutorial", "beginner guide"]
      },
      "display": {
        "max_items": 8,
        "style": "detailed"
      }
    }
  ]
}

bash

python3 scripts/run-pipeline.py \
  --defaults config/defaults [--config CONFIG_DIR] \
  --hours 48 --freshness pd \
  --archive-dir workspace/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force

bash

python3 scripts/fetch-rss.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--verbose]
Extracted Files

SKILL.md

---
name: tech-news-digest
description: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, GitHub releases, Reddit, and web search. Pipeline-based scripts with retry mechanisms and deduplication. Supports Discord, email, and markdown templates.
version: "3.11.0"
homepage: https://github.com/draco-agent/tech-news-digest
source: https://github.com/draco-agent/tech-news-digest
metadata:
  openclaw:
    requires:
      bins: ["python3"]
    optionalBins: ["mail", "msmtp", "gog", "gh", "openssl", "weasyprint"]
env:
  - name: TWITTER_API_BACKEND
    required: false
    description: "Twitter API backend: 'official', 'twitterapiio', or 'auto' (default: auto)"
  - name: X_BEARER_TOKEN
    required: false
    description: Twitter/X API bearer token for KOL monitoring (official backend)
  - name: TWITTERAPI_IO_KEY
    required: false
    description: twitterapi.io API key for KOL monitoring (twitterapiio backend)
  - name: TAVILY_API_KEY
    required: false
    description: Tavily Search API key (alternative to Brave)
  - name: WEB_SEARCH_BACKEND
    required: false
    description: "Web search backend: auto (default), brave, or tavily"
  - name: BRAVE_API_KEYS
    required: false
    description: Brave Search API keys (comma-separated for rotation)
  - name: BRAVE_API_KEY
    required: false
    description: Brave Search API key (single key fallback)
  - name: GITHUB_TOKEN
    required: false
    description: GitHub token for higher API rate limits (auto-generated from GitHub App if not set)
  - name: GH_APP_ID
    required: false
    description: GitHub App ID for automatic installation token generation
  - name: GH_APP_INSTALL_ID
    required: false
    description: GitHub App Installation ID for automatic token generation
  - name: GH_APP_KEY_FILE
    required: false
    description: Path to GitHub App private key PEM file
tools:
  - python3: Required. Runs data collection and merge scripts.
  - mail: Optional. msmtp-based mail command for email delivery (preferred).
  - gog: Optional. Gmail CLI for email delivery (fallback if mail not available).
files:
  read:
    - config/defaults/: Default source and topic configurations
    - references/: Prompt templates and output templates
    - scripts/: Python pipeline scripts
    - <workspace>/archive/tech-news-digest/: Previous digests for dedup
  write:
    - /tmp/td-*.json: Temporary pipeline intermediate outputs
    - /tmp/td-email.html: Temporary email HTML body
    - /tmp/td-digest.pdf: Generated PDF digest
    - <workspace>/archive/tech-news-digest/: Saved digest archives
---

# Tech News Digest

Automated tech news digest system with unified data source model, quality scoring pipeline, and template-based output generation.

## Quick Start

1. **Configuration Setup**: Default configs are in `config/defaults/`. Copy to workspace for customization:
   ```bash
   mkdir -p workspace/config
   cp config/d

README.md

# Tech News Digest

> Automated tech news digest — 138 sources, 5-layer pipeline, one chat message to install.

**English** | [中文](README_CN.md)

[![Tests](https://github.com/draco-agent/tech-news-digest/actions/workflows/test.yml/badge.svg)](https://github.com/draco-agent/tech-news-digest/actions/workflows/test.yml)
[![Python 3.8+](https://img.shields.io/badge/python-3.8+-blue.svg)](https://www.python.org/downloads/)
[![ClawHub](https://img.shields.io/badge/ClawHub-tech--news--digest-blueviolet)](https://clawhub.com/draco-agent/tech-news-digest)
[![MIT License](https://img.shields.io/badge/license-MIT-green.svg)](LICENSE)

## 💬 Install in One Message

Tell your [OpenClaw](https://openclaw.ai) AI assistant:

> **"Install tech-news-digest and send a daily digest to #tech-news every morning at 9am"**

That's it. Your bot handles installation, configuration, scheduling, and delivery — all through conversation.

More examples:

> 🗣️ "Set up a weekly AI digest, only LLM and AI Agent topics, deliver to Discord #ai-weekly every Monday"

> 🗣️ "Install tech-news-digest, add my RSS feeds, and send crypto news to Telegram"

> 🗣️ "Give me a tech digest right now, skip Twitter sources"

Or install via CLI:
```bash
clawhub install tech-news-digest
```

## 📊 What You Get

A quality-scored, deduplicated tech digest built from **138 sources**:

| Layer | Sources | What |
|-------|---------|------|
| 📡 RSS | 49 feeds | OpenAI, Anthropic, Ben's Bites, HN, 36氪, CoinDesk… |
| 🐦 Twitter/X | 48 KOLs | @karpathy, @VitalikButerin, @sama, @elonmusk… |
| 🔍 Web Search | 4 topics | Brave Search API with freshness filters |
| 🐙 GitHub | 28 repos | Releases from key projects (LangChain, vLLM, DeepSeek, Llama…) |
| 🗣️ Reddit | 13 subs | r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency… |

### Pipeline

```
       run-pipeline.py (~30s)
              ↓
  RSS ─┐
  Twitter ─┤
  Web ─────┤── parallel fetch ──→ merge-sources.py
  GitHub ──┤
  Reddit ──┘
              ↓
  Quality Scoring → Deduplication → Topic Grouping
              ↓
    Discord / Email / PDF output
```

**Quality scoring**: priority source (+3), multi-source cross-ref (+5), recency (+2), engagement (+1), Reddit score bonus (+1/+3/+5), already reported (-5).

## ⚙️ Configuration

- `config/defaults/sources.json` — 138 built-in sources
- `config/defaults/topics.json` — 4 topics with search queries & Twitter queries
- User overrides in `workspace/config/` take priority

## 🎨 Customize Your Sources

Works out of the box with 138 built-in sources — but fully customizable. Copy the defaults to your workspace config and override:

```bash
# Copy and customize
cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json
```

Your overlay file **merges** with defaults:
- **Override** a source by matching its `id` — your version replaces the default
- **Add** new sources with a unique `id` — appended to the list
- **Di

_meta.json

{
  "ownerId": "kn74589cx1nbhnc3x0f3nwre39814699",
  "slug": "tech-news-digest",
  "version": "3.11.0",
  "publishedAt": 1772295728196
}

references/digest-prompt.md

# Digest Prompt Template

Replace `<...>` placeholders before use. Daily defaults shown; weekly overrides in parentheses.

## Placeholders

| Placeholder | Default | Weekly Override |
|-------------|---------|----------------|
| `<MODE>` | `daily` | `weekly` |
| `<TIME_WINDOW>` | `past 1-2 days` | `past 7 days` |
| `<FRESHNESS>` | `pd` | `pw` |
| `<RSS_HOURS>` | `48` | `168` |
| `<ITEMS_PER_SECTION>` | `3-5` | `5-8` |
| `<BLOG_PICKS_COUNT>` | `2-3` | `3-5` |
| `<EXTRA_SECTIONS>` | *(none)* | `📊 Weekly Trend Summary` |
| `<SUBJECT>` | `Daily Tech Digest - YYYY-MM-DD` | `Weekly Tech Digest - YYYY-MM-DD` |
| `<WORKSPACE>` | Your workspace path | |
| `<SKILL_DIR>` | Installed skill directory | |
| `<DISCORD_CHANNEL_ID>` | Target channel ID | |
| `<EMAIL>` | *(optional)* Recipient email | |
| `<EMAIL_FROM>` | *(optional)* e.g. `MyBot <bot@example.com>` | |
| `<LANGUAGE>` | `Chinese` | |
| `<TEMPLATE>` | `discord` / `email` / `markdown` | |
| `<DATE>` | Today's date YYYY-MM-DD (caller provides) | |
| `<VERSION>` | Read from SKILL.md frontmatter | |

---

Generate the <MODE> tech digest for **<DATE>**. Use `<DATE>` as the report date — do NOT infer it.

## Configuration

Read config files (workspace overrides take priority over defaults):
1. **Sources**: `<WORKSPACE>/config/tech-news-digest-sources.json` → fallback `<SKILL_DIR>/config/defaults/sources.json`
2. **Topics**: `<WORKSPACE>/config/tech-news-digest-topics.json` → fallback `<SKILL_DIR>/config/defaults/topics.json`

## Context: Previous Report

Read the most recent file from `<WORKSPACE>/archive/tech-news-digest/` to avoid repeats and follow up on developing stories. Skip if none exists.

## Data Collection Pipeline

**Use the unified pipeline** (runs all 5 sources in parallel, ~30s):

```bash
python3 <SKILL_DIR>/scripts/run-pipeline.py \
  --defaults <SKILL_DIR>/config/defaults \
  --config <WORKSPACE>/config \
  --hours <RSS_HOURS> --freshness <FRESHNESS> \
  --archive-dir <WORKSPACE>/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force
```

If it fails, run individual scripts in `<SKILL_DIR>/scripts/` (see each script's `--help`), then merge with `merge-sources.py`.

## Report Generation

Get a structured overview:
```bash
python3 <SKILL_DIR>/scripts/summarize-merged.py --input /tmp/td-merged.json --top <ITEMS_PER_SECTION>
```

Use this output to select articles — **do NOT write ad-hoc Python to parse the JSON**. Apply the template from `<SKILL_DIR>/references/templates/<TEMPLATE>.md`.

Select articles **purely by quality_score regardless of source type**. Articles in merged JSON are already sorted by quality_score descending within each topic — respect this order. For Reddit posts, append `*[Reddit r/xxx, {{score}}↑]*`.

Each article line must include its quality score using 🔥 prefix. Format: `🔥{score} | {summary with link}`. This makes scoring transparent and helps readers identify the most important news at a glance.

### Executive Summary
2-4 sentences between t

references/templates/discord.md

# Tech Digest Discord Template

Discord-optimized format with bullet points and link suppression.

## Template Structure

```markdown
# 🚀 Tech Digest - {{DATE}}

{{#topics}}
## {{emoji}} {{label}}

{{#articles}}
• 🔥{{quality_score}} | {{title}}
  <{{link}}>
  {{#multi_source}}*[{{source_count}} sources]*{{/multi_source}}

{{/articles}}
{{/topics}}

---
📊 Data Sources: RSS {{rss_count}} | Twitter {{twitter_count}} | Reddit {{reddit_count}} | Web {{web_count}} | GitHub {{github_count}} releases | After dedup: {{merged_count}} articles
🤖 Generated by tech-news-digest v{{version}} | <https://github.com/draco-agent/tech-news-digest> | Powered by OpenClaw
```

## Delivery

- **Default: Channel** — Send to the Discord channel specified by `DISCORD_CHANNEL_ID`
- Use `message` tool with `target` set to the channel ID for channel delivery
- For DM delivery instead, set `target` to a user ID

## Discord-Specific Features

- **Link suppression**: Wrap links in `<>` to prevent embeds
- **Bullet format**: Use `•` for clean mobile display  
- **No tables**: Discord mobile doesn't handle markdown tables well
- **Emoji headers**: Visual hierarchy with topic emojis
- **Concise metadata**: Source count and multi-source indicators
- **Character limits**: Discord messages have 2000 char limit, may need splitting

## Example Output

```markdown
# 🚀 Tech Digest - 2026-02-15

## 🧠 LLM / Large Models

• 🔥15 | OpenAI releases GPT-5 with breakthrough reasoning capabilities
  <https://openai.com/blog/gpt5-announcement>
  *[3 sources]*

• 🔥12 | Meta's Llama 3.1 achieves new MMLU benchmarks
  <https://ai.meta.com/blog/llama-31-release>

## 🤖 AI Agent

• 🔥14 | LangChain launches production-ready agent framework
  <https://blog.langchain.dev/production-agents>

## 💰 Cryptocurrency

• 🔥18 | Bitcoin reaches new ATH at $67,000 amid ETF approval
  <https://coindesk.com/markets/btc-ath-etf>
  *[2 sources]*

## 📢 KOL Updates

• **Elon Musk** (@elonmusk) — Confirmed X's crypto trading feature `👁 2.1M | 💬 12.3K | 🔁 8.5K | ❤️ 49.8K`
  <https://twitter.com/elonmusk/status/123456789>
• **@saylor** — Valentine's BTC enthusiasm `👁 450K | 💬 1.2K | 🔁 3.1K | ❤️ 13K`
  <https://twitter.com/saylor/status/987654321>

---
📊 Data Sources: RSS 285 | Twitter 67 | Reddit 45 | Web 60 | GitHub 29 releases | After dedup: 95 articles
```

## Variables

- `{{DATE}}` - Report date (YYYY-MM-DD format)
- `{{topics}}` - Array of topic objects
- `{{emoji}}` - Topic emoji 
- `{{label}}` - Topic display name
- `{{articles}}` - Array of article objects per topic
- `{{title}}` - Article title (truncated if needed)
- `{{link}}` - Article URL
- `{{quality_score}}` - Article quality score (higher = more important)
- `{{multi_source}}` - Boolean, true if article from multiple sources
- `{{source_count}}` - Number of sources for this article
- `{{total_sources}}` - Total number of sources used
- `{{total_articles}}` - Total articles in digest

Editorial read

Docs & README

Docs source

CLAWHUB

Editorial quality

ready

Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Skill: Tech News Digest Owner: dinstein Summary: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi... Tags: latest:3.11.0 Version history: v3.11.0 | 2026-02-28T16:22:08.196Z | user Tavily backend, quality scores, domain limit fix, tests, CI v3.10.3 | 2026-02-27T15:46:08.147Z | user Docs alignment + config namin

Full README

Skill: Tech News Digest

Owner: dinstein

Summary: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, Gi...

Tags: latest:3.11.0

Version history:

v3.11.0 | 2026-02-28T16:22:08.196Z | user

Tavily backend, quality scores, domain limit fix, tests, CI

v3.10.3 | 2026-02-27T15:46:08.147Z | user

Docs alignment + config naming refactor

v3.10.2 | 2026-02-27T12:01:25.639Z | user

Fix domain limits for multi-author platforms + Brave multi-key rotation

v3.6.2 | 2026-02-22T06:23:22.415Z | user

Add 3 GitHub sources: cloudflare/moltworker, sipeed/picoclaw, HKUDS/nanobot

v3.6.1 | 2026-02-22T06:17:37.818Z | user

v3.6.1: prompt review pass

v3.6.0 | 2026-02-21T06:57:08.561Z | user

v3.6.0

v3.5.1 | 2026-02-21T06:28:06.959Z | user

v3.5.1

v3.5.0 | 2026-02-18T14:47:10.890Z | user

P1 fixes: per-run cache isolation, per-topic domain limits, fix 6 Twitter handles, zsh compat

v3.4.11 | 2026-02-18T13:48:07.508Z | user

Fix 6 Twitter handles, zsh compat for test-pipeline, github token debug logs

v3.4.10 | 2026-02-18T00:46:12.449Z | user

Republish

v3.4.9 | 2026-02-17T15:37:13.956Z | auto

Version 3.4.9

  • Added summarize-merged.py script for output summarization.
  • Updated SKILL.md to include "openssl" as an optional dependency.
  • Documentation improvements and updates to changelog and reference prompts.

v3.4.7 | 2026-02-17T14:56:10.812Z | auto

Version 3.4.7

  • Updated fetch-github.py script.
  • Documentation improvements in SKILL.md, including clarified template and output details.
  • Version bump and changelog maintenance.

v3.4.6 | 2026-02-17T14:08:50.097Z | auto

  • Improved security note for GH_APP_TOKEN_SCRIPT: now warns users that only trusted scripts should be referenced, as the script will be executed via subprocess.
  • Updated all documentation and schema references to clarify workspace and archive directory naming (now references 'tech-news-digest' instead of 'tech-digest').
  • Corrected environment variable documentation for generating GitHub App tokens.
  • Minor documentation and schema consistency updates across config, template, and reference files for clarity and safety.

v3.4.5 | 2026-02-17T10:47:05.936Z | auto

  • Improved reliability and internal consistency across all data collection scripts.
  • Updated documentation in SKILL.md, README.md, and CHANGELOG.md for recent enhancements.
  • Refactored source configuration loader and validation for clearer error handling.
  • Enhanced logging and status reporting in fetch and merge scripts.
  • Minor bug fixes and code clean-up.

v3.4.4 | 2026-02-17T10:35:49.853Z | auto

tech-news-digest 3.4.4

  • Updated version to 3.4.4 in documentation.
  • No changes to code or behavior; documentation/version consistency update only.

v3.4.3 | 2026-02-17T09:41:09.209Z | auto

Version 3.4.3

  • Added "gh" to the list of optional CLI tools in the requirements.
  • No functional or pipeline changes; documentation and metadata updated.

v3.4.2 | 2026-02-17T09:39:10.447Z | auto

  • Added new environment variables for GitHub App authentication: GH_APP_ID, GH_APP_INSTALL_ID, GH_APP_KEY_FILE, and GH_APP_TOKEN_SCRIPT.
  • Updated documentation to reflect additional GitHub App token generation options.
  • Improved GitHub token generation logic in scripts/fetch-github.py to support new environment variables for automatic installation token creation.

v3.4.1 | 2026-02-17T07:00:20.427Z | auto

  • Improved merge-sources.py script to handle additional merging logic for topic and source overrides.
  • Updated default sources.json with new tech sources and refined topic mappings.
  • Enhanced Discord, email, and Telegram template files for clearer digest formatting.
  • Minor documentation fixes and clarifications in SKILL.md and prompts.
  • Improved configuration validation and documentation for workspace overrides.

v3.4.0 | 2026-02-17T01:44:28.875Z | auto

tech-news-digest 3.4.0 introduces a unified pipeline script and improved programmatic control.

  • Added scripts/run-pipeline.py: a new single-entry script running all collectors in parallel and merging results.
  • Updated all fetch scripts to accept unified --defaults and --config arguments.
  • Auth improvements: GitHub token now auto-generates from a GitHub App if $GITHUB_TOKEN is unset.
  • Enhanced configuration flexibility and environment variable descriptions in documentation.
  • README and SKILL.md revised to highlight the unified workflow and clarify recommended usage.

v3.3.2 | 2026-02-16T16:00:33.444Z | user

Declare tools (python3, gog) and file access paths in metadata; addresses VT undeclared tools audit

v3.3.1 | 2026-02-16T15:53:58.637Z | user

Remove anthropic-rss third-party mirror (supply chain risk); update source count to 131

v3.2.0 | 2026-02-16T03:30:14.448Z | auto

Version 3.2.0

  • Updated and improved output templates for Discord, email, Markdown, and Telegram.
  • Enhanced reference prompt and documentation in reference files.
  • Minor documentation and changelog updates for clarity and accuracy.
  • No changes to core pipeline logic or configuration schema.

v3.1.0 | 2026-02-16T03:23:00.182Z | auto

  • Added detailed inline documentation and usage examples for all supported output templates (Discord, email, Telegram) and digest prompts.
  • Improved template files for Discord, email, and Telegram to clarify formatting and message structure.
  • Updated references and prompt documentation to streamline multi-format output customization.
  • Expanded SKILL.md with clearer template usage instructions and examples.
  • General content and documentation enhancements for usability and clarity.

v3.0.0 | 2026-02-16T03:02:23.033Z | auto

Major update: Reddit support and expanded source coverage.

  • Added Reddit as a new data layer with fetch-reddit.py, enabling five-layer news aggregation.
  • Increased default sources to 132, including 13 major Reddit communities.
  • Updated config and pipeline scripts to integrate Reddit alongside RSS, Twitter/X, GitHub, and web search.
  • Enhanced SKILL.md, documentation, and example configs to reflect the new Reddit integration.
  • Improved source model and topic tagging for broader and more customizable tech news coverage.

v2.8.1 | 2026-02-16T01:55:39.092Z | auto

  • Refreshed digest templates for Discord, email, and Telegram with improved formatting and metadata.
  • Updated default source list and topic definitions for better coverage and tag accuracy.
  • Revised documentation in SKILL.md for clarity and up-to-date usage instructions.
  • Enhanced prompt guidance in references/digest-prompt.md.
  • Minor improvements to default config (sources.json), template usability, and config instructions.

v2.7.0 | 2026-02-16T00:36:45.486Z | auto

tech-news-digest v2.7.0

  • Updated prompt and template files for improved digest formatting and clarity.
  • Refined Discord, email, and Telegram output templates for more consistent multi-platform presentation.
  • Enhanced reference documentation for prompt design and template usage.
  • Minor content and metadata adjustments in skill documentation.

v2.6.1 | 2026-02-16T00:09:56.542Z | auto

  • Improved Twitter/X KOL collection: Now skips over empty user handles for better data integrity.
  • Minor update to documentation in SKILL.md.
  • Patch version bump to 2.6.1.

v2.6.0 | 2026-02-16T00:04:39.583Z | auto

  • Adds a new config option to set both defaults and workspace paths in the config validator (validate-config.py).
  • Updates validate-config.py argument interface to --defaults / --config for better clarity (was formerly --config-dir).
  • Updates documentation and usage examples in SKILL.md for consistent interface references.
  • Enhances configuration validation flexibility and accuracy.
  • Minor documentation cleanup and consistency fixes across references and scripts.

v2.5.0 | 2026-02-15T16:46:14.680Z | user

Fix twitter reply filter, scoring consistency, template versions, article count, pipeline resume, source health monitoring, e2e test, archive cleanup, rate limiting, web scoring, dead code removal

v2.4.1 | 2026-02-15T16:02:18.260Z | user

Remove version from SKILL.md title, move changelog to CHANGELOG.md

v2.4.0 | 2026-02-15T15:55:35.351Z | user

Performance: batch Twitter lookup, smart dedup, conditional fetch for RSS/GitHub

v2.2.1 | 2026-02-15T14:36:27.427Z | user

Security hardening: shell injection prevention for email delivery, added Security Notes section documenting execution model, network access, file access scope

v2.2.0 | 2026-02-15T12:17:30.123Z | user

Renamed from tech-digest. 109 sources, 4-layer pipeline (RSS/Twitter/Web/GitHub), quality scoring, multi-template output, cron prompt pattern.

Archive index:

Archive v3.11.0: 37 files, 113229 bytes

Files: CHANGELOG.md (17649b), config/defaults/sources.json (38686b), config/defaults/topics.json (2951b), config/schema.json (4561b), CONTRIBUTING.md (2283b), README_CN.md (4592b), README.md (4583b), references/digest-prompt.md (5933b), references/templates/discord.md (3011b), references/templates/email.md (5805b), references/templates/pdf.md (2119b), requirements.txt (606b), scripts/config_loader.py (8490b), scripts/fetch-github.py (20476b), scripts/fetch-reddit.py (12211b), scripts/fetch-rss.py (19352b), scripts/fetch-twitter.py (28449b), scripts/fetch-web.py (26778b), scripts/generate-pdf.py (9751b), scripts/merge-sources.py (24247b), scripts/run-pipeline.py (9960b), scripts/sanitize-html.py (7357b), scripts/send-email.py (5199b), scripts/source-health.py (4931b), scripts/summarize-merged.py (3526b), scripts/test-pipeline.sh (10975b), scripts/validate-config.py (9659b), SKILL.md (21091b), tests/fixtures/github.json (3488b), tests/fixtures/merged.json (13075b), tests/fixtures/reddit.json (6097b), tests/fixtures/rss.json (3374b), tests/fixtures/twitter.json (5039b), tests/fixtures/web.json (4138b), tests/test_config.py (4574b), tests/test_merge.py (9779b), _meta.json (136b)

File v3.11.0:SKILL.md


name: tech-news-digest description: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, GitHub releases, Reddit, and web search. Pipeline-based scripts with retry mechanisms and deduplication. Supports Discord, email, and markdown templates. version: "3.11.0" homepage: https://github.com/draco-agent/tech-news-digest source: https://github.com/draco-agent/tech-news-digest metadata: openclaw: requires: bins: ["python3"] optionalBins: ["mail", "msmtp", "gog", "gh", "openssl", "weasyprint"] env:

  • name: TWITTER_API_BACKEND required: false description: "Twitter API backend: 'official', 'twitterapiio', or 'auto' (default: auto)"
  • name: X_BEARER_TOKEN required: false description: Twitter/X API bearer token for KOL monitoring (official backend)
  • name: TWITTERAPI_IO_KEY required: false description: twitterapi.io API key for KOL monitoring (twitterapiio backend)
  • name: TAVILY_API_KEY required: false description: Tavily Search API key (alternative to Brave)
  • name: WEB_SEARCH_BACKEND required: false description: "Web search backend: auto (default), brave, or tavily"
  • name: BRAVE_API_KEYS required: false description: Brave Search API keys (comma-separated for rotation)
  • name: BRAVE_API_KEY required: false description: Brave Search API key (single key fallback)
  • name: GITHUB_TOKEN required: false description: GitHub token for higher API rate limits (auto-generated from GitHub App if not set)
  • name: GH_APP_ID required: false description: GitHub App ID for automatic installation token generation
  • name: GH_APP_INSTALL_ID required: false description: GitHub App Installation ID for automatic token generation
  • name: GH_APP_KEY_FILE required: false description: Path to GitHub App private key PEM file tools:
  • python3: Required. Runs data collection and merge scripts.
  • mail: Optional. msmtp-based mail command for email delivery (preferred).
  • gog: Optional. Gmail CLI for email delivery (fallback if mail not available). files: read:
    • config/defaults/: Default source and topic configurations
    • references/: Prompt templates and output templates
    • scripts/: Python pipeline scripts
    • <workspace>/archive/tech-news-digest/: Previous digests for dedup write:
    • /tmp/td-*.json: Temporary pipeline intermediate outputs
    • /tmp/td-email.html: Temporary email HTML body
    • /tmp/td-digest.pdf: Generated PDF digest
    • <workspace>/archive/tech-news-digest/: Saved digest archives

Tech News Digest

Automated tech news digest system with unified data source model, quality scoring pipeline, and template-based output generation.

Quick Start

  1. Configuration Setup: Default configs are in config/defaults/. Copy to workspace for customization:

    mkdir -p workspace/config
    cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
    cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json
    
  2. Environment Variables:

    • TWITTERAPI_IO_KEY - twitterapi.io API key (optional, preferred)
    • X_BEARER_TOKEN - Twitter/X official API bearer token (optional, fallback)
    • TAVILY_API_KEY - Tavily Search API key, alternative to Brave (optional)
    • WEB_SEARCH_BACKEND - Web search backend: auto|brave|tavily (optional, default: auto)
    • BRAVE_API_KEYS - Brave Search API keys, comma-separated for rotation (optional)
    • BRAVE_API_KEY - Single Brave key fallback (optional)
    • GITHUB_TOKEN - GitHub personal access token (optional, improves rate limits)
  3. Generate Digest:

    # Unified pipeline (recommended) — runs all 5 sources in parallel + merge
    python3 scripts/run-pipeline.py \
      --defaults config/defaults \
      --config workspace/config \
      --hours 48 --freshness pd \
      --archive-dir workspace/archive/tech-news-digest/ \
      --output /tmp/td-merged.json --verbose --force
    
  4. Use Templates: Apply Discord, email, or PDF templates to merged output

Configuration Files

sources.json - Unified Data Sources

{
  "sources": [
    {
      "id": "openai-rss",
      "type": "rss",
      "name": "OpenAI Blog",
      "url": "https://openai.com/blog/rss.xml",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "ai-agent"],
      "note": "Official OpenAI updates"
    },
    {
      "id": "sama-twitter",
      "type": "twitter", 
      "name": "Sam Altman",
      "handle": "sama",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "frontier-tech"],
      "note": "OpenAI CEO"
    }
  ]
}

topics.json - Enhanced Topic Definitions

{
  "topics": [
    {
      "id": "llm",
      "emoji": "🧠",
      "label": "LLM / Large Models",
      "description": "Large Language Models, foundation models, breakthroughs",
      "search": {
        "queries": ["LLM latest news", "large language model breakthroughs"],
        "must_include": ["LLM", "large language model", "foundation model"],
        "exclude": ["tutorial", "beginner guide"]
      },
      "display": {
        "max_items": 8,
        "style": "detailed"
      }
    }
  ]
}

Scripts Pipeline

run-pipeline.py - Unified Pipeline (Recommended)

python3 scripts/run-pipeline.py \
  --defaults config/defaults [--config CONFIG_DIR] \
  --hours 48 --freshness pd \
  --archive-dir workspace/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force
  • Features: Runs all 5 fetch steps in parallel, then merges + deduplicates + scores
  • Output: Final merged JSON ready for report generation (~30s total)
  • Metadata: Saves per-step timing and counts to *.meta.json
  • GitHub Auth: Auto-generates GitHub App token if $GITHUB_TOKEN not set
  • Fallback: If this fails, run individual scripts below

Individual Scripts (Fallback)

fetch-rss.py - RSS Feed Fetcher

python3 scripts/fetch-rss.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--verbose]
  • Parallel fetching (10 workers), retry with backoff, feedparser + regex fallback
  • Timeout: 30s per feed, ETag/Last-Modified caching

fetch-twitter.py - Twitter/X KOL Monitor

python3 scripts/fetch-twitter.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--backend auto|official|twitterapiio]
  • Backend auto-detection: uses twitterapi.io if TWITTERAPI_IO_KEY set, else official X API v2 if X_BEARER_TOKEN set
  • Rate limit handling, engagement metrics, retry with backoff

fetch-web.py - Web Search Engine

python3 scripts/fetch-web.py [--defaults DIR] [--config DIR] [--freshness pd] [--output FILE]
  • Auto-detects Brave API rate limit: paid plans → parallel queries, free → sequential
  • Without API: generates search interface for agents

fetch-github.py - GitHub Releases Monitor

python3 scripts/fetch-github.py [--defaults DIR] [--config DIR] [--hours 168] [--output FILE]
  • Parallel fetching (10 workers), 30s timeout
  • Auth priority: $GITHUB_TOKEN → GitHub App auto-generate → gh CLI → unauthenticated (60 req/hr)

fetch-reddit.py - Reddit Posts Fetcher

python3 scripts/fetch-reddit.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE]
  • Parallel fetching (4 workers), public JSON API (no auth required)
  • 13 subreddits with score filtering

merge-sources.py - Quality Scoring & Deduplication

python3 scripts/merge-sources.py --rss FILE --twitter FILE --web FILE --github FILE --reddit FILE
  • Quality scoring, title similarity dedup (85%), previous digest penalty
  • Output: topic-grouped articles sorted by score

validate-config.py - Configuration Validator

python3 scripts/validate-config.py [--defaults DIR] [--config DIR] [--verbose]
  • JSON schema validation, topic reference checks, duplicate ID detection

generate-pdf.py - PDF Report Generator

python3 scripts/generate-pdf.py --input report.md --output digest.pdf [--verbose]
  • Converts markdown digest to styled A4 PDF with Chinese typography (Noto Sans CJK SC)
  • Emoji icons, page headers/footers, blue accent theme. Requires weasyprint.

sanitize-html.py - Safe HTML Email Converter

python3 scripts/sanitize-html.py --input report.md --output email.html [--verbose]
  • Converts markdown to XSS-safe HTML email with inline CSS
  • URL whitelist (http/https only), HTML-escaped text content

source-health.py - Source Health Monitor

python3 scripts/source-health.py --rss FILE --twitter FILE --github FILE --reddit FILE --web FILE [--verbose]
  • Tracks per-source success/failure history over 7 days
  • Reports unhealthy sources (>50% failure rate)

summarize-merged.py - Merged Data Summary

python3 scripts/summarize-merged.py --input merged.json [--top N] [--topic TOPIC]
  • Human-readable summary of merged data for LLM consumption
  • Shows top articles per topic with scores and metrics

User Customization

Workspace Configuration Override

Place custom configs in workspace/config/ to override defaults:

  • Sources: Append new sources, disable defaults with "enabled": false
  • Topics: Override topic definitions, search queries, display settings
  • Merge Logic:
    • Sources with same id → user version takes precedence
    • Sources with new id → appended to defaults
    • Topics with same id → user version completely replaces default

Example Workspace Override

// workspace/config/tech-news-digest-sources.json
{
  "sources": [
    {
      "id": "simonwillison-rss",
      "enabled": false,
      "note": "Disabled: too noisy for my use case"
    },
    {
      "id": "my-custom-blog", 
      "type": "rss",
      "name": "My Custom Tech Blog",
      "url": "https://myblog.com/rss",
      "enabled": true,
      "priority": true,
      "topics": ["frontier-tech"]
    }
  ]
}

Templates & Output

Discord Template (references/templates/discord.md)

  • Bullet list format with link suppression (<link>)
  • Mobile-optimized, emoji headers
  • 2000 character limit awareness

Email Template (references/templates/email.md)

  • Rich metadata, technical stats, archive links
  • Executive summary, top articles section
  • HTML-compatible formatting

PDF Template (references/templates/pdf.md)

  • A4 layout with Noto Sans CJK SC font for Chinese support
  • Emoji icons, page headers/footers with page numbers
  • Generated via scripts/generate-pdf.py (requires weasyprint)

Default Sources (138 total)

  • RSS Feeds (49): AI labs, tech blogs, crypto news, Chinese tech media
  • Twitter/X KOLs (48): AI researchers, crypto leaders, tech executives
  • GitHub Repos (28): Major open-source projects (LangChain, vLLM, DeepSeek, Llama, etc.)
  • Reddit (13): r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency, r/ChatGPT, r/OpenAI, etc.
  • Web Search (4 topics): LLM, AI Agent, Crypto, Frontier Tech

All sources pre-configured with appropriate topic tags and priority levels.

Dependencies

pip install -r requirements.txt

Optional but Recommended:

  • feedparser>=6.0.0 - Better RSS parsing (fallback to regex if unavailable)
  • jsonschema>=4.0.0 - Configuration validation

All scripts work with Python 3.8+ standard library only.

Monitoring & Operations

Health Checks

# Validate configuration
python3 scripts/validate-config.py --verbose

# Test RSS feeds
python3 scripts/fetch-rss.py --hours 1 --verbose

# Check Twitter API
python3 scripts/fetch-twitter.py --hours 1 --verbose

Archive Management

  • Digests automatically archived to <workspace>/archive/tech-news-digest/
  • Previous digest titles used for duplicate detection
  • Old archives cleaned automatically (90+ days)

Error Handling

  • Network Failures: Retry with exponential backoff
  • Rate Limits: Automatic retry with appropriate delays
  • Invalid Content: Graceful degradation, detailed logging
  • Configuration Errors: Schema validation with helpful messages

API Keys & Environment

Set in ~/.zshenv or similar:

# Twitter (at least one required for Twitter source)
export TWITTERAPI_IO_KEY="your_key"        # twitterapi.io key (preferred)
export X_BEARER_TOKEN="your_bearer_token"  # Official X API v2 (fallback)
export TWITTER_API_BACKEND="auto"          # auto|twitterapiio|official (default: auto)

# Web Search (optional, enables web search layer)
export WEB_SEARCH_BACKEND="auto"          # auto|brave|tavily (default: auto)
export TAVILY_API_KEY="tvly-xxx"           # Tavily Search API (free 1000/mo)

# Brave Search (alternative)
export BRAVE_API_KEYS="key1,key2,key3"     # Multiple keys, comma-separated rotation
export BRAVE_API_KEY="key1"                # Single key fallback
export BRAVE_PLAN="free"                   # Override rate limit detection: free|pro

# GitHub (optional, improves rate limits)
export GITHUB_TOKEN="ghp_xxx"              # PAT (simplest)
export GH_APP_ID="12345"                   # Or use GitHub App for auto-token
export GH_APP_INSTALL_ID="67890"
export GH_APP_KEY_FILE="/path/to/key.pem"
  • Twitter: TWITTERAPI_IO_KEY preferred ($3-5/mo); X_BEARER_TOKEN as fallback; auto mode tries twitterapiio first
  • Web Search: Tavily (preferred in auto mode) or Brave; optional, fallback to agent web_search if unavailable
  • GitHub: Auto-generates token from GitHub App if PAT not set; unauthenticated fallback (60 req/hr)
  • Reddit: No API key needed (uses public JSON API)

Cron / Scheduled Task Integration

OpenClaw Cron (Recommended)

The cron prompt should NOT hardcode the pipeline steps. Instead, reference references/digest-prompt.md and only pass configuration parameters. This ensures the pipeline logic stays in the skill repo and is consistent across all installations.

Daily Digest Cron Prompt

Read <SKILL_DIR>/references/digest-prompt.md and follow the complete workflow to generate a daily digest.

Replace placeholders with:
- MODE = daily
- TIME_WINDOW = past 1-2 days
- FRESHNESS = pd
- RSS_HOURS = 48
- ITEMS_PER_SECTION = 3-5
- BLOG_PICKS_COUNT = 2-3
- EXTRA_SECTIONS = (none)
- SUBJECT = Daily Tech Digest - YYYY-MM-DD
- WORKSPACE = <your workspace path>
- SKILL_DIR = <your skill install path>
- DISCORD_CHANNEL_ID = <your channel id>
- EMAIL = (optional)
- LANGUAGE = English
- TEMPLATE = discord

Follow every step in the prompt template strictly. Do not skip any steps.

Weekly Digest Cron Prompt

Read <SKILL_DIR>/references/digest-prompt.md and follow the complete workflow to generate a weekly digest.

Replace placeholders with:
- MODE = weekly
- TIME_WINDOW = past 7 days
- FRESHNESS = pw
- RSS_HOURS = 168
- ITEMS_PER_SECTION = 5-8
- BLOG_PICKS_COUNT = 3-5
- EXTRA_SECTIONS = 📊 Weekly Trend Summary (2-3 sentences summarizing macro trends)
- SUBJECT = Weekly Tech Digest - YYYY-MM-DD
- WORKSPACE = <your workspace path>
- SKILL_DIR = <your skill install path>
- DISCORD_CHANNEL_ID = <your channel id>
- EMAIL = (optional)
- LANGUAGE = English
- TEMPLATE = discord

Follow every step in the prompt template strictly. Do not skip any steps.

Why This Pattern?

  • Single source of truth: Pipeline logic lives in digest-prompt.md, not scattered across cron configs
  • Portable: Same skill on different OpenClaw instances, just change paths and channel IDs
  • Maintainable: Update the skill → all cron jobs pick up changes automatically
  • Anti-pattern: Do NOT copy pipeline steps into the cron prompt — it will drift out of sync

Multi-Channel Delivery Limitation

OpenClaw enforces cross-provider isolation: a single session can only send messages to one provider (e.g., Discord OR Telegram, not both). If you need to deliver digests to multiple platforms, create separate cron jobs for each provider:

# Job 1: Discord + Email
- DISCORD_CHANNEL_ID = <your-discord-channel-id>
- EMAIL = user@example.com
- TEMPLATE = discord

# Job 2: Telegram DM
- DISCORD_CHANNEL_ID = (none)
- EMAIL = (none)
- TEMPLATE = telegram

Replace DISCORD_CHANNEL_ID delivery with the target platform's delivery in the second job's prompt.

This is a security feature, not a bug — it prevents accidental cross-context data leakage.

Security Notes

Execution Model

This skill uses a prompt template pattern: the agent reads digest-prompt.md and follows its instructions. This is the standard OpenClaw skill execution model — the agent interprets structured instructions from skill-provided files. All instructions are shipped with the skill bundle and can be audited before installation.

Network Access

The Python scripts make outbound requests to:

  • RSS feed URLs (configured in tech-news-digest-sources.json)
  • Twitter/X API (api.x.com or api.twitterapi.io)
  • Brave Search API (api.search.brave.com)
  • Tavily Search API (api.tavily.com)
  • GitHub API (api.github.com)
  • Reddit JSON API (reddit.com)

No data is sent to any other endpoints. All API keys are read from environment variables declared in the skill metadata.

Shell Safety

Email delivery uses send-email.py which constructs proper MIME multipart messages with HTML body + optional PDF attachment. Subject formats are hardcoded (Daily Tech Digest - YYYY-MM-DD). PDF generation uses generate-pdf.py via weasyprint. The prompt template explicitly prohibits interpolating untrusted content (article titles, tweet text, etc.) into shell arguments. Email addresses and subjects must be static placeholder values only.

File Access

Scripts read from config/ and write to workspace/archive/. No files outside the workspace are accessed.

Support & Troubleshooting

Common Issues

  1. RSS feeds failing: Check network connectivity, use --verbose for details
  2. Twitter rate limits: Reduce sources or increase interval
  3. Configuration errors: Run validate-config.py for specific issues
  4. No articles found: Check time window (--hours) and source enablement

Debug Mode

All scripts support --verbose flag for detailed logging and troubleshooting.

Performance Tuning

  • Parallel Workers: Adjust MAX_WORKERS in scripts for your system
  • Timeout Settings: Increase TIMEOUT for slow networks
  • Article Limits: Adjust MAX_ARTICLES_PER_FEED based on needs

Security Considerations

Shell Execution

The digest prompt instructs agents to run Python scripts via shell commands. All script paths and arguments are skill-defined constants — no user input is interpolated into commands. Two scripts use subprocess:

  • run-pipeline.py orchestrates child fetch scripts (all within scripts/ directory)
  • fetch-github.py has two subprocess calls:
    1. openssl dgst -sha256 -sign for JWT signing (only if GH_APP_* env vars are set — signs a self-constructed JWT payload, no user content involved)
    2. gh auth token CLI fallback (only if gh is installed — reads from gh's own credential store)

No user-supplied or fetched content is ever interpolated into subprocess arguments. Email delivery uses send-email.py which builds MIME messages programmatically — no shell interpolation. PDF generation uses generate-pdf.py via weasyprint. Email subjects are static format strings only — never constructed from fetched data.

Credential & File Access

Scripts do not directly read ~/.config/, ~/.ssh/, or any credential files. All API tokens are read from environment variables declared in the skill metadata. The GitHub auth cascade is:

  1. $GITHUB_TOKEN env var (you control what to provide)
  2. GitHub App token generation (only if you set GH_APP_ID, GH_APP_INSTALL_ID, and GH_APP_KEY_FILE — uses inline JWT signing via openssl CLI, no external scripts involved)
  3. gh auth token CLI (delegates to gh's own secure credential store)
  4. Unauthenticated (60 req/hr, safe fallback)

If you prefer no automatic credential discovery, simply set $GITHUB_TOKEN and the script will use it directly without attempting steps 2-3.

Dependency Installation

This skill does not install any packages. requirements.txt lists optional dependencies (feedparser, jsonschema) for reference only. All scripts work with Python 3.8+ standard library. Users should install optional deps in a virtualenv if desired — the skill never runs pip install.

Input Sanitization

  • URL resolution rejects non-HTTP(S) schemes (javascript:, data:, etc.)
  • RSS fallback parsing uses simple, non-backtracking regex patterns (no ReDoS risk)
  • All fetched content is treated as untrusted data for display only

Network Access

Scripts make outbound HTTP requests to configured RSS feeds, Twitter API, GitHub API, Reddit JSON API, Brave Search API, and Tavily Search API. No inbound connections or listeners are created.

File v3.11.0:README.md

Tech News Digest

Automated tech news digest — 138 sources, 5-layer pipeline, one chat message to install.

English | 中文

Tests Python 3.8+ ClawHub MIT License

💬 Install in One Message

Tell your OpenClaw AI assistant:

"Install tech-news-digest and send a daily digest to #tech-news every morning at 9am"

That's it. Your bot handles installation, configuration, scheduling, and delivery — all through conversation.

More examples:

🗣️ "Set up a weekly AI digest, only LLM and AI Agent topics, deliver to Discord #ai-weekly every Monday"

🗣️ "Install tech-news-digest, add my RSS feeds, and send crypto news to Telegram"

🗣️ "Give me a tech digest right now, skip Twitter sources"

Or install via CLI:

clawhub install tech-news-digest

📊 What You Get

A quality-scored, deduplicated tech digest built from 138 sources:

| Layer | Sources | What | |-------|---------|------| | 📡 RSS | 49 feeds | OpenAI, Anthropic, Ben's Bites, HN, 36氪, CoinDesk… | | 🐦 Twitter/X | 48 KOLs | @karpathy, @VitalikButerin, @sama, @elonmusk… | | 🔍 Web Search | 4 topics | Brave Search API with freshness filters | | 🐙 GitHub | 28 repos | Releases from key projects (LangChain, vLLM, DeepSeek, Llama…) | | 🗣️ Reddit | 13 subs | r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency… |

Pipeline

       run-pipeline.py (~30s)
              ↓
  RSS ─┐
  Twitter ─┤
  Web ─────┤── parallel fetch ──→ merge-sources.py
  GitHub ──┤
  Reddit ──┘
              ↓
  Quality Scoring → Deduplication → Topic Grouping
              ↓
    Discord / Email / PDF output

Quality scoring: priority source (+3), multi-source cross-ref (+5), recency (+2), engagement (+1), Reddit score bonus (+1/+3/+5), already reported (-5).

⚙️ Configuration

  • config/defaults/sources.json — 138 built-in sources
  • config/defaults/topics.json — 4 topics with search queries & Twitter queries
  • User overrides in workspace/config/ take priority

🎨 Customize Your Sources

Works out of the box with 138 built-in sources — but fully customizable. Copy the defaults to your workspace config and override:

# Copy and customize
cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json

Your overlay file merges with defaults:

  • Override a source by matching its id — your version replaces the default
  • Add new sources with a unique id — appended to the list
  • Disable a built-in source — set "enabled": false on the matching id
{
  "sources": [
    {"id": "my-blog", "type": "rss", "enabled": true, "url": "https://myblog.com/feed", "topics": ["llm"]},
    {"id": "openai-blog", "enabled": false}
  ]
}

No need to copy the entire file — just include what you want to change.

🔧 Optional Setup

All environment variables are optional. The pipeline runs with whatever sources are available.

export TWITTERAPI_IO_KEY="..."  # twitterapi.io (~$5/mo) — enables Twitter layer
export X_BEARER_TOKEN="..."     # Twitter/X official API — alternative Twitter backend
export TAVILY_API_KEY="tvly-xxx"  # Tavily Search API (alternative, free 1000/mo)
export BRAVE_API_KEYS="k1,k2,k3" # Brave Search API keys (comma-separated, rotation)
export BRAVE_API_KEY="..."       # Fallback: single Brave key
export GITHUB_TOKEN="..."       # GitHub API — higher rate limits (auto-generated from GitHub App if unset)
export TWITTER_API_BACKEND="auto" # auto|twitterapiio|official (default: auto)
export BRAVE_PLAN="free"         # Override Brave rate limit detection: free|pro
export WEB_SEARCH_BACKEND="auto" # auto|brave|tavily (default: auto)
pip install weasyprint           # Enables PDF report generation

📂 Repository

GitHub: github.com/draco-agent/tech-news-digest

📄 License

MIT License — see LICENSE for details.

File v3.11.0:_meta.json

{ "ownerId": "kn74589cx1nbhnc3x0f3nwre39814699", "slug": "tech-news-digest", "version": "3.11.0", "publishedAt": 1772295728196 }

File v3.11.0:references/digest-prompt.md

Digest Prompt Template

Replace <...> placeholders before use. Daily defaults shown; weekly overrides in parentheses.

Placeholders

| Placeholder | Default | Weekly Override | |-------------|---------|----------------| | <MODE> | daily | weekly | | <TIME_WINDOW> | past 1-2 days | past 7 days | | <FRESHNESS> | pd | pw | | <RSS_HOURS> | 48 | 168 | | <ITEMS_PER_SECTION> | 3-5 | 5-8 | | <BLOG_PICKS_COUNT> | 2-3 | 3-5 | | <EXTRA_SECTIONS> | (none) | 📊 Weekly Trend Summary | | <SUBJECT> | Daily Tech Digest - YYYY-MM-DD | Weekly Tech Digest - YYYY-MM-DD | | <WORKSPACE> | Your workspace path | | | <SKILL_DIR> | Installed skill directory | | | <DISCORD_CHANNEL_ID> | Target channel ID | | | <EMAIL> | (optional) Recipient email | | | <EMAIL_FROM> | (optional) e.g. MyBot <bot@example.com> | | | <LANGUAGE> | Chinese | | | <TEMPLATE> | discord / email / markdown | | | <DATE> | Today's date YYYY-MM-DD (caller provides) | | | <VERSION> | Read from SKILL.md frontmatter | |


Generate the <MODE> tech digest for <DATE>. Use <DATE> as the report date — do NOT infer it.

Configuration

Read config files (workspace overrides take priority over defaults):

  1. Sources: <WORKSPACE>/config/tech-news-digest-sources.json → fallback <SKILL_DIR>/config/defaults/sources.json
  2. Topics: <WORKSPACE>/config/tech-news-digest-topics.json → fallback <SKILL_DIR>/config/defaults/topics.json

Context: Previous Report

Read the most recent file from <WORKSPACE>/archive/tech-news-digest/ to avoid repeats and follow up on developing stories. Skip if none exists.

Data Collection Pipeline

Use the unified pipeline (runs all 5 sources in parallel, ~30s):

python3 <SKILL_DIR>/scripts/run-pipeline.py \
  --defaults <SKILL_DIR>/config/defaults \
  --config <WORKSPACE>/config \
  --hours <RSS_HOURS> --freshness <FRESHNESS> \
  --archive-dir <WORKSPACE>/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force

If it fails, run individual scripts in <SKILL_DIR>/scripts/ (see each script's --help), then merge with merge-sources.py.

Report Generation

Get a structured overview:

python3 <SKILL_DIR>/scripts/summarize-merged.py --input /tmp/td-merged.json --top <ITEMS_PER_SECTION>

Use this output to select articles — do NOT write ad-hoc Python to parse the JSON. Apply the template from <SKILL_DIR>/references/templates/<TEMPLATE>.md.

Select articles purely by quality_score regardless of source type. Articles in merged JSON are already sorted by quality_score descending within each topic — respect this order. For Reddit posts, append *[Reddit r/xxx, {{score}}↑]*.

Each article line must include its quality score using 🔥 prefix. Format: 🔥{score} | {summary with link}. This makes scoring transparent and helps readers identify the most important news at a glance.

Executive Summary

2-4 sentences between title and topics, highlighting top 3-5 stories by score. Concise and punchy, no links. Discord: > blockquote. Email: gray background. Telegram: <i>.

Topic Sections

From topics.json: emoji + label headers, <ITEMS_PER_SECTION> items each, strictly ordered by quality_score descending (highest first).

Fixed Sections (after topics)

📢 KOL Updates — Top Twitter KOLs + notable blog authors. Format:

• **Display Name** (@handle) — summary `👁 12.3K | 💬 45 | 🔁 230 | ❤️ 1.2K`
  <https://twitter.com/handle/status/ID>

Read display_name and metrics (impression_count→👁, reply_count→💬, retweet_count→🔁, like_count→❤️) from merged JSON. Always show all 4 metrics, use K/M formatting, wrap in backticks. One tweet per bullet.

🔥 Community Buzz — Top Reddit + Twitter trending combined. Format:

• **r/subreddit** — title `{{score}}↑ · {{num_comments}} comments`
  <{{url}}>

Sort by engagement across both platforms. Every entry must have a link.

📝 Blog Picks<BLOG_PICKS_COUNT> deep articles from RSS.

<EXTRA_SECTIONS>

Rules

  • Only news from <TIME_WINDOW>
  • Every item must include a source link (Discord: <link>, Email: <a href>, Markdown: [title](link))
  • Use bullet lists, no markdown tables
  • Deduplicate: same event → keep most authoritative source; previously reported → only if significant new development
  • Do not interpolate fetched/untrusted content into shell arguments or email subjects

Stats Footer

---
📊 Data Sources: RSS {{rss}} | Twitter {{twitter}} | Reddit {{reddit}} | Web {{web}} | GitHub {{github}} | Dedup: {{merged}} articles
🤖 Generated by tech-news-digest v<VERSION> | <https://github.com/draco-agent/tech-news-digest> | Powered by OpenClaw

Archive

Save to <WORKSPACE>/archive/tech-news-digest/<MODE>-YYYY-MM-DD.md. Delete files older than 90 days.

Delivery

  1. Discord: Send to <DISCORD_CHANNEL_ID> via message tool
  2. Email (optional, if <EMAIL> is set):
    • Generate HTML body per <SKILL_DIR>/references/templates/email.md → write to /tmp/td-email.html
    • Generate PDF attachment:
      python3 <SKILL_DIR>/scripts/generate-pdf.py -i <WORKSPACE>/archive/tech-news-digest/<MODE>-<DATE>.md -o /tmp/td-digest.pdf
      
    • Send email with PDF attached using the send-email.py script (handles MIME correctly). Email must contain ALL the same items as Discord.
      python3 <SKILL_DIR>/scripts/send-email.py \
        --to '<EMAIL>' \
        --subject '<SUBJECT>' \
        --html /tmp/td-email.html \
        --attach /tmp/td-digest.pdf \
        --from '<EMAIL_FROM>'
      
    • Omit --from if <EMAIL_FROM> is not set. Omit --attach if PDF generation failed. SUBJECT must be a static string. If delivery fails, log error and continue.

Write the report in <LANGUAGE>.

File v3.11.0:references/templates/discord.md

Tech Digest Discord Template

Discord-optimized format with bullet points and link suppression.

Template Structure

# 🚀 Tech Digest - {{DATE}}

{{#topics}}
## {{emoji}} {{label}}

{{#articles}}
• 🔥{{quality_score}} | {{title}}
  <{{link}}>
  {{#multi_source}}*[{{source_count}} sources]*{{/multi_source}}

{{/articles}}
{{/topics}}

---
📊 Data Sources: RSS {{rss_count}} | Twitter {{twitter_count}} | Reddit {{reddit_count}} | Web {{web_count}} | GitHub {{github_count}} releases | After dedup: {{merged_count}} articles
🤖 Generated by tech-news-digest v{{version}} | <https://github.com/draco-agent/tech-news-digest> | Powered by OpenClaw

Delivery

  • Default: Channel — Send to the Discord channel specified by DISCORD_CHANNEL_ID
  • Use message tool with target set to the channel ID for channel delivery
  • For DM delivery instead, set target to a user ID

Discord-Specific Features

  • Link suppression: Wrap links in <> to prevent embeds
  • Bullet format: Use for clean mobile display
  • No tables: Discord mobile doesn't handle markdown tables well
  • Emoji headers: Visual hierarchy with topic emojis
  • Concise metadata: Source count and multi-source indicators
  • Character limits: Discord messages have 2000 char limit, may need splitting

Example Output

# 🚀 Tech Digest - 2026-02-15

## 🧠 LLM / Large Models

• 🔥15 | OpenAI releases GPT-5 with breakthrough reasoning capabilities
  <https://openai.com/blog/gpt5-announcement>
  *[3 sources]*

• 🔥12 | Meta's Llama 3.1 achieves new MMLU benchmarks
  <https://ai.meta.com/blog/llama-31-release>

## 🤖 AI Agent

• 🔥14 | LangChain launches production-ready agent framework
  <https://blog.langchain.dev/production-agents>

## 💰 Cryptocurrency

• 🔥18 | Bitcoin reaches new ATH at $67,000 amid ETF approval
  <https://coindesk.com/markets/btc-ath-etf>
  *[2 sources]*

## 📢 KOL Updates

• **Elon Musk** (@elonmusk) — Confirmed X's crypto trading feature `👁 2.1M | 💬 12.3K | 🔁 8.5K | ❤️ 49.8K`
  <https://twitter.com/elonmusk/status/123456789>
• **@saylor** — Valentine's BTC enthusiasm `👁 450K | 💬 1.2K | 🔁 3.1K | ❤️ 13K`
  <https://twitter.com/saylor/status/987654321>

---
📊 Data Sources: RSS 285 | Twitter 67 | Reddit 45 | Web 60 | GitHub 29 releases | After dedup: 95 articles

Variables

  • {{DATE}} - Report date (YYYY-MM-DD format)
  • {{topics}} - Array of topic objects
  • {{emoji}} - Topic emoji
  • {{label}} - Topic display name
  • {{articles}} - Array of article objects per topic
  • {{title}} - Article title (truncated if needed)
  • {{link}} - Article URL
  • {{quality_score}} - Article quality score (higher = more important)
  • {{multi_source}} - Boolean, true if article from multiple sources
  • {{source_count}} - Number of sources for this article
  • {{total_sources}} - Total number of sources used
  • {{total_articles}} - Total articles in digest

File v3.11.0:references/templates/email.md

Tech Digest Email Template

HTML email format optimized for Gmail/Outlook rendering.

Delivery

Send via gog gmail send with --body-html flag:

gog gmail send --to '<EMAIL>' --subject '<SUBJECT>' --body-html '<HTML_CONTENT>'

Important: Use --body-html, NOT --body. Plain text markdown will not render properly in email clients.

Template Structure

The agent should generate an HTML email body. Use inline styles (email clients strip <style> blocks).

<div style="max-width:640px;margin:0 auto;font-family:-apple-system,BlinkMacSystemFont,'Segoe UI',Roboto,sans-serif;color:#1a1a1a;line-height:1.6">

  <h1 style="font-size:22px;border-bottom:2px solid #e5e5e5;padding-bottom:8px">
    🐉 {{TITLE}}
  </h1>

  <!-- Optional: Executive Summary for weekly -->
  <p style="color:#555;font-size:14px;background:#f8f9fa;padding:12px;border-radius:6px">
    {{SUMMARY}}
  </p>

  <!-- Topic Section -->
  <h2 style="font-size:17px;margin-top:24px;color:#333">{{emoji}} {{label}}</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>🔥{{quality_score}}</strong> {{title}} — {{description}}
      <br><a href="{{link}}" style="color:#0969da;font-size:13px">{{link}}</a>
    </li>
  </ul>

  <!-- Repeat for each topic -->

  <!-- KOL Section: Read metrics from twitter JSON data (metrics.impression_count, reply_count, retweet_count, like_count). One tweet per <li>. -->
  <h2 style="font-size:17px;margin-top:24px;color:#333">📢 KOL Updates</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>{{display_name}}</strong> (@{{handle}}) — {{summary}}
      <br><code style="font-size:12px;color:#888;background:#f4f4f4;padding:2px 6px;border-radius:3px">👁 {{views}} | 💬 {{replies}} | 🔁 {{retweets}} | ❤️ {{likes}}</code>
      <br><a href="{{tweet_link}}" style="color:#0969da;font-size:13px">{{tweet_link}}</a>
    </li>
  </ul>

  <!-- Twitter/X Trending Section: Each entry must include at least one reference link -->
  <h2 style="font-size:17px;margin-top:24px;color:#333">🔥 Community Buzz</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>{{trending_topic}}</strong> — {{summary}}
      <br><a href="{{reference_link}}" style="color:#0969da;font-size:13px">{{reference_link}}</a>
    </li>
  </ul>

  <!-- Blog / Releases sections -->

  <!-- Footer -->
  <hr style="border:none;border-top:1px solid #e5e5e5;margin:24px 0">
  <p style="font-size:12px;color:#888">
    📊 Data Sources: RSS {{rss_count}} | Twitter {{twitter_count}} | Reddit {{reddit_count}} | Web {{web_count}} | GitHub {{github_count}} releases | After dedup: {{merged_count}} articles
    <br>🤖 Generated by <a href="https://github.com/draco-agent/tech-news-digest" style="color:#0969da">tech-news-digest</a> v{{version}} | Powered by <a href="https://openclaw.ai" style="color:#0969da">OpenClaw</a>
  </p>

</div>

Style Guidelines

  • Max width: 640px centered (mobile-friendly)
  • Fonts: System font stack (no web fonts in email)
  • All styles inline: Email clients strip <style> tags
  • Links: Use full URLs, styled with color:#0969da
  • Headings: h1 for title (22px), h2 for topics (17px)
  • Lists: <ul> with <li>, adequate spacing
  • Footer: Small gray text with stats
  • No images: Pure text/HTML for maximum compatibility
  • No tables for layout: Use div + inline styles

Example Output

<div style="max-width:640px;margin:0 auto;font-family:-apple-system,BlinkMacSystemFont,'Segoe UI',Roboto,sans-serif;color:#1a1a1a;line-height:1.6">

  <h1 style="font-size:22px;border-bottom:2px solid #e5e5e5;padding-bottom:8px">
    🐉 Daily Tech Digest — 2026-02-15
  </h1>

  <h2 style="font-size:17px;margin-top:24px;color:#333">🧠 LLM / Large Models</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>GPT-5.2 achieves first theoretical physics discovery</strong> — Collaboration with IAS, Cambridge, Harvard on gluon interactions
      <br><a href="https://twitter.com/OpenAI/status/2022390096625078389" style="color:#0969da;font-size:13px">twitter.com/OpenAI</a>
    </li>
    <li style="margin-bottom:10px">
      <strong>ByteDance releases Doubao 2.0</strong> — Full upgrade across Agent, image, and video
      <br><a href="https://www.jiqizhixin.com/articles/2026-02-14-9" style="color:#0969da;font-size:13px">jiqizhixin.com</a>
    </li>
    <li style="margin-bottom:10px">
      <strong>Dario Amodei: nearing the end of exponential growth</strong> — In-depth Anthropic CEO interview
      <br><a href="https://www.dwarkesh.com/p/dario-amodei-2" style="color:#0969da;font-size:13px">dwarkesh.com</a>
    </li>
  </ul>

  <h2 style="font-size:17px;margin-top:24px;color:#333">🤖 AI Agent</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>Stanford AI Town startup raises $100M</strong> — Backed by Fei-Fei Li, Karpathy
      <br><a href="https://www.qbitai.com/2026/02/380347.html" style="color:#0969da;font-size:13px">qbitai.com</a>
    </li>
  </ul>

  <h2 style="font-size:17px;margin-top:24px;color:#333">💰 Cryptocurrency</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>X to launch crypto & stock trading</strong> — Smart Cashtags feature coming soon
      <br><a href="https://www.theblock.co/post/389952" style="color:#0969da;font-size:13px">theblock.co</a>
    </li>
  </ul>

  <hr style="border:none;border-top:1px solid #e5e5e5;margin:24px 0">
  <p style="font-size:12px;color:#888">
    📊 Data Sources: RSS 287 | Twitter 71 | Reddit 45 | Web 60 | GitHub 29 releases | After dedup: 140 articles
    <br>Generated by Tech News Digest
  </p>

</div>

File v3.11.0:references/templates/pdf.md

Tech Digest PDF Template

Professional PDF output with Chinese typography, emoji icons, and A4 layout.

Generation

Generate PDF from the markdown report using generate-pdf.py:

python3 scripts/generate-pdf.py --input /tmp/td-report.md --output /tmp/td-digest.pdf

Prerequisites

  • weasyprint: pip install weasyprint
  • Chinese fonts: apt install fonts-noto-cjk (Noto Sans CJK SC)

Workflow

  1. Generate the digest in markdown format first (same as Discord template output)
  2. Save the markdown to a temp file (e.g., /tmp/td-report.md)
  3. Run generate-pdf.py to convert to PDF
  4. Optionally attach the PDF to Discord or email

Features

  • A4 layout with 2cm/2.5cm margins
  • Noto Sans CJK SC font for native Chinese rendering
  • Emoji support — section icons (🧠🤖💰🔬) render correctly
  • Page headers/footers — "Tech Digest" header, page numbers
  • Blue accent color scheme — headers, links, blockquote borders
  • Blockquote summary — highlighted executive summary area
  • Source links — compact, below each item
  • Responsive bullet lists — clean indentation

Example Markdown Input

The PDF generator accepts the same markdown format as the Discord template:

# 🚀 科技日报 - 2026-02-25

> 今日要闻:OpenAI 发布新模型,Anthropic 推出 Claude 4...

## 🧠 LLM / 大语言模型

• **OpenAI 发布 GPT-5** — 全新推理能力突破
  <https://openai.com/blog/gpt5>

• **Anthropic Claude 4 上线** — 更强的代码能力
  <https://anthropic.com/claude-4>

## 💰 Crypto / 区块链

• **以太坊 Pectra 升级完成** — EIP-7702 正式上线
  <https://ethereum.org/pectra>

---
📊 数据源: RSS 180 | Twitter 98 | Reddit 45 | Web 20 | GitHub 15
🤖 Generated by tech-news-digest v3.9.1

Delivery

# Generate PDF
python3 scripts/generate-pdf.py -i /tmp/td-report.md -o /tmp/td-digest.pdf

# Attach to Discord
# (use message tool with filePath parameter)

# Attach to email
mail -a /tmp/td-digest.pdf -s "Tech Digest" recipient@example.com < /dev/null

File v3.11.0:CHANGELOG.md

Changelog

v3.11.0

  • Tavily Search backend: Alternative to Brave Search via TAVILY_API_KEY + WEB_SEARCH_BACKEND env
  • Quality scores in output: 🔥 score prefix on every article, strict descending order per topic
  • Domain limit fix: Exempt x.com/github.com/reddit.com from per-topic domain limits (#1)
  • Brave multi-key: BRAVE_API_KEYS for comma-separated key rotation
  • Config naming: User overlay files renamed to tech-news-digest-sources.json / tech-news-digest-topics.json
  • Tests: 41 unit + integration tests with real fixture data, GitHub Actions CI (Python 3.9 + 3.12)
  • Docs: Full env var alignment, Network Access/Shell Safety updates, README badges, CN sync

v3.10.3

  • Docs: Align API Keys & Environment with all 10 actual env vars
  • Docs: Update Network Access (add Reddit) and Shell Safety (send-email.py + generate-pdf.py)
  • Refactor: Rename user overlay configs to tech-news-digest-sources.json / tech-news-digest-topics.json to avoid naming conflicts

v3.10.2

  • Fix domain limits: Exempt multi-author platforms (x.com, github.com, reddit.com) from per-topic domain limits — previously 77 tweets were cut to 12 (#1)
  • Brave multi-key: Prefer BRAVE_API_KEYS (comma-separated) over BRAVE_API_KEY for key rotation in fetch-web.py

v3.10.1

  • Fix email MIME: New send-email.py — proper multipart MIME construction for HTML body + PDF attachment (replaces broken mail -a -A approach)
  • Docs alignment: README + SKILL.md updated to v3.10 (source counts, PDF, all scripts documented)

v3.10.0

  • PDF generation: New generate-pdf.py script — converts markdown digest to styled A4 PDF with Chinese typography (Noto Sans CJK SC), emoji icons, page headers/footers, blue accent theme. Requires weasyprint.
  • PDF template: references/templates/pdf.md with usage docs and examples

v3.9.1

  • Remove unused markdown and telegram templates
  • Add sanitize-html.py for safe markdown→HTML email conversion (XSS-safe, inline CSS)

v3.9.0

  • URL-based dedup: merge-sources now deduplicates by normalized URL (domain+path) before title similarity, catching cross-source duplicates
  • Brave rate limit caching: detect_brave_rate_limit() results cached for 24h; supports BRAVE_PLAN=free|pro env override
  • source-health: Now tracks Reddit (--reddit) and Web (--web) sources; flexible key detection
  • run-pipeline: --skip (comma-separated step names) and --reuse-dir (reuse intermediate outputs) for partial reruns

v3.8.1

  • merge-sources: Fix getattr → direct args.reddit; domain limit stats now accurate; SequenceMatcher early-exit for >30% length diff
  • merge-sources: RSS priority sources get +2 extra score bonus (prevent drowning by low-engagement tweets)
  • run-pipeline: Add --twitter-backend parameter (transparent passthrough); clean up tmp dir after success
  • fetch-rss: Warn when feedparser not installed (basic regex parser fallback)
  • config_loader: Validate required fields (id, type, enabled) on source load, skip invalid with warning

v3.8.0

  • twitterapiio pagination: Fetches up to 2 pages (40 tweets) for high-volume users; logs truncation warning
  • Unified tweet limit: MAX_TWEETS_PER_USER 10→20 for official backend (matches twitterapiio)
  • Shared result helpers: _make_result() / _make_error() on base class, reduces duplication
  • Smarter rate limiting: RateLimiter class with threading.Lock for twitterapiio (5 QPS); replaces per-thread sleep
  • Retry improvements: RETRY_COUNT 1→2 (3 attempts); twitterapiio 429 wait 60s→5s
  • Tweet text limit: 200→280 chars (matches Twitter's actual limit)
  • Empty result format: Now matches normal output structure for consistent downstream parsing
  • Removed redundant isReply filter in twitterapiio (API already excludes replies)

v3.7.1

  • twitterapi.io bugfix: Fix response envelope parsing (data.tweets not top-level tweets)
  • twitterapi.io concurrency: 3-worker parallel fetch with progress logs showing tweet counts and top likes
  • test-pipeline.sh revamp: --only, --skip, --topics, --ids, --twitter-backend filtering; per-step timing; detailed --help

v3.7.0

  • twitterapi.io backend: Alternative Twitter data source via TWITTERAPI_IO_KEY — no username→ID resolution needed, simpler API, same normalized output format
  • Backend auto-detection: TWITTER_API_BACKEND=auto (default) uses twitterapi.io if key is set, else falls back to official X API v2
  • --backend CLI arg: Override env var per invocation (official, twitterapiio, auto)
  • Backend abstraction: fetch-twitter.py refactored with TwitterBackend base class and two implementations (OfficialBackend, TwitterApiIoBackend)

v3.6.3

  • Add GitHub source: zeroclaw-labs/zeroclaw (137→138 total, 27→28 GitHub)

v3.6.2

  • Add 3 GitHub sources: cloudflare/moltworker, sipeed/picoclaw, HKUDS/nanobot (134→137 total, 24→27 GitHub)

v3.6.1

  • Prompt review & optimization pass (no functional changes)

v3.6.0

  • Simplify digest-prompt: 232→122 lines (-47%), remove fallback scripts block, merge redundant rules
  • Add optional <EMAIL_FROM> placeholder for sender display name
  • Add "Environment vs Code" separation rule to CONTRIBUTING.md

v3.5.1

  • Email delivery: prefer mail (msmtp) over gog, remove redundant fallback options
  • Require email content to match Discord (no abbreviation or skipped sections)
  • Add CONTRIBUTING.md with development conventions

v3.5.0

  • Unified source count: 134 sources (49 RSS + 48 Twitter + 24 GitHub + 13 Reddit)
  • Updated README source counts and sub-totals

v3.4.9

  • Declare openssl as optional binary in SKILL.md (used for GitHub App JWT signing)

v3.4.8

  • New summarize-merged.py helper: Outputs structured human-readable summary of merged data, sorted by quality score with metrics/sources
  • Prevent ad-hoc JSON parsing: digest-prompt.md now instructs agents to use summarize-merged.py instead of writing inline Python (which often failed with AttributeError on nested structures)

v3.4.7

  • Inline GitHub App JWT signing: Remove GH_APP_TOKEN_SCRIPT env var entirely. Token generation now built into fetch-github.py using openssl CLI for RS256 signing — no external scripts executed, no arbitrary code execution risk.
  • Only 3 env vars needed: GH_APP_ID, GH_APP_INSTALL_ID, GH_APP_KEY_FILE
  • Remove unused imports, fix bare excepts across all scripts

v3.4.6

  • Add reddit to config/schema.json source type enum (was missing, caused validation mismatch)
  • Rename all archive paths tech-digest/tech-news-digest/ for consistency
  • Fix Discord template: default delivery is channel (via DISCORD_CHANNEL_ID), not DM
  • GH_APP_TOKEN_SCRIPT: add trust warning in code and env var description
  • Path placeholders: SKILL.md uses <workspace>/ consistently with digest-prompt.md

v3.4.5

  • Fix source count inconsistencies across docs (131/132 → 133: 49 RSS + 49 Twitter + 22 GitHub + 13 Reddit)
  • Rename legacy tech-digest references to tech-news-digest in comments, descriptions, and cache file paths

v3.4.4

  • Remove hardcoded Discord channel ID from SKILL.md (use <your-discord-channel-id> placeholder)
  • Cron prompt examples: Chinese → English, default LANGUAGE = English
  • Remove outdated "Migration from v1.x" section

v3.4.3

  • Audit compliance: Address all ClawHub Code Insights findings:
    • Declare gh as optional binary in SKILL.md metadata
    • Document credential access cascade and file access scope in security section
    • Add "Dependency Installation" section clarifying skill never runs pip install
    • Explicitly state scripts do not read ~/.config/, ~/.ssh/, or arbitrary credential files

v3.4.2

  • Remove hardcoded GitHub App credentials: App ID, install ID, key file path, and token script path now read exclusively from env vars (GH_APP_ID, GH_APP_INSTALL_ID, GH_APP_KEY_FILE, GH_APP_TOKEN_SCRIPT). No defaults — if not set, this auth method is silently skipped.
  • Declare new env vars in SKILL.md: All 4 GitHub App env vars declared in metadata
  • Fix security docs: Updated Shell Execution section to accurately describe subprocess.run() usage in run-pipeline.py and fetch-github.py

v3.4.1

  • KOL Display Names: KOL Updates section now shows "Sam Altman (@sama)" instead of bare "@sama" across all templates (Discord, Email, Telegram)
  • display_name in Merged JSON: merge-sources.py propagates Twitter source name to article-level display_name field, eliminating need to re-read raw Twitter data
  • New Twitter Sources: Added @OpenClawAI (official) and @steipete (Peter Steinberger), total 49 Twitter KOLs / 133 sources
  • Enforce Unified Pipeline: digest-prompt.md now says "You MUST use" run-pipeline.py, individual steps demoted to <details> fallback with --force flags

v3.4.0

  • Unified Pipeline: New run-pipeline.py runs all 5 fetch steps (RSS, Twitter, GitHub, Reddit, Web) in parallel, then merges — total ~30s vs ~3-4min sequential. Digest prompt updated to use this by default.
  • Reddit Parallel Fetch: fetch-reddit.py now uses ThreadPoolExecutor(max_workers=4) instead of sequential requests with sleep(1)
  • Reddit 403 Fix: Added explicit ssl.create_default_context() and Accept-Language header to fix Reddit blocking Python's default urllib TLS fingerprint
  • Brave API Auto-Concurrency: fetch-web.py probes x-ratelimit-limit header at startup — paid plans auto-switch to parallel queries, free plans stay sequential
  • GitHub Auto-Auth: fetch-github.py resolves tokens in priority order: $GITHUB_TOKEN → GitHub App auto-generate → gh CLI → unauthenticated. No manual token setup needed if GitHub App credentials exist.
  • Timeout Increase: All fetch scripts 15s → 30s per HTTP request; pipeline per-step subprocess 120s → 180s
  • Pipeline Metadata: run-pipeline.py saves *.meta.json with per-step timing, counts, and status

v3.3.2

  • Declare tools and file access: Added tools (python3 required, gog optional) and files (read/write paths) to SKILL.md metadata, addressing VirusTotal "undeclared tools/binaries" and "modify workspace files" audit findings
  • Added metadata.openclaw.requires: Declares python3 binary dependency

v3.3.1

  • Remove anthropic-rss mirror: Removed third-party community RSS mirror (anthropic-rss) to eliminate supply chain risk flagged by VirusTotal Code Insights. Anthropic coverage remains via Twitter KOL, GitHub releases, and Reddit sources.
  • Remove Third-Party RSS Sources section from SKILL.md security docs (no longer applicable)

v3.3.0

  • RSS Domain Validation: New expected_domains field in sources.json rejects articles from unexpected origins (applied to anthropic-rss mirror)
  • Email Shell Safety: HTML body written to temp file before CLI delivery; subjects restricted to static format strings
  • Discord Embed Suppression: Footer links wrapped in <> to prevent preview embeds

v3.2.1

  • Mandatory Reddit Execution: Agent explicitly required to run fetch-reddit.py script — cannot skip or generate fake output

v3.2.0

  • Unified English Templates: All prompt instructions, section titles, stats footer, and example content standardized to English. Output language controlled by <LANGUAGE> placeholder at runtime.

v3.1.0

  • Executive Summary: 2-4 sentence overview of top stories at the beginning of each digest
  • Community Buzz Section: Merged Twitter/X Trending and Reddit Hot Discussions into unified 🔥 社区热议
  • Reddit in Topic Sections: Reddit posts now selected by quality_score alongside other sources
  • Digest Footer Branding: Shows skill version and OpenClaw link
  • Prompt Fix: Agent explicitly instructed to read Reddit data from merged JSON

v3.0.0

  • Reddit Data Source: New fetch-reddit.py script — 5th data layer using Reddit's public JSON API (no auth required). 13 subreddits: r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency, r/artificial, r/ethereum, r/ChatGPT, r/singularity, r/OpenAI, r/Bitcoin, r/programming, r/Anthropic, r/defi, r/ExperiencedDevs
  • Reddit Score Bonus: Posts with score > 500 get +5, > 200 get +3, > 100 get +1 in quality scoring
  • 10 New Non-Reddit Sources: Ben's Bites, The Decoder, a16z Crypto, Bankless (RSS); @ClementDelangue, @GregBrockman, @zuck (Twitter); MCP Servers, DeepSeek-V3, Meta Llama (GitHub)
  • Tweet Engagement Metrics: KOL entries display 👁|💬|🔁|❤️ stats in inline code blocks across all templates
  • Date Timezone Fix: Report date explicitly provided via <DATE> placeholder, preventing UTC/local mismatch
  • Mandatory Links: KOL Updates and Twitter/X Trending sections require source URLs for every entry
  • Graceful Twitter Degradation: Missing X_BEARER_TOKEN outputs empty JSON instead of failing
  • URL Sanitization: resolve_link() rejects non-HTTP(S) schemes
  • Security Documentation: Added Security Considerations section to SKILL.md
  • Total Sources: 132 (50 RSS + 47 Twitter + 22 GitHub + 13 Reddit + 4 web search topics)

v2.8.1

  • Metrics Data Fix: Agent now required to read actual metrics values from Twitter JSON data instead of defaulting to 0
  • Email Template Enhancement: Added KOL metrics and Twitter/X Trending section to email template

v2.8.0

  • Tweet Metrics Display: KOL entries show 👁|💬|🔁|❤️ engagement stats wrapped in inline code to prevent emoji enlargement on Discord
  • Standardized Metrics Format: Fixed 4-metric order, show 0 for missing values, one tweet per bullet with own URL
  • 10 New Sources (119 total): Ben's Bites, The Decoder, a16z Crypto, Bankless (RSS); @ClementDelangue, @GregBrockman, @zuck (Twitter); MCP Servers, DeepSeek-V3, Meta Llama (GitHub)

v2.7.0

  • Tweet Engagement Metrics: KOL Updates now display 👁 views, 💬 replies, 🔁 retweets, ❤️ likes from Twitter public_metrics across all templates (Discord, Email, Telegram)

v2.6.1

  • Graceful Twitter Degradation: Missing X_BEARER_TOKEN now outputs empty JSON and exits 0 instead of failing with exit code 1, allowing the pipeline to continue without Twitter data

v2.6.0

  • Date Timezone Fix: Added <DATE> placeholder to digest prompt — report date now explicitly provided by caller, preventing UTC/local timezone mismatch
  • Mandatory Links in KOL/Trending: KOL Updates and Twitter/X Trending sections now require source URLs for every entry (no link-free entries allowed)
  • URL Sanitization: resolve_link() in fetch-rss.py rejects non-HTTP(S) schemes (javascript:, data:, etc.)
  • Third-Party Source Annotation: Community-maintained RSS mirrors (e.g. anthropic-rss) are annotated with notes in sources.json
  • Security Documentation: Added Security Considerations section to SKILL.md covering shell execution model, input sanitization, and network access

v2.5.0

  • Twitter Reply Filter Fix: Use referenced_tweets field instead of text prefix to distinguish replies from mentions
  • Scoring Consistency: digest-prompt.md now matches code (PENALTY_OLD_REPORT = -5)
  • Template Version Cleanup: Removed hardcoded version numbers from email/markdown/telegram templates
  • Article Count Fix: merge-sources.py uses deduplicated count instead of inflated topic-grouped sum
  • Pipeline Resume Support: All fetch scripts support --force flag; skip if cached output < 1 hour old
  • Source Health Monitoring: New scripts/source-health.py tracks per-source success/failure history
  • End-to-End Test: New scripts/test-pipeline.sh smoke test for the full pipeline
  • Archive Auto-Cleanup: digest-prompt.md documents 90-day archive retention policy
  • Twitter Rate Limiting: Moved sleep into fetch_user_tweets for actual per-request rate limiting
  • Web Article Scoring: Web articles now use calculate_base_score instead of hardcoded 1.0
  • Dead Code Removal: Removed unused load_sources_with_overlay / load_topics_with_overlay wrappers

v2.4.0

  • Batch Twitter Lookup: Single API call for all username→ID resolution + 7-day local cache (~88→~45 API calls)
  • Smart Dedup: Token-based bucketing replaces O(n²) SequenceMatcher — only compares articles sharing 2+ key tokens
  • Conditional Fetch (RSS): ETag/Last-Modified caching, 304 responses skip parsing
  • Conditional Fetch (GitHub): Same caching pattern + prominent warning when GITHUB_TOKEN is unset
  • --no-cache flag: All fetch scripts support bypassing cache

v2.3.0

  • GitHub Releases: 19 tracked repositories as a fourth data source
  • Data Source Stats Footer: Pipeline statistics in all templates
  • Twitter Queries: Added to all 4 topics for better coverage
  • Simplified Cron Prompts: Reference digest-prompt.md with parameters only

v2.1.0

  • Unified Source Model: Single sources.json for RSS, Twitter, and web sources
  • Enhanced Topics: Richer topic definitions with search queries and filters
  • Pipeline Scripts: Modular fetch → merge → template workflow
  • Quality Scoring: Multi-source detection, deduplication, priority weighting
  • Multiple Templates: Discord, email, and markdown output formats
  • Configuration Validation: JSON schema validation and consistency checks
  • User Customization: Workspace config overrides for personalization

File v3.11.0:CONTRIBUTING.md

Contributing / Development Conventions

Version Management

  • SemVer: SKILL.md frontmatter version field is the single source of truth
  • CHANGELOG.md: reverse-chronological, update with every version bump
  • Every change must update both SKILL.md version + CHANGELOG.md + git commit & push
  • Changelog version format: ## v3.5.0 (prefixed with v)

Code Conventions

  • All prompts, templates, comments, and code in English
  • Output language controlled at runtime via LANGUAGE variable
  • Python: use except Exception: — never bare except:
  • No hardcoded credentials — all secrets via environment variables
  • When adding data sources, update sources.json schema and README source count

Security

  • ClawHub audit compliance: declare all tools/bins, file read/write paths, credential access in SKILL.md metadata
  • No third-party untrusted RSS mirrors (supply chain risk)
  • HTML email bodies written to temp files before CLI delivery
  • Subjects restricted to static format strings (no injection)
  • Discord embed suppression: wrap links in <> to prevent previews

Debugging

  • Full pipeline: python3 scripts/run-pipeline.py --verbose --force
  • Each step generates *.meta.json with timing, counts, and status
  • Individual scripts can be run standalone for targeted debugging

File Structure

SKILL.md          — Skill metadata (version, env vars, tools, files)
CHANGELOG.md      — Version history
README.md         — English docs
README_CN.md      — Chinese docs
config/defaults/  — Default sources.json, topics.json
references/       — digest-prompt.md, output templates
scripts/          — Python pipeline scripts

Environment vs Code

  • Never push environment-specific config to repo — email sender names, API keys, file paths, channel IDs, timezone settings, etc. belong in local workspace config or env vars, not in skill code
  • Repo code uses <PLACEHOLDER> patterns; actual values are substituted at runtime
  • Local overrides go in workspace/config/, not in config/defaults/

Git Workflow

  • Commit messages: concise English, describe what changed
  • Push to main branch on github.com/draco-agent/tech-news-digest
  • No feature branches for solo development (direct to main)

File v3.11.0:README_CN.md

Tech News Digest

自动化科技资讯汇总 — 138 个数据源,5 层管道,一句话安装。

English | 中文

Tests Python 3.8+ ClawHub MIT License

💬 一句话安装

跟你的 OpenClaw AI 助手说:

"安装 tech-news-digest,每天早上 9 点发科技日报到 #tech-news 频道"

搞定。Bot 会自动安装、配置、定时、推送——全程对话完成。

更多示例:

🗣️ "配置一个每周 AI 周报,只要 LLM 和 AI Agent 板块,每周一发到 Discord #ai-weekly"

🗣️ "安装 tech-news-digest,加上我的 RSS 源,发送科技新闻到 Telegram"

🗣️ "现在就给我生成一份科技日报,跳过 Twitter 数据源"

或通过 CLI 安装:

clawhub install tech-news-digest

📊 你会得到什么

基于 138 个数据源 的质量评分、去重科技日报:

| 层级 | 数量 | 内容 | |------|------|------| | 📡 RSS | 49 个订阅源 | OpenAI、Anthropic、Ben's Bites、HN、36氪、CoinDesk… | | 🐦 Twitter/X | 48 个 KOL | @karpathy、@VitalikButerin、@sama、@elonmusk… | | 🔍 Web 搜索 | 4 个主题 | Brave Search API + 时效过滤 | | 🐙 GitHub | 28 个仓库 | 关键项目的 Release 跟踪(LangChain、vLLM、DeepSeek、Llama…) | | 🗣️ Reddit | 13 个子版块 | r/MachineLearning、r/LocalLLaMA、r/CryptoCurrency… |

数据管道

       run-pipeline.py (~30秒)
              ↓
  RSS ─┐
  Twitter ─┤
  Web ─────┤── 并行采集 ──→ merge-sources.py
  GitHub ──┤
  Reddit ──┘
              ↓
  质量评分 → 去重 → 主题分组
              ↓
    Discord / 邮件 / PDF 输出

质量评分:优先级源 (+3)、多源交叉验证 (+5)、时效性 (+2)、互动度 (+1~+5)、Reddit 热度加分 (+1/+3/+5)、已报道过 (-5)。

⚙️ 配置

  • config/defaults/sources.json — 138 个内置数据源
  • config/defaults/topics.json — 4 个主题,含搜索查询和 Twitter 查询
  • 用户自定义配置放 workspace/config/,优先级更高

🎨 自定义数据源

开箱即用,内置 138 个数据源——但完全可自定义。将默认配置复制到 workspace 并覆盖:

# 复制并自定义
cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json

你的配置文件会与默认配置合并

  • 覆盖id 匹配的源会被你的版本替换
  • 新增:使用新的 id 即可添加自定义源
  • 禁用:对匹配的 id 设置 "enabled": false
{
  "sources": [
    {"id": "my-blog", "type": "rss", "enabled": true, "url": "https://myblog.com/feed", "topics": ["llm"]},
    {"id": "openai-blog", "enabled": false}
  ]
}

不需要复制整个文件——只写你要改的部分。

🔧 可选配置

所有环境变量均为可选,管道会自动使用可用的数据源。

export TWITTERAPI_IO_KEY="..."    # twitterapi.io (~$5/月) — 启用 Twitter 数据层
export X_BEARER_TOKEN="..."       # Twitter/X 官方 API — 备选 Twitter 后端
export TWITTER_API_BACKEND="auto" # auto|twitterapiio|official(默认: auto)
export TAVILY_API_KEY="tvly-xxx"  # Tavily Search API(替代方案,免费 1000 次/月)
export BRAVE_API_KEYS="k1,k2,k3" # Brave Search API 密钥(逗号分隔,自动轮换)
export BRAVE_API_KEY="..."        # 单密钥回退
export BRAVE_PLAN="free"          # 覆盖速率限制检测: free|pro
export WEB_SEARCH_BACKEND="auto" # auto|brave|tavily(默认: auto)
export GITHUB_TOKEN="..."         # GitHub API — 提高速率限制(未设置时自动从 GitHub App 生成)
pip install weasyprint             # 启用 PDF 报告生成

🧪 测试

python -m unittest discover -s tests -v   # 41 个测试,纯标准库

📂 仓库地址

GitHub: github.com/draco-agent/tech-news-digest

📄 开源协议

MIT License — 详见 LICENSE

File v3.11.0:config/defaults/sources.json

{ "_description": "Unified data sources configuration. RSS feeds, Twitter/X KOLs, and web search sources. Each source binds to topics and has enabled/priority fields.", "_updated": "2025-06-02", "_version": "2.5.0", "sources": [ { "id": "simonwillison-rss", "type": "rss", "name": "Simon Willison", "url": "https://simonwillison.net/atom/everything/", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "LLM/AI tooling, prolific blogger" }, { "id": "garymarcus-rss", "type": "rss", "name": "Gary Marcus", "url": "https://garymarcus.substack.com/feed", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "AI critic, industry analysis" }, { "id": "huggingface-rss", "type": "rss", "name": "Hugging Face Blog", "url": "https://huggingface.co/blog/feed.xml", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ], "note": "Open source AI/ML" }, { "id": "openai-rss", "type": "rss", "name": "OpenAI Blog", "url": "https://openai.com/blog/rss.xml", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "Official OpenAI updates" }, { "id": "sebas-rss", "type": "rss", "name": "Sebastian Raschka", "url": "https://magazine.sebastianraschka.com/feed", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "LLM research and tutorials" }, { "id": "lilian-rss", "type": "rss", "name": "Lil'Log (Lilian Weng)", "url": "https://lilianweng.github.io/index.xml", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "In-depth ML tutorials" }, { "id": "gwern-rss", "type": "rss", "name": "Gwern", "url": "https://gwern.substack.com/feed", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "Deep AI/ML research essays" }, { "id": "dwarkesh-rss", "type": "rss", "name": "Dwarkesh Patel", "url": "https://www.dwarkeshpatel.com/feed", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ], "note": "AI interviews and analysis" }, { "id": "minimaxir-rss", "type": "rss", "name": "minimaxir (Max Woolf)", "url": "https://minimaxir.com/index.xml", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "LLM benchmarks and experiments" }, { "id": "googleai-rss", "type": "rss", "name": "Google AI Blog", "url": "https://blog.google/technology/ai/rss/", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "Google AI research" }, { "id": "vitalik-rss", "type": "rss", "name": "Vitalik Buterin", "url": "https://vitalik.eth.limo/feed.xml", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Ethereum founder" }, { "id": "coindesk-rss", "type": "rss", "name": "CoinDesk", "url": "https://www.coindesk.com/arc/outboundfeeds/rss/", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Major crypto news" }, { "id": "theblock-rss", "type": "rss", "name": "The Block", "url": "https://www.theblock.co/rss.xml", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Crypto news" }, { "id": "decrypt-rss", "type": "rss", "name": "Decrypt", "url": "https://decrypt.co/feed", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Crypto and web3 news" }, { "id": "cointelegraph-rss", "type": "rss", "name": "Cointelegraph", "url": "https://cointelegraph.com/rss", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Crypto news and analysis" }, { "id": "hn-rss", "type": "rss", "name": "Hacker News Frontpage", "url": "https://hnrss.org/frontpage", "enabled": true, "priority": true, "topics": [ "frontier-tech" ], "note": "HN top stories" }, { "id": "ars-rss", "type": "rss", "name": "Ars Technica", "url": "https://feeds.arstechnica.com/arstechnica/index", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Tech news" }, { "id": "techcrunch-rss", "type": "rss", "name": "TechCrunch", "url": "https://techcrunch.com/feed/", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Startup and tech news" }, { "id": "verge-rss", "type": "rss", "name": "The Verge", "url": "https://www.theverge.com/rss/index.xml", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Tech news" }, { "id": "krebs-rss", "type": "rss", "name": "Krebs on Security", "url": "https://krebsonsecurity.com/feed/", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Cybersecurity" }, { "id": "daringfireball-rss", "type": "rss", "name": "Daring Fireball", "url": "https://daringfireball.net/feeds/main", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Apple/tech commentary" }, { "id": "pg-rss", "type": "rss", "name": "Paul Graham", "url": "http://www.aaronsw.com/2002/feeds/pgessays.rss", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Startups and tech essays" }, { "id": "troyhunt-rss", "type": "rss", "name": "Troy Hunt", "url": "https://www.troyhunt.com/rss/", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Security, HIBP" }, { "id": "antirez-rss", "type": "rss", "name": "antirez", "url": "http://antirez.com/rss", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Redis creator, systems" }, { "id": "mitchellh-rss", "type": "rss", "name": "Mitchell Hashimoto", "url": "https://mitchellh.com/feed.xml", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Ghostty, infrastructure" }, { "id": "geohot-rss", "type": "rss", "name": "Geohot", "url": "https://geohot.github.io/blog/feed.xml", "enabled": true, "priority": false, "topics": [ "frontier-tech", "llm" ], "note": "tinygrad, AI infrastructure" }, { "id": "ml-reddit-rss", "type": "rss", "name": "r/MachineLearning", "url": "https://www.reddit.com/r/MachineLearning/.rss", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "Reddit ML community" }, { "id": "36kr-rss", "type": "rss", "name": "36氪", "url": "https://36kr.com/feed", "enabled": true, "priority": true, "topics": [ "frontier-tech", "crypto" ], "note": "中文科技媒体" }, { "id": "synced-rss", "type": "rss", "name": "机器之心 Synced", "url": "https://www.jiqizhixin.com/rss", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "中文AI媒体" }, { "id": "qbitai-rss", "type": "rss", "name": "量子位 QbitAI", "url": "https://www.qbitai.com/feed", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ], "note": "中文AI媒体 (may 403)" }, { "id": "infoq-rss", "type": "rss", "name": "InfoQ 中文", "url": "https://www.infoq.cn/feed", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "技术社区" }, { "id": "sama-twitter", "type": "twitter", "name": "Sam Altman (OpenAI CEO)", "handle": "sama", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "OpenAI CEO" }, { "id": "openai-twitter", "type": "twitter", "name": "OpenAI official", "handle": "OpenAI", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "OpenAI official" }, { "id": "anthropic-twitter", "type": "twitter", "name": "Anthropic official", "handle": "AnthropicAI", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "Anthropic official" }, { "id": "ylecun-twitter", "type": "twitter", "name": "Yann LeCun (Meta AI)", "handle": "ylecun", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "Meta AI" }, { "id": "mistral-twitter", "type": "twitter", "name": "Mistral AI official", "handle": "MistralAI", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "Mistral AI official" }, { "id": "deepmind-twitter", "type": "twitter", "name": "Google DeepMind official", "handle": "GoogleDeepMind", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "Google DeepMind official" }, { "id": "googleai-twitter", "type": "twitter", "name": "Google AI official", "handle": "GoogleAI", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "Google AI official" }, { "id": "xai-twitter", "type": "twitter", "name": "xAI official", "handle": "xai", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "xAI official" }, { "id": "karpathy-twitter", "type": "twitter", "name": "Andrej Karpathy", "handle": "karpathy", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "AI researcher" }, { "id": "andrewng-twitter", "type": "twitter", "name": "Andrew Ng", "handle": "AndrewYNg", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ], "note": "AI educator" }, { "id": "jimfan-twitter", "type": "twitter", "name": "Jim Fan (NVIDIA)", "handle": "DrJimFan", "enabled": true, "priority": true, "topics": [ "ai-agent", "frontier-tech" ], "note": "NVIDIA AI" }, { "id": "hf-twitter", "type": "twitter", "name": "Hugging Face official", "handle": "huggingface", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ], "note": "Hugging Face official" }, { "id": "langchain-twitter", "type": "twitter", "name": "LangChain official", "handle": "LangChain", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "LangChain official" }, { "id": "llamaindex-twitter", "type": "twitter", "name": "LlamaIndex official", "handle": "llama_index", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "LlamaIndex official" }, { "id": "emad-twitter", "type": "twitter", "name": "Emad Mostaque", "handle": "EMostaque", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "Stability AI" }, { "id": "sebastian-twitter", "type": "twitter", "name": "Sebastian Raschka", "handle": "rasbt", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "AI researcher" }, { "id": "vitalik-twitter", "type": "twitter", "name": "Vitalik Buterin (Ethereum)", "handle": "VitalikButerin", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Ethereum founder" }, { "id": "cz-twitter", "type": "twitter", "name": "CZ (Binance)", "handle": "cz_binance", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Binance" }, { "id": "brian-twitter", "type": "twitter", "name": "Brian Armstrong (Coinbase)", "handle": "brian_armstrong", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Coinbase" }, { "id": "saylor-twitter", "type": "twitter", "name": "Michael Saylor (MicroStrategy)", "handle": "saylor", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "MicroStrategy" }, { "id": "pomp-twitter", "type": "twitter", "name": "Anthony Pompliano", "handle": "APompliano", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Crypto influencer" }, { "id": "zachxbt-twitter", "type": "twitter", "name": "ZachXBT", "handle": "zachxbt", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "on-chain investigator" }, { "id": "wu-twitter", "type": "twitter", "name": "Wu Blockchain", "handle": "WuBlockchain", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "吴说区块链" }, { "id": "discus-twitter", "type": "twitter", "name": "神鱼 DiscusFish", "handle": "bitfish", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "F2Pool/Cobo co-founder" }, { "id": "mindao-twitter", "type": "twitter", "name": "Mindao", "handle": "mindaoyang", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "dForce founder" }, { "id": "herbert-twitter", "type": "twitter", "name": "Herbert", "handle": "herbertcrypto", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "PANews founder" }, { "id": "elon-twitter", "type": "twitter", "name": "Elon Musk", "handle": "elonmusk", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Entrepreneur" }, { "id": "sundar-twitter", "type": "twitter", "name": "Sundar Pichai", "handle": "sundarpichai", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Google CEO" }, { "id": "pmarca-twitter", "type": "twitter", "name": "Marc Andreessen", "handle": "pmarca", "enabled": true, "priority": false, "topics": [ "frontier-tech", "crypto" ], "note": "a16z" }, { "id": "levie-twitter", "type": "twitter", "name": "Aaron Levie", "handle": "levie", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Box CEO" }, { "id": "satya-twitter", "type": "twitter", "name": "Satya Nadella", "handle": "satyanadella", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Microsoft CEO" }, { "id": "mit-tech-review-rss", "type": "rss", "name": "MIT Technology Review", "url": "https://www.technologyreview.com/feed", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent", "frontier-tech" ], "note": "AI policy + deep analysis" }, { "id": "venturebeat-ai-rss", "type": "rss", "name": "VentureBeat AI", "url": "https://venturebeat.com/category/ai/feed/", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ], "note": "AI industry news" }, { "id": "404media-rss", "type": "rss", "name": "404 Media", "url": "https://www.404media.co/rss", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Independent tech investigative journalism" }, { "id": "aisnakeoil-rss", "type": "rss", "name": "AI Snake Oil", "url": "https://aisnakeoil.substack.com/feed", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "Princeton professor, critical AI perspective" }, { "id": "bytebytego-rss", "type": "rss", "name": "ByteByteGo", "url": "https://blog.bytebytego.com/feed", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "System design + engineering by Alex Xu" }, { "id": "nvidia-ai-rss", "type": "rss", "name": "NVIDIA AI Blog", "url": "https://blogs.nvidia.com/feed/", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "GPU/AI infrastructure" }, { "id": "deepmind-rss", "type": "rss", "name": "Google DeepMind Blog", "url": "https://deepmind.google/blog/rss.xml", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "Frontier AI research" }, { "id": "producthunt-rss", "type": "rss", "name": "Product Hunt", "url": "https://www.producthunt.com/feed", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "New product discovery, AI tools" }, { "id": "messari-rss", "type": "rss", "name": "Messari", "url": "https://messari.io/rss", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Crypto research reports" }, { "id": "defiant-rss", "type": "rss", "name": "The Defiant", "url": "https://thedefiant.io/feed", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "DeFi professional media" }, { "id": "ifanr-rss", "type": "rss", "name": "爱范儿", "url": "https://www.ifanr.com/feed", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Chinese tech product reviews" }, { "id": "sspai-rss", "type": "rss", "name": "少数派", "url": "https://sspai.com/feed", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Chinese productivity + tech depth" }, { "id": "wired-rss", "type": "rss", "name": "Wired", "url": "https://www.wired.com/feed/rss", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Tech culture deep reporting" }, { "id": "ieee-spectrum-rss", "type": "rss", "name": "IEEE Spectrum", "url": "https://spectrum.ieee.org/feeds/feed.rss", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Engineering + frontier tech authority" }, { "id": "rowancheung-twitter", "type": "twitter", "name": "Rowan Cheung (The Rundown AI)", "handle": "rowancheung", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ], "note": "AI newsletter founder, 567K followers" }, { "id": "yudkowsky-twitter", "type": "twitter", "name": "Eliezer Yudkowsky", "handle": "ESYudkowsky", "enabled": true, "priority": false, "topics": [ "llm", "frontier-tech" ], "note": "AI safety pioneer" }, { "id": "demis-twitter", "type": "twitter", "name": "Demis Hassabis (DeepMind CEO)", "handle": "demishassabis", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "DeepMind CEO, Nobel laureate" }, { "id": "dario-twitter", "type": "twitter", "name": "Dario Amodei (Anthropic CEO)", "handle": "DarioAmodei", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ], "note": "Anthropic CEO" }, { "id": "hwchase-twitter", "type": "twitter", "name": "Harrison Chase (LangChain)", "handle": "hwchase17", "enabled": true, "priority": true, "topics": [ "ai-agent" ], "note": "LangChain founder, Agent ecosystem" }, { "id": "swyx-twitter", "type": "twitter", "name": "Swyx", "handle": "swyx", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ], "note": "AI Engineer community, Latent Space podcast" }, { "id": "erikbryn-twitter", "type": "twitter", "name": "Erik Brynjolfsson", "handle": "erikbryn", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Stanford Digital Economy Lab" }, { "id": "balaji-twitter", "type": "twitter", "name": "Balaji Srinivasan", "handle": "balaji", "enabled": true, "priority": true, "topics": [ "crypto", "frontier-tech" ], "note": "Former Coinbase CTO, macro thinker" }, { "id": "cobie-twitter", "type": "twitter", "name": "Cobie", "handle": "cobie", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Influential independent crypto analyst" }, { "id": "hsaka-twitter", "type": "twitter", "name": "Hsaka", "handle": "HsakaTrades", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Crypto trading analysis" }, { "id": "cochran-twitter", "type": "twitter", "name": "Adam Cochran", "handle": "adamscochran", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Cinneamhain Ventures, on-chain analysis" }, { "id": "lcermak-twitter", "type": "twitter", "name": "Larry Cermak", "handle": "lawmaster", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "The Block research director" }, { "id": "pytorch-github", "type": "github", "name": "PyTorch", "repo": "pytorch/pytorch", "enabled": true, "priority": true, "topics": [ "llm", "frontier-tech" ], "note": "Leading deep learning framework from Meta" }, { "id": "transformers-github", "type": "github", "name": "Hugging Face Transformers", "repo": "huggingface/transformers", "enabled": true, "priority": true, "topics": [ "llm" ], "note": "State-of-the-art transformer models library" }, { "id": "langchain-github", "type": "github", "name": "LangChain", "repo": "langchain-ai/langchain", "enabled": true, "priority": true, "topics": [ "ai-agent" ], "note": "Framework for building LLM applications" }, { "id": "llamaindex-github", "type": "github", "name": "LlamaIndex", "repo": "run-llama/llama_index", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "Data framework for LLM applications" }, { "id": "ollama-github", "type": "github", "name": "Ollama", "repo": "ollama/ollama", "enabled": true, "priority": true, "topics": [ "llm" ], "note": "Run LLMs locally with ease" }, { "id": "vllm-github", "type": "github", "name": "vLLM", "repo": "vllm-project/vllm", "enabled": true, "priority": false, "topics": [ "llm" ], "note": "High-throughput LLM inference engine" }, { "id": "openai-python-github", "type": "github", "name": "OpenAI Python SDK", "repo": "openai/openai-python", "enabled": true, "priority": false, "topics": [ "llm" ], "note": "Official OpenAI Python client library" }, { "id": "anthropic-sdk-github", "type": "github", "name": "Anthropic SDK", "repo": "anthropics/anthropic-sdk-python", "enabled": true, "priority": false, "topics": [ "llm" ], "note": "Official Anthropic Python SDK" }, { "id": "crewai-github", "type": "github", "name": "CrewAI", "repo": "crewAIInc/crewAI", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "Multi-agent AI collaboration framework" }, { "id": "autogen-github", "type": "github", "name": "AutoGen", "repo": "microsoft/autogen", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "Multi-agent conversation framework from Microsoft" }, { "id": "dify-github", "type": "github", "name": "Dify", "repo": "langgenius/dify", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "LLM app development platform" }, { "id": "openclaw-github", "type": "github", "name": "OpenClaw", "repo": "openclaw/openclaw", "enabled": true, "priority": true, "topics": [ "ai-agent" ], "note": "Open source AI assistant platform" }, { "id": "go-ethereum-github", "type": "github", "name": "go-ethereum (Geth)", "repo": "ethereum/go-ethereum", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Official Go implementation of Ethereum" }, { "id": "solidity-github", "type": "github", "name": "Solidity", "repo": "ethereum/solidity", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Ethereum smart contract programming language" }, { "id": "foundry-github", "type": "github", "name": "Foundry", "repo": "foundry-rs/foundry", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "Fast, portable and modular Ethereum toolkit" }, { "id": "eips-github", "type": "github", "name": "Ethereum EIPs", "repo": "ethereum/EIPs", "enabled": true, "priority": true, "topics": [ "crypto" ], "note": "Ethereum Improvement Proposals repository" }, { "id": "linux-github", "type": "github", "name": "Linux Kernel", "repo": "torvalds/linux", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Linux kernel source code maintained by Linus Torvalds" }, { "id": "rust-github", "type": "github", "name": "Rust", "repo": "rust-lang/rust", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "The Rust programming language compiler and standard library" }, { "id": "agno-github", "type": "github", "name": "Agno", "repo": "agno-agi/agno", "enabled": true, "priority": true, "topics": [ "ai-agent" ], "note": "Lightweight AI agent framework" }, { "id": "bensbites-rss", "type": "rss", "name": "Ben's Bites", "url": "https://www.bensbites.com/feed", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "the-decoder-rss", "type": "rss", "name": "The Decoder", "url": "https://the-decoder.com/feed/", "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ] }, { "id": "a16zcrypto-rss", "type": "rss", "name": "a16z Crypto", "url": "https://a16zcrypto.substack.com/feed", "enabled": true, "priority": true, "topics": [ "crypto" ] }, { "id": "bankless-rss", "type": "rss", "name": "Bankless", "url": "https://newsletter.banklesshq.com/feed", "enabled": true, "priority": true, "topics": [ "crypto" ] }, { "id": "twitter-clementdelangue", "type": "twitter", "name": "Clement Delangue", "handle": "ClementDelangue", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "twitter-gaborhm", "type": "twitter", "name": "Greg Brockman", "handle": "GregBrockman", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "twitter-zuck", "type": "twitter", "name": "Mark Zuckerberg", "handle": "finkd", "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "github-mcp-servers", "type": "github", "name": "MCP Servers", "repo": "modelcontextprotocol/servers", "enabled": true, "priority": true, "topics": [ "ai-agent" ] }, { "id": "github-deepseek-v3", "type": "github", "name": "DeepSeek V3", "repo": "deepseek-ai/DeepSeek-V3", "enabled": true, "priority": true, "topics": [ "llm" ] }, { "id": "github-meta-llama", "type": "github", "name": "Meta Llama", "repo": "meta-llama/llama-models", "enabled": true, "priority": true, "topics": [ "llm" ] }, { "id": "reddit-machinelearning", "type": "reddit", "name": "r/MachineLearning", "subreddit": "MachineLearning", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "reddit-localllama", "type": "reddit", "name": "r/LocalLLaMA", "subreddit": "LocalLLaMA", "sort": "hot", "limit": 25, "min_score": 30, "enabled": true, "priority": false, "topics": [ "llm" ] }, { "id": "reddit-cryptocurrency", "type": "reddit", "name": "r/CryptoCurrency", "subreddit": "CryptoCurrency", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": true, "topics": [ "crypto" ] }, { "id": "reddit-artificial", "type": "reddit", "name": "r/artificial", "subreddit": "artificial", "sort": "hot", "limit": 25, "min_score": 30, "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ] }, { "id": "reddit-ethereum", "type": "reddit", "name": "r/ethereum", "subreddit": "ethereum", "sort": "hot", "limit": 25, "min_score": 30, "enabled": true, "priority": false, "topics": [ "crypto" ] }, { "id": "reddit-chatgpt", "type": "reddit", "name": "r/ChatGPT", "subreddit": "ChatGPT", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "reddit-singularity", "type": "reddit", "name": "r/singularity", "subreddit": "singularity", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": false, "topics": [ "llm" ] }, { "id": "reddit-openai", "type": "reddit", "name": "r/OpenAI", "subreddit": "OpenAI", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "reddit-bitcoin", "type": "reddit", "name": "r/Bitcoin", "subreddit": "Bitcoin", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": true, "topics": [ "crypto" ] }, { "id": "reddit-programming", "type": "reddit", "name": "r/programming", "subreddit": "programming", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": false, "topics": [ "llm", "ai-agent", "crypto" ] }, { "id": "reddit-anthropic", "type": "reddit", "name": "r/Anthropic", "subreddit": "Anthropic", "sort": "hot", "limit": 25, "min_score": 30, "enabled": true, "priority": true, "topics": [ "llm", "ai-agent" ] }, { "id": "reddit-defi", "type": "reddit", "name": "r/defi", "subreddit": "defi", "sort": "hot", "limit": 25, "min_score": 30, "enabled": true, "priority": false, "topics": [ "crypto" ] }, { "id": "reddit-experienceddevs", "type": "reddit", "name": "r/ExperiencedDevs", "subreddit": "ExperiencedDevs", "sort": "hot", "limit": 25, "min_score": 50, "enabled": true, "priority": false, "topics": [ "llm", "ai-agent" ] }, { "id": "openclaw-twitter", "type": "twitter", "name": "OpenClaw", "handle": "OpenClawAI", "enabled": true, "priority": true, "topics": [ "ai-agent" ], "note": "Official OpenClaw account" }, { "id": "steipete-twitter", "type": "twitter", "name": "Peter Steinberger", "handle": "steipete", "enabled": true, "priority": true, "topics": [ "ai-agent", "frontier-tech" ], "note": "OpenClaw creator, now at OpenAI" }, { "id": "mem0-github", "type": "github", "repo": "mem0ai/mem0", "name": "Mem0", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "Memory layer for AI agents" }, { "id": "openviking-github", "type": "github", "repo": "volcengine/OpenViking", "name": "OpenViking", "enabled": true, "priority": false, "topics": [ "llm" ], "note": "Volcengine/ByteDance open-source LLM" }, { "id": "moltworker-github", "type": "github", "name": "Cloudflare MoltWorker", "repo": "cloudflare/moltworker", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Cloudflare MoltWorker project" }, { "id": "picoclaw-github", "type": "github", "name": "Sipeed PicoClaw", "repo": "sipeed/picoclaw", "enabled": true, "priority": false, "topics": [ "frontier-tech" ], "note": "Sipeed PicoClaw embedded AI project" }, { "id": "nanobot-github", "type": "github", "name": "HKUDS NanoBot", "repo": "HKUDS/nanobot", "enabled": true, "priority": false, "topics": [ "ai-agent" ], "note": "HKUDS NanoBot AI agent project" }, { "id": "zeroclaw-github", "type": "github", "name": "ZeroClaw", "repo": "zeroclaw-labs/zeroclaw", "enabled": true, "priority": false, "topics": [ "crypto" ], "note": "ZeroClaw Labs project" } ] }

File v3.11.0:config/defaults/topics.json

{ "_description": "Enhanced topic definitions for tech digest. Each topic defines a report section with search queries, filters, and display preferences.", "_updated": "2026-02-15", "_version": "2.5.0", "topics": [ { "id": "llm", "emoji": "🧠", "label": "LLM / Large Models", "description": "Large Language Models, foundation models, model releases, benchmarks, and breakthroughs in generative AI", "search": { "queries": ["LLM latest news", "large language model breakthroughs", "大模型最新动态", "GPT Claude Gemini updates"], "twitter_queries": ["GPT-5", "Claude", "大模型"], "must_include": ["LLM", "large language model", "foundation model", "language model", "大模型"], "exclude": ["tutorial", "how to use", "beginner guide"] }, "display": { "max_items": 8, "style": "detailed" } }, { "id": "ai-agent", "emoji": "🤖", "label": "AI Agent", "description": "Autonomous agents, agent frameworks, AI assistants, and agentic AI systems", "search": { "queries": ["AI Agent latest developments", "autonomous agent framework", "AI assistant breakthrough"], "twitter_queries": ["AI agent", "autonomous agent", "AI 智能体"], "must_include": ["AI agent", "autonomous agent", "agent framework", "agentic", "multi-agent"], "exclude": ["game agent", "travel agent"] }, "display": { "max_items": 6, "style": "compact" } }, { "id": "crypto", "emoji": "💰", "label": "Cryptocurrency", "description": "Bitcoin, Ethereum, DeFi, NFTs, blockchain technology, and crypto market developments", "search": { "queries": ["cryptocurrency bitcoin ethereum latest news", "加密货币最新新闻", "DeFi breakthrough", "blockchain development"], "twitter_queries": ["Bitcoin", "Ethereum", "加密货币"], "must_include": ["crypto", "bitcoin", "ethereum", "blockchain", "DeFi", "NFT", "web3"], "exclude": ["scam", "pump dump", "get rich quick"] }, "display": { "max_items": 6, "style": "compact" } }, { "id": "frontier-tech", "emoji": "🔬", "label": "Frontier Tech", "description": "Cutting-edge technology, research breakthroughs, quantum computing, biotech, and emerging technologies", "search": { "queries": ["artificial intelligence breakthroughs", "frontier technology latest", "quantum computing progress", "biotech breakthrough"], "twitter_queries": ["AI breakthrough", "量子计算", "机器人"], "must_include": ["breakthrough", "research", "technology", "innovation", "quantum", "biotech", "robotics"], "exclude": ["rumor", "speculation", "unverified"] }, "display": { "max_items": 8, "style": "detailed" } } ] }

Archive v3.10.3: 29 files, 96907 bytes

Files: CHANGELOG.md (16948b), config/defaults/sources.json (38686b), config/defaults/topics.json (2951b), config/schema.json (4561b), CONTRIBUTING.md (2283b), README_CN.md (2950b), README.md (3198b), references/digest-prompt.md (5549b), references/templates/discord.md (2879b), references/templates/email.md (5783b), references/templates/pdf.md (2119b), requirements.txt (606b), scripts/config_loader.py (8490b), scripts/fetch-github.py (20476b), scripts/fetch-reddit.py (12211b), scripts/fetch-rss.py (19352b), scripts/fetch-twitter.py (28449b), scripts/fetch-web.py (20565b), scripts/generate-pdf.py (9751b), scripts/merge-sources.py (24247b), scripts/run-pipeline.py (9960b), scripts/sanitize-html.py (7357b), scripts/send-email.py (5199b), scripts/source-health.py (4931b), scripts/summarize-merged.py (3526b), scripts/test-pipeline.sh (10975b), scripts/validate-config.py (9659b), SKILL.md (20410b), _meta.json (136b)

File v3.10.3:SKILL.md


name: tech-news-digest description: Generate tech news digests with unified source model, quality scoring, and multi-format output. Five-layer data collection from RSS feeds, Twitter/X KOLs, GitHub releases, Reddit, and web search. Pipeline-based scripts with retry mechanisms and deduplication. Supports Discord, email, and markdown templates. version: "3.10.3" homepage: https://github.com/draco-agent/tech-news-digest source: https://github.com/draco-agent/tech-news-digest metadata: openclaw: requires: bins: ["python3"] optionalBins: ["mail", "msmtp", "gog", "gh", "openssl", "weasyprint"] env:

  • name: TWITTER_API_BACKEND required: false description: "Twitter API backend: 'official', 'twitterapiio', or 'auto' (default: auto)"
  • name: X_BEARER_TOKEN required: false description: Twitter/X API bearer token for KOL monitoring (official backend)
  • name: TWITTERAPI_IO_KEY required: false description: twitterapi.io API key for KOL monitoring (twitterapiio backend)
  • name: BRAVE_API_KEYS required: false description: Brave Search API keys (comma-separated for rotation)
  • name: BRAVE_API_KEY required: false description: Brave Search API key (single key fallback)
  • name: GITHUB_TOKEN required: false description: GitHub token for higher API rate limits (auto-generated from GitHub App if not set)
  • name: GH_APP_ID required: false description: GitHub App ID for automatic installation token generation
  • name: GH_APP_INSTALL_ID required: false description: GitHub App Installation ID for automatic token generation
  • name: GH_APP_KEY_FILE required: false description: Path to GitHub App private key PEM file tools:
  • python3: Required. Runs data collection and merge scripts.
  • mail: Optional. msmtp-based mail command for email delivery (preferred).
  • gog: Optional. Gmail CLI for email delivery (fallback if mail not available). files: read:
    • config/defaults/: Default source and topic configurations
    • references/: Prompt templates and output templates
    • scripts/: Python pipeline scripts
    • <workspace>/archive/tech-news-digest/: Previous digests for dedup write:
    • /tmp/td-*.json: Temporary pipeline intermediate outputs
    • /tmp/td-email.html: Temporary email HTML body
    • /tmp/td-digest.pdf: Generated PDF digest
    • <workspace>/archive/tech-news-digest/: Saved digest archives

Tech News Digest

Automated tech news digest system with unified data source model, quality scoring pipeline, and template-based output generation.

Quick Start

  1. Configuration Setup: Default configs are in config/defaults/. Copy to workspace for customization:

    mkdir -p workspace/config
    cp config/defaults/sources.json workspace/config/tech-news-digest-sources.json
    cp config/defaults/topics.json workspace/config/tech-news-digest-topics.json
    
  2. Environment Variables:

    • TWITTERAPI_IO_KEY - twitterapi.io API key (optional, preferred)
    • X_BEARER_TOKEN - Twitter/X official API bearer token (optional, fallback)
    • BRAVE_API_KEYS - Brave Search API keys, comma-separated for rotation (optional)
    • BRAVE_API_KEY - Single Brave key fallback (optional)
    • GITHUB_TOKEN - GitHub personal access token (optional, improves rate limits)
  3. Generate Digest:

    # Unified pipeline (recommended) — runs all 5 sources in parallel + merge
    python3 scripts/run-pipeline.py \
      --defaults config/defaults \
      --config workspace/config \
      --hours 48 --freshness pd \
      --archive-dir workspace/archive/tech-news-digest/ \
      --output /tmp/td-merged.json --verbose --force
    
  4. Use Templates: Apply Discord, email, or PDF templates to merged output

Configuration Files

sources.json - Unified Data Sources

{
  "sources": [
    {
      "id": "openai-rss",
      "type": "rss",
      "name": "OpenAI Blog",
      "url": "https://openai.com/blog/rss.xml",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "ai-agent"],
      "note": "Official OpenAI updates"
    },
    {
      "id": "sama-twitter",
      "type": "twitter", 
      "name": "Sam Altman",
      "handle": "sama",
      "enabled": true,
      "priority": true,
      "topics": ["llm", "frontier-tech"],
      "note": "OpenAI CEO"
    }
  ]
}

topics.json - Enhanced Topic Definitions

{
  "topics": [
    {
      "id": "llm",
      "emoji": "🧠",
      "label": "LLM / Large Models",
      "description": "Large Language Models, foundation models, breakthroughs",
      "search": {
        "queries": ["LLM latest news", "large language model breakthroughs"],
        "must_include": ["LLM", "large language model", "foundation model"],
        "exclude": ["tutorial", "beginner guide"]
      },
      "display": {
        "max_items": 8,
        "style": "detailed"
      }
    }
  ]
}

Scripts Pipeline

run-pipeline.py - Unified Pipeline (Recommended)

python3 scripts/run-pipeline.py \
  --defaults config/defaults [--config CONFIG_DIR] \
  --hours 48 --freshness pd \
  --archive-dir workspace/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force
  • Features: Runs all 5 fetch steps in parallel, then merges + deduplicates + scores
  • Output: Final merged JSON ready for report generation (~30s total)
  • Metadata: Saves per-step timing and counts to *.meta.json
  • GitHub Auth: Auto-generates GitHub App token if $GITHUB_TOKEN not set
  • Fallback: If this fails, run individual scripts below

Individual Scripts (Fallback)

fetch-rss.py - RSS Feed Fetcher

python3 scripts/fetch-rss.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--verbose]
  • Parallel fetching (10 workers), retry with backoff, feedparser + regex fallback
  • Timeout: 30s per feed, ETag/Last-Modified caching

fetch-twitter.py - Twitter/X KOL Monitor

python3 scripts/fetch-twitter.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE] [--backend auto|official|twitterapiio]
  • Backend auto-detection: uses twitterapi.io if TWITTERAPI_IO_KEY set, else official X API v2 if X_BEARER_TOKEN set
  • Rate limit handling, engagement metrics, retry with backoff

fetch-web.py - Web Search Engine

python3 scripts/fetch-web.py [--defaults DIR] [--config DIR] [--freshness pd] [--output FILE]
  • Auto-detects Brave API rate limit: paid plans → parallel queries, free → sequential
  • Without API: generates search interface for agents

fetch-github.py - GitHub Releases Monitor

python3 scripts/fetch-github.py [--defaults DIR] [--config DIR] [--hours 168] [--output FILE]
  • Parallel fetching (10 workers), 30s timeout
  • Auth priority: $GITHUB_TOKEN → GitHub App auto-generate → gh CLI → unauthenticated (60 req/hr)

fetch-reddit.py - Reddit Posts Fetcher

python3 scripts/fetch-reddit.py [--defaults DIR] [--config DIR] [--hours 48] [--output FILE]
  • Parallel fetching (4 workers), public JSON API (no auth required)
  • 13 subreddits with score filtering

merge-sources.py - Quality Scoring & Deduplication

python3 scripts/merge-sources.py --rss FILE --twitter FILE --web FILE --github FILE --reddit FILE
  • Quality scoring, title similarity dedup (85%), previous digest penalty
  • Output: topic-grouped articles sorted by score

validate-config.py - Configuration Validator

python3 scripts/validate-config.py [--defaults DIR] [--config DIR] [--verbose]
  • JSON schema validation, topic reference checks, duplicate ID detection

generate-pdf.py - PDF Report Generator

python3 scripts/generate-pdf.py --input report.md --output digest.pdf [--verbose]
  • Converts markdown digest to styled A4 PDF with Chinese typography (Noto Sans CJK SC)
  • Emoji icons, page headers/footers, blue accent theme. Requires weasyprint.

sanitize-html.py - Safe HTML Email Converter

python3 scripts/sanitize-html.py --input report.md --output email.html [--verbose]
  • Converts markdown to XSS-safe HTML email with inline CSS
  • URL whitelist (http/https only), HTML-escaped text content

source-health.py - Source Health Monitor

python3 scripts/source-health.py --rss FILE --twitter FILE --github FILE --reddit FILE --web FILE [--verbose]
  • Tracks per-source success/failure history over 7 days
  • Reports unhealthy sources (>50% failure rate)

summarize-merged.py - Merged Data Summary

python3 scripts/summarize-merged.py --input merged.json [--top N] [--topic TOPIC]
  • Human-readable summary of merged data for LLM consumption
  • Shows top articles per topic with scores and metrics

User Customization

Workspace Configuration Override

Place custom configs in workspace/config/ to override defaults:

  • Sources: Append new sources, disable defaults with "enabled": false
  • Topics: Override topic definitions, search queries, display settings
  • Merge Logic:
    • Sources with same id → user version takes precedence
    • Sources with new id → appended to defaults
    • Topics with same id → user version completely replaces default

Example Workspace Override

// workspace/config/tech-news-digest-sources.json
{
  "sources": [
    {
      "id": "simonwillison-rss",
      "enabled": false,
      "note": "Disabled: too noisy for my use case"
    },
    {
      "id": "my-custom-blog", 
      "type": "rss",
      "name": "My Custom Tech Blog",
      "url": "https://myblog.com/rss",
      "enabled": true,
      "priority": true,
      "topics": ["frontier-tech"]
    }
  ]
}

Templates & Output

Discord Template (references/templates/discord.md)

  • Bullet list format with link suppression (<link>)
  • Mobile-optimized, emoji headers
  • 2000 character limit awareness

Email Template (references/templates/email.md)

  • Rich metadata, technical stats, archive links
  • Executive summary, top articles section
  • HTML-compatible formatting

PDF Template (references/templates/pdf.md)

  • A4 layout with Noto Sans CJK SC font for Chinese support
  • Emoji icons, page headers/footers with page numbers
  • Generated via scripts/generate-pdf.py (requires weasyprint)

Default Sources (138 total)

  • RSS Feeds (49): AI labs, tech blogs, crypto news, Chinese tech media
  • Twitter/X KOLs (48): AI researchers, crypto leaders, tech executives
  • GitHub Repos (28): Major open-source projects (LangChain, vLLM, DeepSeek, Llama, etc.)
  • Reddit (13): r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency, r/ChatGPT, r/OpenAI, etc.
  • Web Search (4 topics): LLM, AI Agent, Crypto, Frontier Tech

All sources pre-configured with appropriate topic tags and priority levels.

Dependencies

pip install -r requirements.txt

Optional but Recommended:

  • feedparser>=6.0.0 - Better RSS parsing (fallback to regex if unavailable)
  • jsonschema>=4.0.0 - Configuration validation

All scripts work with Python 3.8+ standard library only.

Monitoring & Operations

Health Checks

# Validate configuration
python3 scripts/validate-config.py --verbose

# Test RSS feeds
python3 scripts/fetch-rss.py --hours 1 --verbose

# Check Twitter API
python3 scripts/fetch-twitter.py --hours 1 --verbose

Archive Management

  • Digests automatically archived to <workspace>/archive/tech-news-digest/
  • Previous digest titles used for duplicate detection
  • Old archives cleaned automatically (90+ days)

Error Handling

  • Network Failures: Retry with exponential backoff
  • Rate Limits: Automatic retry with appropriate delays
  • Invalid Content: Graceful degradation, detailed logging
  • Configuration Errors: Schema validation with helpful messages

API Keys & Environment

Set in ~/.zshenv or similar:

# Twitter (at least one required for Twitter source)
export TWITTERAPI_IO_KEY="your_key"        # twitterapi.io key (preferred)
export X_BEARER_TOKEN="your_bearer_token"  # Official X API v2 (fallback)
export TWITTER_API_BACKEND="auto"          # auto|twitterapiio|official (default: auto)

# Brave Search (optional, enables web search layer)
export BRAVE_API_KEYS="key1,key2,key3"     # Multiple keys, comma-separated rotation
export BRAVE_API_KEY="key1"                # Single key fallback
export BRAVE_PLAN="free"                   # Override rate limit detection: free|pro

# GitHub (optional, improves rate limits)
export GITHUB_TOKEN="ghp_xxx"              # PAT (simplest)
export GH_APP_ID="12345"                   # Or use GitHub App for auto-token
export GH_APP_INSTALL_ID="67890"
export GH_APP_KEY_FILE="/path/to/key.pem"
  • Twitter: TWITTERAPI_IO_KEY preferred ($3-5/mo); X_BEARER_TOKEN as fallback; auto mode tries twitterapiio first
  • Brave Search: Optional, fallback to agent web_search if unavailable
  • GitHub: Auto-generates token from GitHub App if PAT not set; unauthenticated fallback (60 req/hr)
  • Reddit: No API key needed (uses public JSON API)

Cron / Scheduled Task Integration

OpenClaw Cron (Recommended)

The cron prompt should NOT hardcode the pipeline steps. Instead, reference references/digest-prompt.md and only pass configuration parameters. This ensures the pipeline logic stays in the skill repo and is consistent across all installations.

Daily Digest Cron Prompt

Read <SKILL_DIR>/references/digest-prompt.md and follow the complete workflow to generate a daily digest.

Replace placeholders with:
- MODE = daily
- TIME_WINDOW = past 1-2 days
- FRESHNESS = pd
- RSS_HOURS = 48
- ITEMS_PER_SECTION = 3-5
- BLOG_PICKS_COUNT = 2-3
- EXTRA_SECTIONS = (none)
- SUBJECT = Daily Tech Digest - YYYY-MM-DD
- WORKSPACE = <your workspace path>
- SKILL_DIR = <your skill install path>
- DISCORD_CHANNEL_ID = <your channel id>
- EMAIL = (optional)
- LANGUAGE = English
- TEMPLATE = discord

Follow every step in the prompt template strictly. Do not skip any steps.

Weekly Digest Cron Prompt

Read <SKILL_DIR>/references/digest-prompt.md and follow the complete workflow to generate a weekly digest.

Replace placeholders with:
- MODE = weekly
- TIME_WINDOW = past 7 days
- FRESHNESS = pw
- RSS_HOURS = 168
- ITEMS_PER_SECTION = 5-8
- BLOG_PICKS_COUNT = 3-5
- EXTRA_SECTIONS = 📊 Weekly Trend Summary (2-3 sentences summarizing macro trends)
- SUBJECT = Weekly Tech Digest - YYYY-MM-DD
- WORKSPACE = <your workspace path>
- SKILL_DIR = <your skill install path>
- DISCORD_CHANNEL_ID = <your channel id>
- EMAIL = (optional)
- LANGUAGE = English
- TEMPLATE = discord

Follow every step in the prompt template strictly. Do not skip any steps.

Why This Pattern?

  • Single source of truth: Pipeline logic lives in digest-prompt.md, not scattered across cron configs
  • Portable: Same skill on different OpenClaw instances, just change paths and channel IDs
  • Maintainable: Update the skill → all cron jobs pick up changes automatically
  • Anti-pattern: Do NOT copy pipeline steps into the cron prompt — it will drift out of sync

Multi-Channel Delivery Limitation

OpenClaw enforces cross-provider isolation: a single session can only send messages to one provider (e.g., Discord OR Telegram, not both). If you need to deliver digests to multiple platforms, create separate cron jobs for each provider:

# Job 1: Discord + Email
- DISCORD_CHANNEL_ID = <your-discord-channel-id>
- EMAIL = user@example.com
- TEMPLATE = discord

# Job 2: Telegram DM
- DISCORD_CHANNEL_ID = (none)
- EMAIL = (none)
- TEMPLATE = telegram

Replace DISCORD_CHANNEL_ID delivery with the target platform's delivery in the second job's prompt.

This is a security feature, not a bug — it prevents accidental cross-context data leakage.

Security Notes

Execution Model

This skill uses a prompt template pattern: the agent reads digest-prompt.md and follows its instructions. This is the standard OpenClaw skill execution model — the agent interprets structured instructions from skill-provided files. All instructions are shipped with the skill bundle and can be audited before installation.

Network Access

The Python scripts make outbound requests to:

  • RSS feed URLs (configured in tech-news-digest-sources.json)
  • Twitter/X API (api.x.com or api.twitterapi.io)
  • Brave Search API (api.search.brave.com)
  • GitHub API (api.github.com)
  • Reddit JSON API (reddit.com)

No data is sent to any other endpoints. All API keys are read from environment variables declared in the skill metadata.

Shell Safety

Email delivery uses send-email.py which constructs proper MIME multipart messages with HTML body + optional PDF attachment. Subject formats are hardcoded (Daily Tech Digest - YYYY-MM-DD). PDF generation uses generate-pdf.py via weasyprint. The prompt template explicitly prohibits interpolating untrusted content (article titles, tweet text, etc.) into shell arguments. Email addresses and subjects must be static placeholder values only.

File Access

Scripts read from config/ and write to workspace/archive/. No files outside the workspace are accessed.

Support & Troubleshooting

Common Issues

  1. RSS feeds failing: Check network connectivity, use --verbose for details
  2. Twitter rate limits: Reduce sources or increase interval
  3. Configuration errors: Run validate-config.py for specific issues
  4. No articles found: Check time window (--hours) and source enablement

Debug Mode

All scripts support --verbose flag for detailed logging and troubleshooting.

Performance Tuning

  • Parallel Workers: Adjust MAX_WORKERS in scripts for your system
  • Timeout Settings: Increase TIMEOUT for slow networks
  • Article Limits: Adjust MAX_ARTICLES_PER_FEED based on needs

Security Considerations

Shell Execution

The digest prompt instructs agents to run Python scripts via shell commands. All script paths and arguments are skill-defined constants — no user input is interpolated into commands. Two scripts use subprocess:

  • run-pipeline.py orchestrates child fetch scripts (all within scripts/ directory)
  • fetch-github.py has two subprocess calls:
    1. openssl dgst -sha256 -sign for JWT signing (only if GH_APP_* env vars are set — signs a self-constructed JWT payload, no user content involved)
    2. gh auth token CLI fallback (only if gh is installed — reads from gh's own credential store)

No user-supplied or fetched content is ever interpolated into subprocess arguments. Email delivery uses send-email.py which builds MIME messages programmatically — no shell interpolation. PDF generation uses generate-pdf.py via weasyprint. Email subjects are static format strings only — never constructed from fetched data.

Credential & File Access

Scripts do not directly read ~/.config/, ~/.ssh/, or any credential files. All API tokens are read from environment variables declared in the skill metadata. The GitHub auth cascade is:

  1. $GITHUB_TOKEN env var (you control what to provide)
  2. GitHub App token generation (only if you set GH_APP_ID, GH_APP_INSTALL_ID, and GH_APP_KEY_FILE — uses inline JWT signing via openssl CLI, no external scripts involved)
  3. gh auth token CLI (delegates to gh's own secure credential store)
  4. Unauthenticated (60 req/hr, safe fallback)

If you prefer no automatic credential discovery, simply set $GITHUB_TOKEN and the script will use it directly without attempting steps 2-3.

Dependency Installation

This skill does not install any packages. requirements.txt lists optional dependencies (feedparser, jsonschema) for reference only. All scripts work with Python 3.8+ standard library. Users should install optional deps in a virtualenv if desired — the skill never runs pip install.

Input Sanitization

  • URL resolution rejects non-HTTP(S) schemes (javascript:, data:, etc.)
  • RSS fallback parsing uses simple, non-backtracking regex patterns (no ReDoS risk)
  • All fetched content is treated as untrusted data for display only

Network Access

Scripts make outbound HTTP requests to configured RSS feeds, Twitter API, GitHub API, Reddit JSON API, and Brave Search API. No inbound connections or listeners are created.

File v3.10.3:README.md

Tech News Digest

Automated tech news digest — 138 sources, 5-layer pipeline, one chat message to install.

Python 3.8+ MIT License

💬 Install in One Message

Tell your OpenClaw AI assistant:

"Install tech-news-digest and send a daily digest to #tech-news every morning at 9am"

That's it. Your bot handles installation, configuration, scheduling, and delivery — all through conversation.

More examples:

🗣️ "Set up a weekly AI digest, only LLM and AI Agent topics, deliver to Discord #ai-weekly every Monday"

🗣️ "Install tech-news-digest, add my RSS feeds, and send crypto news to Telegram"

🗣️ "Give me a tech digest right now, skip Twitter sources"

Or install via CLI:

clawhub install tech-news-digest

📊 What You Get

A quality-scored, deduplicated tech digest built from 138 sources:

| Layer | Sources | What | |-------|---------|------| | 📡 RSS | 49 feeds | OpenAI, Anthropic, Ben's Bites, HN, 36氪, CoinDesk… | | 🐦 Twitter/X | 48 KOLs | @karpathy, @VitalikButerin, @sama, @elonmusk… | | 🔍 Web Search | 4 topics | Brave Search API with freshness filters | | 🐙 GitHub | 28 repos | Releases from key projects (LangChain, vLLM, DeepSeek, Llama…) | | 🗣️ Reddit | 13 subs | r/MachineLearning, r/LocalLLaMA, r/CryptoCurrency… |

Pipeline

       run-pipeline.py (~30s)
              ↓
  RSS ─┐
  Twitter ─┤
  Web ─────┤── parallel fetch ──→ merge-sources.py
  GitHub ──┤
  Reddit ──┘
              ↓
  Quality Scoring → Deduplication → Topic Grouping
              ↓
    Discord / Email / PDF output

Quality scoring: priority source (+3), multi-source cross-ref (+5), recency (+2), engagement (+1), Reddit score bonus (+1/+3/+5), already reported (-5).

⚙️ Configuration

  • config/defaults/sources.json — 138 built-in sources
  • config/defaults/topics.json — 4 topics with search queries & Twitter queries
  • User overrides in workspace/config/ take priority

🔧 Optional Setup

All environment variables are optional. The pipeline runs with whatever sources are available.

export TWITTERAPI_IO_KEY="..."  # twitterapi.io (~$5/mo) — enables Twitter layer
export X_BEARER_TOKEN="..."     # Twitter/X official API — alternative Twitter backend
export BRAVE_API_KEYS="k1,k2,k3" # Brave Search API keys (comma-separated, rotation)
export BRAVE_API_KEY="..."       # Fallback: single Brave key
export GITHUB_TOKEN="..."       # GitHub API — higher rate limits (auto-generated from GitHub App if unset)
export TWITTER_API_BACKEND="auto" # auto|twitterapiio|official (default: auto)
export BRAVE_PLAN="free"         # Override Brave rate limit detection: free|pro
pip install weasyprint           # Enables PDF report generation

📂 Repository

GitHub: github.com/draco-agent/tech-news-digest

📄 License

MIT License — see LICENSE for details.

File v3.10.3:_meta.json

{ "ownerId": "kn74589cx1nbhnc3x0f3nwre39814699", "slug": "tech-news-digest", "version": "3.10.3", "publishedAt": 1772207168147 }

File v3.10.3:references/digest-prompt.md

Digest Prompt Template

Replace <...> placeholders before use. Daily defaults shown; weekly overrides in parentheses.

Placeholders

| Placeholder | Default | Weekly Override | |-------------|---------|----------------| | <MODE> | daily | weekly | | <TIME_WINDOW> | past 1-2 days | past 7 days | | <FRESHNESS> | pd | pw | | <RSS_HOURS> | 48 | 168 | | <ITEMS_PER_SECTION> | 3-5 | 5-8 | | <BLOG_PICKS_COUNT> | 2-3 | 3-5 | | <EXTRA_SECTIONS> | (none) | 📊 Weekly Trend Summary | | <SUBJECT> | Daily Tech Digest - YYYY-MM-DD | Weekly Tech Digest - YYYY-MM-DD | | <WORKSPACE> | Your workspace path | | | <SKILL_DIR> | Installed skill directory | | | <DISCORD_CHANNEL_ID> | Target channel ID | | | <EMAIL> | (optional) Recipient email | | | <EMAIL_FROM> | (optional) e.g. MyBot <bot@example.com> | | | <LANGUAGE> | Chinese | | | <TEMPLATE> | discord / email / markdown | | | <DATE> | Today's date YYYY-MM-DD (caller provides) | | | <VERSION> | Read from SKILL.md frontmatter | |


Generate the <MODE> tech digest for <DATE>. Use <DATE> as the report date — do NOT infer it.

Configuration

Read config files (workspace overrides take priority over defaults):

  1. Sources: <WORKSPACE>/config/tech-news-digest-sources.json → fallback <SKILL_DIR>/config/defaults/sources.json
  2. Topics: <WORKSPACE>/config/tech-news-digest-topics.json → fallback <SKILL_DIR>/config/defaults/topics.json

Context: Previous Report

Read the most recent file from <WORKSPACE>/archive/tech-news-digest/ to avoid repeats and follow up on developing stories. Skip if none exists.

Data Collection Pipeline

Use the unified pipeline (runs all 5 sources in parallel, ~30s):

python3 <SKILL_DIR>/scripts/run-pipeline.py \
  --defaults <SKILL_DIR>/config/defaults \
  --config <WORKSPACE>/config \
  --hours <RSS_HOURS> --freshness <FRESHNESS> \
  --archive-dir <WORKSPACE>/archive/tech-news-digest/ \
  --output /tmp/td-merged.json --verbose --force

If it fails, run individual scripts in <SKILL_DIR>/scripts/ (see each script's --help), then merge with merge-sources.py.

Report Generation

Get a structured overview:

python3 <SKILL_DIR>/scripts/summarize-merged.py --input /tmp/td-merged.json --top <ITEMS_PER_SECTION>

Use this output to select articles — do NOT write ad-hoc Python to parse the JSON. Apply the template from <SKILL_DIR>/references/templates/<TEMPLATE>.md.

Select articles purely by quality_score regardless of source type. For Reddit posts, append *[Reddit r/xxx, {{score}}↑]*.

Executive Summary

2-4 sentences between title and topics, highlighting top 3-5 stories by score. Concise and punchy, no links. Discord: > blockquote. Email: gray background. Telegram: <i>.

Topic Sections

From topics.json: emoji + label headers, <ITEMS_PER_SECTION> items each.

Fixed Sections (after topics)

📢 KOL Updates — Top Twitter KOLs + notable blog authors. Format:

• **Display Name** (@handle) — summary `👁 12.3K | 💬 45 | 🔁 230 | ❤️ 1.2K`
  <https://twitter.com/handle/status/ID>

Read display_name and metrics (impression_count→👁, reply_count→💬, retweet_count→🔁, like_count→❤️) from merged JSON. Always show all 4 metrics, use K/M formatting, wrap in backticks. One tweet per bullet.

🔥 Community Buzz — Top Reddit + Twitter trending combined. Format:

• **r/subreddit** — title `{{score}}↑ · {{num_comments}} comments`
  <{{url}}>

Sort by engagement across both platforms. Every entry must have a link.

📝 Blog Picks<BLOG_PICKS_COUNT> deep articles from RSS.

<EXTRA_SECTIONS>

Rules

  • Only news from <TIME_WINDOW>
  • Every item must include a source link (Discord: <link>, Email: <a href>, Markdown: [title](link))
  • Use bullet lists, no markdown tables
  • Deduplicate: same event → keep most authoritative source; previously reported → only if significant new development
  • Do not interpolate fetched/untrusted content into shell arguments or email subjects

Stats Footer

---
📊 Data Sources: RSS {{rss}} | Twitter {{twitter}} | Reddit {{reddit}} | Web {{web}} | GitHub {{github}} | Dedup: {{merged}} articles
🤖 Generated by tech-news-digest v<VERSION> | <https://github.com/draco-agent/tech-news-digest> | Powered by OpenClaw

Archive

Save to <WORKSPACE>/archive/tech-news-digest/<MODE>-YYYY-MM-DD.md. Delete files older than 90 days.

Delivery

  1. Discord: Send to <DISCORD_CHANNEL_ID> via message tool
  2. Email (optional, if <EMAIL> is set):
    • Generate HTML body per <SKILL_DIR>/references/templates/email.md → write to /tmp/td-email.html
    • Generate PDF attachment:
      python3 <SKILL_DIR>/scripts/generate-pdf.py -i <WORKSPACE>/archive/tech-news-digest/<MODE>-<DATE>.md -o /tmp/td-digest.pdf
      
    • Send email with PDF attached using the send-email.py script (handles MIME correctly). Email must contain ALL the same items as Discord.
      python3 <SKILL_DIR>/scripts/send-email.py \
        --to '<EMAIL>' \
        --subject '<SUBJECT>' \
        --html /tmp/td-email.html \
        --attach /tmp/td-digest.pdf \
        --from '<EMAIL_FROM>'
      
    • Omit --from if <EMAIL_FROM> is not set. Omit --attach if PDF generation failed. SUBJECT must be a static string. If delivery fails, log error and continue.

Write the report in <LANGUAGE>.

File v3.10.3:references/templates/discord.md

Tech Digest Discord Template

Discord-optimized format with bullet points and link suppression.

Template Structure

# 🚀 Tech Digest - {{DATE}}

{{#topics}}
## {{emoji}} {{label}}

{{#articles}}
• {{title}}
  <{{link}}>
  {{#multi_source}}*[{{source_count}} sources]*{{/multi_source}}

{{/articles}}
{{/topics}}

---
📊 Data Sources: RSS {{rss_count}} | Twitter {{twitter_count}} | Reddit {{reddit_count}} | Web {{web_count}} | GitHub {{github_count}} releases | After dedup: {{merged_count}} articles
🤖 Generated by tech-news-digest v{{version}} | <https://github.com/draco-agent/tech-news-digest> | Powered by OpenClaw

Delivery

  • Default: Channel — Send to the Discord channel specified by DISCORD_CHANNEL_ID
  • Use message tool with target set to the channel ID for channel delivery
  • For DM delivery instead, set target to a user ID

Discord-Specific Features

  • Link suppression: Wrap links in <> to prevent embeds
  • Bullet format: Use for clean mobile display
  • No tables: Discord mobile doesn't handle markdown tables well
  • Emoji headers: Visual hierarchy with topic emojis
  • Concise metadata: Source count and multi-source indicators
  • Character limits: Discord messages have 2000 char limit, may need splitting

Example Output

# 🚀 Tech Digest - 2026-02-15

## 🧠 LLM / Large Models

• OpenAI releases GPT-5 with breakthrough reasoning capabilities
  <https://openai.com/blog/gpt5-announcement>
  *[3 sources]*

• Meta's Llama 3.1 achieves new MMLU benchmarks
  <https://ai.meta.com/blog/llama-31-release>

## 🤖 AI Agent

• LangChain launches production-ready agent framework
  <https://blog.langchain.dev/production-agents>

## 💰 Cryptocurrency

• Bitcoin reaches new ATH at $67,000 amid ETF approval
  <https://coindesk.com/markets/btc-ath-etf>
  *[2 sources]*

## 📢 KOL Updates

• **Elon Musk** (@elonmusk) — Confirmed X's crypto trading feature `👁 2.1M | 💬 12.3K | 🔁 8.5K | ❤️ 49.8K`
  <https://twitter.com/elonmusk/status/123456789>
• **@saylor** — Valentine's BTC enthusiasm `👁 450K | 💬 1.2K | 🔁 3.1K | ❤️ 13K`
  <https://twitter.com/saylor/status/987654321>

---
📊 Data Sources: RSS 285 | Twitter 67 | Reddit 45 | Web 60 | GitHub 29 releases | After dedup: 95 articles

Variables

  • {{DATE}} - Report date (YYYY-MM-DD format)
  • {{topics}} - Array of topic objects
  • {{emoji}} - Topic emoji
  • {{label}} - Topic display name
  • {{articles}} - Array of article objects per topic
  • {{title}} - Article title (truncated if needed)
  • {{link}} - Article URL
  • {{multi_source}} - Boolean, true if article from multiple sources
  • {{source_count}} - Number of sources for this article
  • {{total_sources}} - Total number of sources used
  • {{total_articles}} - Total articles in digest

File v3.10.3:references/templates/email.md

Tech Digest Email Template

HTML email format optimized for Gmail/Outlook rendering.

Delivery

Send via gog gmail send with --body-html flag:

gog gmail send --to '<EMAIL>' --subject '<SUBJECT>' --body-html '<HTML_CONTENT>'

Important: Use --body-html, NOT --body. Plain text markdown will not render properly in email clients.

Template Structure

The agent should generate an HTML email body. Use inline styles (email clients strip <style> blocks).

<div style="max-width:640px;margin:0 auto;font-family:-apple-system,BlinkMacSystemFont,'Segoe UI',Roboto,sans-serif;color:#1a1a1a;line-height:1.6">

  <h1 style="font-size:22px;border-bottom:2px solid #e5e5e5;padding-bottom:8px">
    🐉 {{TITLE}}
  </h1>

  <!-- Optional: Executive Summary for weekly -->
  <p style="color:#555;font-size:14px;background:#f8f9fa;padding:12px;border-radius:6px">
    {{SUMMARY}}
  </p>

  <!-- Topic Section -->
  <h2 style="font-size:17px;margin-top:24px;color:#333">{{emoji}} {{label}}</h2>
  <ul style="padding-left:20px">
    <li style="margin-bottom:10px">
      <strong>{{title}}</strong> — {{description}}
      <br><a href="{{link}}" style="color:#0969da;font-size:13px">{{link}}</a>
    </li>
  </ul>

  <!-- Repeat for each topic -->

  <!-- KOL Section: Read metrics from twitter JSON data (metrics.impression_count, reply_

API & Reliability

Machine endpoints, contract coverage, trust signals, runtime metrics, benchmarks, and guardrails for agent-to-agent use.

MissingCLAWHUB

Machine interfaces

Contract & API

Contract coverage

Status

missing

Auth

None

Streaming

No

Data region

Unspecified

Protocol support

OpenClaw: self-declared

Requires: none

Forbidden: none

Guardrails

Operational confidence: low

No positive guardrails captured.
Invocation examples
curl -s "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/snapshot"
curl -s "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/contract"
curl -s "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/trust"

Operational fit

Reliability & Benchmarks

Trust signals

Handshake

UNKNOWN

Confidence

unknown

Attempts 30d

unknown

Fallback rate

unknown

Runtime metrics

Observed P50

unknown

Observed P95

unknown

Rate limit

unknown

Estimated cost

unknown

Do not use if

Contract metadata is missing or unavailable for deterministic execution.
No benchmark suites or observed failure patterns are available.

Machine Appendix

Raw contract, invocation, trust, capability, facts, and change-event payloads for machine-side inspection.

MissingCLAWHUB

Contract JSON

{
  "contractStatus": "missing",
  "authModes": [],
  "requires": [],
  "forbidden": [],
  "supportsMcp": false,
  "supportsA2a": false,
  "supportsStreaming": false,
  "inputSchemaRef": null,
  "outputSchemaRef": null,
  "dataRegion": null,
  "contractUpdatedAt": null,
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Invocation Guide

{
  "preferredApi": {
    "snapshotUrl": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/snapshot",
    "contractUrl": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/contract",
    "trustUrl": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/trust"
  },
  "curlExamples": [
    "curl -s \"https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/snapshot\"",
    "curl -s \"https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/contract\"",
    "curl -s \"https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/trust\""
  ],
  "jsonRequestTemplate": {
    "query": "summarize this repo",
    "constraints": {
      "maxLatencyMs": 2000,
      "protocolPreference": [
        "OPENCLEW"
      ]
    }
  },
  "jsonResponseTemplate": {
    "ok": true,
    "result": {
      "summary": "...",
      "confidence": 0.9
    },
    "meta": {
      "source": "CLAWHUB",
      "generatedAt": "2026-04-17T04:49:50.416Z"
    }
  },
  "retryPolicy": {
    "maxAttempts": 3,
    "backoffMs": [
      500,
      1500,
      3500
    ],
    "retryableConditions": [
      "HTTP_429",
      "HTTP_503",
      "NETWORK_TIMEOUT"
    ]
  }
}

Trust JSON

{
  "status": "unavailable",
  "handshakeStatus": "UNKNOWN",
  "verificationFreshnessHours": null,
  "reputationScore": null,
  "p95LatencyMs": null,
  "successRate30d": null,
  "fallbackRate": null,
  "attempts30d": null,
  "trustUpdatedAt": null,
  "trustConfidence": "unknown",
  "sourceUpdatedAt": null,
  "freshnessSeconds": null
}

Capability Matrix

{
  "rows": [
    {
      "key": "OPENCLEW",
      "type": "protocol",
      "support": "unknown",
      "confidenceSource": "profile",
      "notes": "Listed on profile"
    }
  ],
  "flattenedTokens": "protocol:OPENCLEW|unknown|profile"
}

Facts JSON

[
  {
    "factKey": "vendor",
    "category": "vendor",
    "label": "Vendor",
    "value": "Clawhub",
    "href": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceUrl": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-04-15T00:45:39.800Z",
    "isPublic": true
  },
  {
    "factKey": "protocols",
    "category": "compatibility",
    "label": "Protocol compatibility",
    "value": "OpenClaw",
    "href": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/contract",
    "sourceUrl": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/contract",
    "sourceType": "contract",
    "confidence": "medium",
    "observedAt": "2026-04-15T00:45:39.800Z",
    "isPublic": true
  },
  {
    "factKey": "traction",
    "category": "adoption",
    "label": "Adoption signal",
    "value": "2K downloads",
    "href": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceUrl": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceType": "profile",
    "confidence": "medium",
    "observedAt": "2026-04-15T00:45:39.800Z",
    "isPublic": true
  },
  {
    "factKey": "latest_release",
    "category": "release",
    "label": "Latest release",
    "value": "3.11.0",
    "href": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceUrl": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceType": "release",
    "confidence": "medium",
    "observedAt": "2026-02-28T16:22:08.196Z",
    "isPublic": true
  },
  {
    "factKey": "handshake_status",
    "category": "security",
    "label": "Handshake status",
    "value": "UNKNOWN",
    "href": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/trust",
    "sourceUrl": "https://xpersona.co/api/v1/agents/clawhub-dinstein-tech-news-digest/trust",
    "sourceType": "trust",
    "confidence": "medium",
    "observedAt": null,
    "isPublic": true
  }
]

Change Events JSON

[
  {
    "eventType": "release",
    "title": "Release 3.11.0",
    "description": "Tavily backend, quality scores, domain limit fix, tests, CI",
    "href": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceUrl": "https://clawhub.ai/dinstein/tech-news-digest",
    "sourceType": "release",
    "confidence": "medium",
    "observedAt": "2026-02-28T16:22:08.196Z",
    "isPublic": true
  }
]

Sponsored

Ads related to Tech News Digest and adjacent AI workflows.