finviz-crawler
name: finviz-crawler
by camopel · published 2026-03-22
$ claw add gh:camopel/camopel-finviz-crawler---
name: finviz-crawler
description: Continuous financial news crawler for finviz.com with SQLite storage, article extraction, and query tool. Use when monitoring financial markets, building news digests, or needing a local financial news database. Runs as a background daemon or systemd service.
metadata: {"openclaw":{"requires":{"bins":["python3"]}}}
---
# finviz-crawler
Why This Skill?
📰 **Your own financial news database** — most finance skills just wrap an API for one-shot queries. This skill runs continuously, building a local archive of every headline and article from Finviz. Query your history anytime — no API limits, no missing data.
🆓 **No API key, no subscription** — scrapes finviz.com directly using Crawl4AI + RSS. Bloomberg, Reuters, Yahoo Finance, CNBC articles extracted automatically. Zero cost.
🤖 **Built for AI summarization** — the query tool outputs clean text/JSON optimized for LLM digests. Pair with an OpenClaw cron job for automated morning briefings, evening wrap-ups, or weekly investment summaries.
💾 **Auto-cleanup** — configurable expiry automatically deletes old articles from both the database and disk. Set `--expiry-days 30` to keep a month of history, or `0` to keep everything forever.
🔄 **Daemon architecture** — runs as a background service that starts/stops with OpenClaw. No manual intervention needed after setup. Works with systemd (Linux) and launchd (macOS).
Install
python3 scripts/install.pyWorks on **macOS, Linux, and Windows**. Installs Python packages (`crawl4ai`, `feedparser`), sets up Playwright browsers, creates data directories, and verifies everything.
Manual install
pip install crawl4ai feedparser
crawl4ai-setup # or: python -m playwright install chromiumUsage
Run the crawler
# Default: ~/workspace/finviz/, 7-day expiry
python3 scripts/finviz_crawler.py
# Custom paths and settings
python3 scripts/finviz_crawler.py --db /path/to/finviz.db --articles-dir /path/to/articles/
# Keep 30 days of articles
python3 scripts/finviz_crawler.py --expiry-days 30
# Never auto-delete (keep everything)
python3 scripts/finviz_crawler.py --expiry-days 0
# Custom crawl interval (default: 300s)
python3 scripts/finviz_crawler.py --sleep 600Query articles
# Last 24 hours of headlines
python3 scripts/finviz_query.py --hours 24
# Titles only (compact, good for LLM summarization)
python3 scripts/finviz_query.py --hours 12 --titles-only
# With full article content
python3 scripts/finviz_query.py --hours 12 --with-content
# List downloaded articles with content status
python3 scripts/finviz_query.py --list-articles --hours 24
# Database stats
python3 scripts/finviz_query.py --statsManage tickers
# List all tracked tickers
python3 scripts/finviz_query.py --list-tickers
# Add single ticker (auto-generates keywords from symbol)
python3 scripts/finviz_query.py --add-ticker NVDA
# Add with custom keywords
python3 scripts/finviz_query.py --add-ticker "NVDA:nvidia,jensen huang"
# Add multiple tickers (batch)
python3 scripts/finviz_query.py --add-ticker NVDA TSLA AAPL
python3 scripts/finviz_query.py --add-ticker "NVDA:nvidia,jensen" "TSLA:tesla,elon musk"
# Remove tickers (batch)
python3 scripts/finviz_query.py --remove-ticker NVDA TSLA
# Custom DB path
python3 scripts/finviz_query.py --list-tickers --db /path/to/finviz.dbTickers are stored in the `tickers` table inside `finviz.db` alongside articles. The crawler reads this table each cycle to know which ticker pages to scrape.
Configuration
| Setting | CLI flag | Env var | Default |
|---------|----------|---------|---------|
| Database path | `--db` | — | `~/workspace/finviz/finviz.db` |
| Articles directory | `--articles-dir` | — | `~/workspace/finviz/articles/` |
| Crawl interval | `--sleep` | — | `300` (5 min) |
| Article expiry | `--expiry-days` | `FINVIZ_EXPIRY_DAYS` | `7` days |
| Timezone | — | `FINVIZ_TZ` or `TZ` | System default |
💬 Chat Commands (OpenClaw Agent)
When this skill is installed, the agent recognizes `/finviz` as a shortcut:
| Command | Action |
|---------|--------|
| `/finviz list` | Show tracked tickers |
| `/finviz add NVDA, TSLA` | Add tickers to track |
| `/finviz remove NVDA` | Remove a ticker |
| `/finviz stats` | Show article/ticker counts |
| `/finviz help` | Show available commands |
The agent runs these via the `finviz_query.py` CLI internally.
📱 PrivateApp Dashboard
A companion mobile dashboard is available in [PrivateApp](https://github.com/camopel/PrivateApp) — a personal PWA dashboard for your home server.
The **Finviz** app provides:
Install PrivateApp, and the Finviz dashboard is built-in — no extra setup needed.
Architecture
**Crawler daemon** (`finviz_crawler.py`):
**Query tool** (`finviz_query.py`):
Run as a service (optional)
systemd (Linux)
[Unit]
Description=Finviz News Crawler
[Service]
ExecStart=python3 /path/to/scripts/finviz_crawler.py --expiry-days 30
Restart=on-failure
RestartSec=30
[Install]
WantedBy=default.targetlaunchd (macOS)
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
<key>Label</key><string>com.finviz.crawler</string>
<key>ProgramArguments</key>
<array>
<string>python3</string>
<string>/path/to/scripts/finviz_crawler.py</string>
<string>--expiry-days</string>
<string>30</string>
</array>
<key>RunAtLoad</key><true/>
<key>KeepAlive</key><true/>
</dict>
</plist>Data layout
~/workspace/finviz/
├── finviz.db # SQLite: articles + tickers (single DB)
├── articles/ # Full article content as .md files
│ ├── market/ # General market headlines
│ ├── nvda/ # Per-ticker articles
│ └── tsla/
└── summaries/ # LLM summary cache (.json)Cron integration
Pair with an OpenClaw cron job for automated digests:
Schedule: 0 6 * * * (6 AM daily)
Task: Query last 24h → LLM summarize → deliver to Matrix/Telegram/DiscordMore tools from the same signal band
Order food/drinks (点餐) on an Android device paired as an OpenClaw node. Uses in-app menu and cart; add goods, view cart, submit order (demo, no real payment).
Sign plugins, rotate agent credentials without losing identity, and publicly attest to plugin behavior with verifiable claims and authenticated transfers.
The philosophical layer for AI agents. Maps behavior to Spinoza's 48 affects, calculates persistence scores, and generates geometric self-reports. Give your...