HomeBrowseUpload
← Back to registry
// Skill profile

Social Scraper

version: "2.0.0"

by ckchzh · published 2026-03-22

社交媒体数据处理
Total installs
0
Stars
★ 0
Last updated
2026-03
// Install command
$ claw add gh:ckchzh/ckchzh-social-scraper
View on GitHub
// Full documentation

---

version: "2.0.0"

name: Twint

description: "An advanced Twitter scraping & OSINT tool written in Python that doesn't use Twitter's API, allowing twitter-scraper, python, elasticsearch, kibana, osint."

---

# Social Scraper

Social Scraper v2.0.0 — a general-purpose utility toolkit for logging, tracking, and managing data entries from the command line. Each command records timestamped entries into its own log file and supports review of recent history.

Commands

The script (`scripts/script.sh`) exposes the following commands via a `case` dispatcher:

| Command | Description |

|---------|-------------|

| `run <input>` | Record a "run" entry. Without args, shows the 20 most recent run entries. |

| `check <input>` | Record a "check" entry. Without args, lists recent check entries. |

| `convert <input>` | Record a "convert" entry. Without args, lists recent convert entries. |

| `analyze <input>` | Record an "analyze" entry. Without args, lists recent analyze entries. |

| `generate <input>` | Record a "generate" entry. Without args, lists recent generate entries. |

| `preview <input>` | Record a "preview" entry. Without args, lists recent preview entries. |

| `batch <input>` | Record a "batch" entry. Without args, lists recent batch entries. |

| `compare <input>` | Record a "compare" entry. Without args, lists recent compare entries. |

| `export <input>` | Record an "export" entry. Without args, lists recent export entries. |

| `config <input>` | Record a "config" entry. Without args, lists recent config entries. |

| `status <input>` | Record a "status" entry. Without args, lists recent status entries. |

| `report <input>` | Record a "report" entry. Without args, lists recent report entries. |

| `stats` | Show summary statistics across all log files (entry counts per type, total, disk usage). |

| `export <fmt>` | Export all data in `json`, `csv`, or `txt` format to `$DATA_DIR/export.<fmt>`. |

| `search <term>` | Search all log files for a term (case-insensitive grep). |

| `recent` | Show the 20 most recent lines from `history.log`. |

| `status` | Health check — shows version, data directory, total entries, disk usage, last activity. |

| `help` | Display the full help/usage message. |

| `version` | Print `social-scraper v2.0.0`. |

> **Note:** The `export` and `status` commands appear twice in the case statement. The first match (entry-logging form) takes precedence. The standalone `_export` and `_status` helper functions are reachable only if the entry-logging branches are bypassed.

How Each Entry Command Works

1. If called **without arguments**, it tails the last 20 lines of `<command>.log`.

2. If called **with arguments**, it:

- Timestamps the input (`YYYY-MM-DD HH:MM|<input>`)

- Appends it to `$DATA_DIR/<command>.log`

- Prints confirmation with the current total count

- Logs the action to `history.log`

Data Storage

All data is stored as plain-text log files under:

~/.local/share/social-scraper/
├── run.log
├── check.log
├── convert.log
├── analyze.log
├── generate.log
├── preview.log
├── batch.log
├── compare.log
├── export.log
├── config.log
├── status.log
├── report.log
└── history.log          # unified activity log

Each log line uses pipe-delimited format: `YYYY-MM-DD HH:MM|<value>`

The `history.log` uses: `MM-DD HH:MM <command>: <value>`

Requirements

  • **Bash** 4.0+ (uses `local` variables, `set -euo pipefail`)
  • **coreutils**: `date`, `wc`, `du`, `tail`, `cat`, `basename`, `grep`, `sed`
  • No external dependencies, API keys, or network access required
  • Works on Linux and macOS
  • When to Use

    1. **Quick data logging** — when you need a lightweight CLI to record timestamped scraping results without setting up a database

    2. **Scrape tracking** — log each scraping run with parameters and review them later with `recent` or `search`

    3. **Batch scrape records** — track batch scraping jobs for auditing and reproducibility

    4. **Data export** — pull all logged entries into JSON, CSV, or TXT for reporting or integration with analytics pipelines

    5. **Health monitoring** — use `stats` and `status` to get a quick overview of scraping activity volume and disk usage

    Examples

    Log a scraping run and review history

    # Record a scraping session
    bash scripts/script.sh run "scraped @user timeline 500 tweets"
    
    # Check recent runs
    bash scripts/script.sh run

    Analyze and report on collected data

    # Log an analysis
    bash scripts/script.sh analyze "sentiment breakdown for #topic"
    
    # Generate a report entry
    bash scripts/script.sh report "weekly scrape summary: 3200 entries"

    Search across all logs

    bash scripts/script.sh search "timeline"

    Export everything as CSV

    bash scripts/script.sh export csv
    # Output: ~/.local/share/social-scraper/export.csv

    View summary statistics

    bash scripts/script.sh stats
    # Shows per-type counts, totals, and disk usage

    Configuration

    Set the `DATA_DIR` variable (or modify it in the script) to change the storage directory. Default: `~/.local/share/social-scraper/`

    Output

    All commands print to stdout. Redirect to a file as needed:

    bash scripts/script.sh report > weekly-report.txt

    ---

    Powered by BytesAgain | bytesagain.com | hello@bytesagain.com

    // Comments
    Sign in with GitHub to leave a comment.
    // Related skills

    More tools from the same signal band