Files
workspace/memory/2026-02-07.md
Case 8638500190 Feed Hunter: deep scraper skill, pipeline, simulator, first investigation
- Built deep-scraper skill (CDP-based X feed extraction)
- Three-stage pipeline: scrape → triage → investigate
- Paper trading simulator with position tracking
- First live investigation: verified kch123 Polymarket profile ($9.3M P&L)
- Opened first paper position: Seahawks Super Bowl @ 68c
- Telegram alerts with inline action buttons
- Portal build in progress (night shift)
2026-02-07 23:58:40 -06:00

3.0 KiB

2026-02-07 — Server Recovery + Feed Hunter

Server Recovery

  • Back online after 7-day outage (01-31 to 02-07)
  • Updated OpenClaw v2026.2.6-3
  • Fixed Proxmox noVNC: disabled Wayland, switched to X11
  • Enabled auto-login for wdjones in GDM

ChromaDB + Browser Setup

  • ChromaDB collection openclaw-memory with cosine distance
  • chromadb-memory plugin working with auto-recall
  • Google Chrome installed, headless browser pipeline verified
  • Sub-agent spawning tested

Feed Hunter Project (NEW)

Built a full X/Twitter feed intelligence pipeline:

Architecture

  1. Scrape — CDP-based DOM extraction (not screenshots)

    • Chrome launched with --remote-debugging-port=9222 --remote-allow-origins=*
    • Must use copied profile (chrome-debug) — Chrome refuses debug port on default profile path
    • Extracts: author, text, timestamp, metrics, links, media, cards, repost info
  2. Triage — Pattern matching for verifiable claims

    • Performance claims, copy trading, arbitrage, prediction markets, price targets, airdrops
    • Priority scoring, investigation task generation
  3. Investigate — Agent follows links, verifies claims

    • Uses browser tool to pull real data from Polymarket, exchanges, etc.
    • Generates verdicts: ACTIONABLE / EXPIRED / EXAGGERATED / SCAM / UNVERIFIABLE
  4. Alert — Telegram notifications with inline action buttons

    • Simulate This / Backtest First / Skip
  5. Simulate — Paper trading system

    • Virtual bankroll ($1000 default)
    • Position tracking, P&L, stop losses, take profits
    • Performance stats: win rate, ROI, by-strategy breakdown

Files

  • skills/deep-scraper/ — scraping skill (SKILL.md + scripts)
  • projects/feed-hunter/ — project home
    • run-pipeline.sh — full pipeline orchestrator
    • simulator.py — paper trading CLI
    • investigate.py — investigation task generator
    • config.json — pipeline settings

Key Discovery

  • Chrome refuses --remote-debugging-port when --user-data-dir is the default path
  • Solution: copy profile to ~/.config/google-chrome-debug/ and launch from there
  • Need --remote-allow-origins=* for WebSocket CDP access
  • Python needs -u flag for unbuffered output in pipeline scripts

First Live Investigation

  • @linie_oo claimed @kch123 has ~$10M Polymarket profit
  • Verified on Polymarket: $9,371,829 all-time P&L
  • 1,862 predictions, $2.3M active positions
  • Sent investigation alert to D J with action buttons

D J's Vision

  • Scrape → investigate → verify → simulate → backtest → if viable, spawn working project
  • Everything paper-traded first to prove it works
  • Backtesting wherever historical data exists
  • Web portal to present reports and implementation details
  • D J headed to bed ~midnight, asked me to refine overnight + build portal

Night Shift Plan

  • Sub-agent building web portal at localhost:8888
  • Refine triage patterns
  • Add positions monitoring
  • Portal shows: dashboard, feed view, investigations, sim tracker, pipeline status