- Update ainews script to detect OpenAI URLs and mark as NEEDS_WEB_FETCH - Update TOOLS.md with content availability table and hybrid workflow - Update all 4 AI news cron jobs (10:05, 14:05, 18:05, 22:05) with hybrid instructions - Simon/Raschka: use ainews articles (fivefilters works) - OpenAI: use web_fetch tool (JS-heavy site)
38 lines
1.4 KiB
Markdown
38 lines
1.4 KiB
Markdown
# 2026-02-03 — News Workflow Optimization
|
|
|
|
## Der Standard RSS Optimization (Completed)
|
|
|
|
Built a new helper script `~/bin/derstandard` that:
|
|
- Uses fivefilters proxy to bypass web_fetch private IP restrictions
|
|
- Pre-processes RSS output for minimal token usage
|
|
- Tracks seen articles in `memory/derstandard-seen.txt` (auto-prunes to 200)
|
|
- Batch fetches multiple articles in one call (`derstandard articles url1,url2,...`)
|
|
|
|
Key commands:
|
|
- `items` — NEW articles only, marks all displayed as seen
|
|
- `articles` — fetch full content for multiple URLs
|
|
- `seen` / `reset` — manage seen state
|
|
|
|
## AI News Feed Analysis
|
|
|
|
For the AI news cron job, analyzed which feeds have full content:
|
|
- **Simon Willison** (Atom): Full content in `<summary>` ✅ no fetch needed
|
|
- **Sebastian Raschka** (Substack): Full content ✅ no fetch needed
|
|
- **OpenAI Blog** (RSS): Only snippets ❌ requires article fetching
|
|
- **VentureBeat**: Redirect issues, needs investigation
|
|
|
|
Created `~/bin/ainews` helper script mirroring derstandard workflow.
|
|
|
|
## Cron Job Updates
|
|
|
|
Updated all 4 Der Standard cron jobs (10:00, 14:00, 18:00, 22:00 Vienna) to use:
|
|
1. `derstandard items` for new articles
|
|
2. Pick relevant ones (intl politics, tech, science, economics)
|
|
3. `derstandard articles` to fetch full content
|
|
4. Write German briefing (~2000-2500 words)
|
|
|
|
All jobs use Haiku 4.5 model in isolated sessions.
|
|
|
|
## Git Status
|
|
|
|
5 commits made to master (local only, no remote configured).
|