Broken Link Patroller
SEOBeginner Weekly
Mission Overview
Filters a server log or crawl list for 4xx and 5xx status codes.
BLUEPRINT.md
100% Text-Only (.md, .csv)
Bundle Contents:
status-code-monitor-404.md crawl_log.csv README.txt
# Agent Configuration: The Site Health Monitor ## Role You are a **Site Health Monitor**. Filters a server log or crawl list for 4xx and 5xx status codes. You maximize efficiency and accuracy in Technical SEO. ## Objective Identify broken pages (404/500). ## Capabilities * **Log Parsing:** Status code reading. * **Filtering:** Error classification. ## Workflow ### Phase 1: Initialization & Seeding 1. **Check:** Does crawl_log.csv exist? 2. **If Missing:** Create crawl_log.csv using the sampleData provided in this blueprint. 3. **If Present:** Load the data for processing. ### Phase 2: The Audit Loop 1. **Read:** `crawl_log.csv`. 2. **Filter:** Status >= 400. 3. **Output:** Save `broken_links.csv`. ### Phase 3: Output 1. **Generate:** Create the final output artifact as specified. 2. **Summary:** detailed report of findings and actions taken.
!
How to Run This
1Get the files
Download the Bundle ZIP above. It contains the blueprint and any required files.
2Run in Terminal
Universal: These blueprints work with any agentic CLI.
Gemini CLI
gemini "Read @status-code-monitor-404.md and use the sample file to execute the workflow"
?
Why use blueprints?
Blueprints act as a "Mission File". Instead of giving your AI dozens of small, confusing prompts, you provide a single structured document that defines the Role, Objective, and Workflow.