🔍 PDF Accessibility Crawler

Submit a website URL to discover all PDF files it contains and assess their accessibility using the checks from simplA11yPDFCrawler.

📊 View Latest Scan Results 📋 View All Reports

Quick Start

Enter a site URL below and click Submit Crawl Request. You will be taken to GitHub with a pre-filled issue titled SCAN: https://… — just click Submit new issue to start the crawl.

Enter the full URL including https://.

Maximum number of pages the spider will visit. Leave blank to use the default (2500). Use a lower value (e.g. 200) for large or slow sites to avoid job timeouts.

Prefer to create the issue yourself? Open a new GitHub issue and set the title to SCAN: https://example.com. Add a Number: 200 line to the issue body to cap the crawl at 200 pages.

How it works

  1. Fill in the form above and click Submit Crawl Request.
  2. You are redirected to GitHub to create a SCAN: issue, which triggers the Crawl Site for PDFs workflow (runs up to 1 hour).
  3. Once the crawl is complete, the Analyse PDFs for Accessibility workflow starts automatically and posts results back to your issue.
  4. PDFs are deleted after analysis; only the YAML manifest is kept (with MD5 hashes, so unchanged files are not rescanned).