Reader is the web infrastructure platform for AI. Scrape, crawl, and automate the web from a single API. Get back clean markdown, HTML, and metadata, or spin up cloud browser sessions with full Playwright/Puppeteer control. One API key. Three primitives. You point it at a URL, we handle the rest:Documentation Index
Fetch the complete documentation index at: https://docs.reader.dev/llms.txt
Use this file to discover all available pages before exploring further.
- Rendering dynamic, JavaScript-heavy sites so you get the page the user sees
- Bypassing bot walls automatically. Pick
standard,stealth, or letautodecide - Extracting main content and stripping navigation, ads, and boilerplate
- Converting to clean markdown optimized for language models
- Automating with cloud browser sessions for login flows, multi-step navigation, and interactive automation
- Caching successful scrapes for 24 hours, so repeat requests are free
What you can build
RAG pipelines
Scrape docs, blog posts, and knowledge bases into your vector store.
AI agents
Give your agent the ability to read the web as a tool call.
Research tools
Batch scrape hundreds of URLs in parallel and stream results live.
Monitoring
Crawl competitor sites and extract structured data on a schedule.
Get started in 60 seconds
Sign up free
Create an account at app.reader.dev. No credit card required - you get 1,000 credits every month on the free tier.
Grab your API key
Open the API Keys page in the dashboard and click Create API Key. Copy it - you’ll only see it once.
Make your first request
Follow the Quickstart to scrape your first page in under a minute.
Open Dashboard →
Sign up free and create your API key in 30 seconds.
How Reader works
Three primitives, each with its own endpoint:| Endpoint | Primitive | What it does |
|---|---|---|
POST /v1/read with single url | Read | Scrape a page. Returns results synchronously |
POST /v1/read with urls array | Batch | Scrape many pages. Returns a jobId to poll or stream |
POST /v1/read with maxDepth / maxPages | Crawl | Discover links and scrape each page asynchronously |
POST /v1/sessions | Browse | Spin up a cloud browser. Connect Playwright/Puppeteer via CDP |
Why developers pick Reader
- Three primitives, one platform. Read, crawl, and browse through one API key, one credit pool. No separate billing tracks or hidden multipliers.
- Browser automation included. Full Playwright and Puppeteer compatibility via CDP. Connect your existing scripts to cloud browsers with anti-bot stealth active.
- Handles the modern web. Dynamic pages render correctly, bot walls get bypassed automatically in
automode, and you never touch the underlying infrastructure. - Built for LLMs. Output is clean markdown by default, with optional HTML and metadata. Main content extraction removes boilerplate so your tokens aren’t wasted.
- Free tier is actually free. 1,000 credits a month, no card required. Enough to build and ship a real prototype.
- Self-hostable. The underlying engine is open source and Apache 2.0 licensed. Host it yourself if you need to. Learn more →
Where to go next
Quickstart
Make your first scrape in 60 seconds.
The read primitive
Understand the mental model behind Reader’s single endpoint.
Guides
Practical recipes for RAG, agents, production patterns, and more.
API Reference
Full API spec for every endpoint.

