The pipeline
Step 1: scrape a corpus
Use batch mode to fetch many URLs in one call:Step 2: chunk the markdown
Break each document into chunks small enough to fit in your embedding model’s context window. A reasonable default is ~500 tokens per chunk with 50-token overlap.tiktoken) instead of character-count estimates.
Step 3: embed and store
Step 4: retrieve at query time
Refreshing the index
Reader’s 24h cache means re-running your ingestion pipeline daily is cheap: anything that hasn’t changed returns from cache (0 credits). Only the genuinely new and updated pages cost credits. For sites that update frequently, run a daily crawl or re-scrape. For stable docs, once a week or on-demand is enough.Cost considerations
Per URL in the ingestion pipeline: 1 credit (standard mode) to 3 credits (stealth). A 10,000-URL corpus inauto typically costs 10,000–15,000 credits depending on escalation rate. See Cost estimation to pilot first.
Next
- Agent tool: Reader behind an LLM tool call
- Structured data: when you need JSON, not prose

