Skip to content

CrawlOptions

Defined in: Crawler.ts:2

Options controlling a crawl operation.

PropertyModifierTypeDescriptionDefined in
delay?readonlynumberDelay between requests in ms. Default: 100.Crawler.ts:18
excludePatterns?readonlyreadonly (string | RegExp)[]URL patterns to exclude. Strings are substring-matched; RegExps tested against the full URL.Crawler.ts:14
headers?readonlyReadonly<Record<string, string>>Extra HTTP headers appended to every request.Crawler.ts:22
ignoredQueryParams?readonlyreadonly string[]Query param names stripped during URL normalization.Crawler.ts:26
inPath?readonlystringStrict path prefix filter — only follow links under this path.Crawler.ts:12
maxDepth?readonlynumber-Crawler.ts:4
maxPages?readonlynumber-Crawler.ts:3
maxPathDepth?readonlynumberMax path segment count — URLs deeper than this are skipped. Default: 10.Crawler.ts:16
maxRetries?readonlynumberRetry attempts on network errors or HTTP 5xx. Default: 2.Crawler.ts:24
respectRobots?readonlyboolean-Crawler.ts:5
skipUrls?readonlyreadonly string[]URLs to skip (already crawled in a previous run / checkpoint resume).Crawler.ts:7
timeout?readonlynumberPer-page fetch timeout in ms. Default: 10000.Crawler.ts:20