@@ -884,6 +884,7 @@ async def crawl(
884884 include_paths : Optional [list [str ]] = None ,
885885 exclude_paths : Optional [list [str ]] = None ,
886886 webhook_url : Optional [str ] = None ,
887+ wait_ms : Optional [int ] = None ,
887888 return_toon : bool = False ,
888889 ):
889890 """Send a crawl request with support for both AI extraction and
@@ -911,6 +912,7 @@ async def crawl(
911912 exclude_paths: List of path patterns to exclude (e.g., ['/admin/*', '/api/*'])
912913 Supports wildcards and takes precedence over include_paths
913914 webhook_url: URL to receive webhook notifications when the crawl completes
915+ wait_ms: Number of milliseconds to wait before scraping each page
914916 return_toon: If True, return response in TOON format (reduces token usage by 30-60%)
915917 """
916918 logger .info ("🔍 Starting crawl request" )
@@ -944,6 +946,8 @@ async def crawl(
944946 logger .debug (f"❌ Exclude paths: { exclude_paths } " )
945947 if webhook_url :
946948 logger .debug (f"🔔 Webhook URL: { webhook_url } " )
949+ if wait_ms is not None :
950+ logger .debug (f"⏱️ Wait ms: { wait_ms } " )
947951 if return_toon :
948952 logger .debug ("🎨 TOON format output enabled" )
949953
@@ -977,6 +981,8 @@ async def crawl(
977981 request_data ["exclude_paths" ] = exclude_paths
978982 if webhook_url is not None :
979983 request_data ["webhook_url" ] = webhook_url
984+ if wait_ms is not None :
985+ request_data ["wait_ms" ] = wait_ms
980986
981987 request = CrawlRequest (** request_data )
982988 logger .debug ("✅ Request validation passed" )
0 commit comments