Skip to main content

ScrapingAnt

概述

ScrapingAnt 是一个具有无头浏览器功能、代理和反机器人绕过能力的网页抓取 API。它允许将网页数据提取为可访问的 LLM markdown。

此特定集成仅使用 Markdown 提取功能,但如果您需要 ScrapingAnt 提供的更多功能,而该功能尚未在此集成中实现,请随时 与我们联系

集成细节

本地可序列化JS 支持
ScrapingAntLoaderlangchain_community

加载器特性

来源文档懒加载异步支持
ScrapingAntLoader

设置

使用 pip 安装 ScrapingAnt Python SDK 和所需的 Langchain 包:

pip install scrapingant-client langchain langchain-community

实例化

from langchain_community.document_loaders import ScrapingAntLoader

scrapingant_loader = ScrapingAntLoader(
["https://scrapingant.com/", "https://example.com/"], # 要抓取的 URL 列表
api_key="<YOUR_SCRAPINGANT_TOKEN>", # 从 https://scrapingant.com/ 获取您的 API 密钥
continue_on_failure=True, # 忽略无法处理的网页并记录其异常
)
[Document(metadata={'url': 'https://scrapingant.com/'}, page_content="![](images/loader.svg)\n\n[![](images/ScrapingAnt-1.svg)](/) Features Pricing\n\nServices\n\n[Web Scraping API](/) [LLM-ready data extraction](/llm-ready-data-extraction)\n[AI data scraping](/ai-data-scraper) [Residential Proxy](/residential-proxies)\n\n[Blog](https://scrapingant.com/blog/)\n\nDocumentatation\n\n[Web Scraping API](https://docs.scrapingant.com) [Residential\nProxies](https://proxydocs.scrapingant.com)\n\nContact Us\n\n[Sign In](https://app.scrapingant.com/login)\n\n![](images/icon-menu.svg)\n\n![](images/Capterra-Rating.png)\n\n# Enterprise-Grade Scraping API.  \nAnt Sized Pricing.\n\n## Get the mission-critical speed, reliability, and features you need at a\nfraction of the cost!  \n\nGot Questions?  \n(get expert advice)\n\n[ Try Our Free Plan (10,000 API Credits) ](https://app.scrapingant.com/signup)\n\n![](images/lines-10-white.svg)![](images/lines-12-white.svg)\n\n### Proudly scaling with us\n\n![](images/_2cd6c6d09d261d19_281d72aa098ecca8.png)![](images/_bb8ca9c8d001abd4_dc29a36ce27bdee8_1_bb8ca9c8d001abd4_dc29a36ce27bdee8.png)![](images/_d84700234b61df23_9abf58d176a2d7fc.png)![](images/_ca6d37170ae5cd25_fca779750afd17ef.png)![](images/Screenshot-2024-05-22-at-23.28.16.png)\n\n### Industry Leading Pricing\n\nFrom our generous 10,000 API credit free plan to our industry leading paid\nplans, we strive to provide unbeatable bang for your buck. That's just what\nants do!  \n\u200d\n\n![](images/industry-leading-prcing--compressed.webp)\n\nCost per 1,000 API Credits - Level 1 Plan\n\n### Unparalleled Value\n\nLow cost per API credit is great, but what’s even more important is how much\ndata you can actually collect for each credit spent. Like any good Ant we\nnever waste a crumb!\n\n![](images/unparalleled-value-compressed.webp)\n\nGoogle SERP API - Cost per 1,000 Requests – Level 1 Plan\n\n![](images/Doodle-4-White.svg)![](images/Doodle-Left-1-White.svg)\n\n## Ultimate Black Box Scraping Solution\n\n### Unlimited Concurrency  \n\u200d\n\nWith unlimited parallel requests easily gather LARGE volumes of data from\nmultiple locations in record time. Available on ALL plan levels.  \n\u200d\n\n### Lightning Fast Scraping WITHOUT Getting Blocked\n\nOur proprietary algo seamlessly switches to the exact right proxy for almost\nany situation, saving you and your dev team countless hours of frustration.  \n\u200d\n\n#### What's inside?\n\n  * Chrome Page Rendering\n\n  * Low Latency Rotating Proxies  \n\n  * Javascript Execution\n\n  * Custom Cookies\n\n  * Fastest AWS & Hetzner Servers\n\n  * Unlimited Parallel Requests\n\n  * Headless Browsers  \n\n  * Residential Proxies\n\n  * Supports All Programming Languages & Proxy\n\n  * CAPTCHA Avoidance\n\n[ Try Our Free Plan (10,000 API Credits) ](https://app.scrapingant.com/signup)\n\n![](images/Doodle-3-White.svg)\n\n###### Metrics\n\n## The most reliable web scraping API\n\nOur clients have saved up to 40% of data collection budgets by integrating\nScrapingAnt API instead of self-made solutions development.\n\n99.99%\n\nUptime over the last year.\n\n85.5%\n\nAnti-scraping avoidance rate with our custom cloud browser solution\n\n![](images/icon-gallery-dark.svg)\n\n### Unlimited parallel requests\n\n![](images/icon-id-dark.svg)\n\n### 3+ million proxy servers across the world\n\n![](images/icon-switcher-white.svg)\n\n### Open your web page as in a real browser\n\n![](images/Doodle-9-Dark.svg)\n\nSimple API integration\n\n1\n\n### Choose your plan\n\nWe offer subscription plans, or you can always request custom pricing.  \n **Free for personal use!**\n\n2\n\n### Test the API\n\nScrape your target website with our **UI request executor** or generate\nscraping code for your preferred language.\n\n3\n\n### Scrape the Web\n\nBuild your data extraction pipeline using our **API** and forget about **rate\nlimits** and **blocks**.\n\n![](images/Doodle-4-White.svg)![](images/Doodle-Left-1-White.svg)\n\n###### Pricing\n\n## Industry leading pricing that scales with your business.\n\n### Enthusiast\n\n#### 100.000 API credits\n\n$19\n\n/mo\n\nIdeal for freelancers or students.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nEmail support\n\n![](images/check-small.svg)\n\nDocumentation-only integration\n\n### Startup\n\n#### 500.000 API credits\n\n$49\n\n/mo\n\nFor small to medium sized teams looking to grow.  \n  \nPopular choice!\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nExpert assistance\n\n![](images/check-small.svg)\n\nIntegration with custom code snippets\n\n### Business\n\n#### 3.000.000 API credits\n\n$249\n\n/mo\n\nFor larger teams and companies.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nLive integration calls\n\n![](images/check-small.svg)\n\nExpert guidance and integration planning\n\n![](images/check-small.svg)\n\nCustom proxy pools\n\n![](images/check-small.svg)\n\nCustom avoidances\n\n![](images/check-small.svg)\n\nDedicated manager\n\n### Business Pro\n\n#### 8.000.000 API credits\n\n$599\n\n/mo\n\nExtended volume Business plan.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nLive integration calls\n\n![](images/check-small.svg)\n\nExpert guidance and integration planning\n\n![](images/check-small.svg)\n\nCustom proxy pools\n\n![](images/check-small.svg)\n\nCustom avoidances\n\n![](images/check-small.svg)\n\nDedicated manager\n\n### Custom Plan\n\n#### 10M+ API credits\n\n$699+\n\n/mo\n\nExplore custom deals and services we could provide for Enterprise level\ncustomers.\n\n[ Contact us ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nFully customisable solution\n\n![](images/check-small.svg)\n\nResidential Proxy special prices\n\n![](images/check-small.svg)\n\nSLA\n\n[![](images/Capterra-\nRating.png)](https://www.capterra.com/p/214735/ScrapingAnt/reviews/)\n\n★ ★ ★ ★ ★\n\n![](images/5521ce5758e089d7d7f5d226a2e995c3.jpg)\n\n#### “Onboarding and API integration was smooth and clear. Everything works\ngreat. The support was excellent. **Overall a great scraper**.”\n\nIllia K., Android Software Developer\n\n★ ★ ★ ★ ★\n\n![](images/e57164aafb18d9a888776c96cf159368.jpg)\n\n#### “Great communication with co-founders helped me to get the job done.\nGreat proxy diversity and good price.”\n\nAndrii M., Senior Software Engineer\n\n★ ★ ★ ★ ★\n\n![](images/Dmytro-T..jpg)\n\n#### “This product helps me to scale and extend my business. The API is easy\nto integrate and support is really good.”\n\nDmytro T., Senior Software Engineer\n\n![](images/Doodle-7-Dark.svg)![](images/Doodle-8-Dark.svg)\n\n#### 常见问题解答。\n\n如果您有任何进一步的问题,请与我们的友好团队[联系](https://scrapingant.com/#contact)\n\n##### 什么是 ScrapingAnt?\n\n![](images/icon-arrow-right.svg)\n\nScrapingAnt 是一个帮助您解决任何复杂性抓取任务的服务。通过使用全球数百万个代理和整个无头浏览器集群,我们可以为您提供最佳的网络采集和抓取体验。  \n  \nScrapingAnt 还提供定制软件开发服务。数据采集、数据存储或数据查询 - 我们可以为您提供最优质和负担得起的定制解决方案,以满足您的所有需求。\n\n##### **什么是 API 信用?**\n\n![](images/icon-arrow-right.svg)\n\n每个订阅计划包含每月特定数量的 API 信用。根据您配置的 API 调用参数,费用从一个到几个信用不等。默认情况下,每个请求的费用为 10 个 API 信用,因为启用了 JavaScript 渲染和标准代理。[了解更多关于请求费用的信息](https://docs.scrapingant.com/api-credits-usage)。\n\n##### 我不是开发人员,您能为我创建定制的抓取解决方案吗?\n\n![](images/icon-arrow-right.svg)\n\n当然可以!我们定期为客户创建定制的抓取脚本和项目。我们还与几家定制软件开发公司合作,因此我们永远不会缺乏资源来帮助您处理任何规模的抓取项目。只需[联系我们](https://scrapingant.com/#contact)并描述您的需求。\n\n##### 我需要信用卡才能开始免费试用吗?\n\n![](images/icon-arrow-right.svg)\n\nScrapingAnt 提供完全免费的订阅计划,其中包含 10,000 个 API 信用,可以在一个月内使用。直到您需要更多 - 它完全免费且不需要信用卡。\n\n### “我们的客户对我们团队的响应速度感到惊讶。”\n\n![](images/oleg-cartoon-image.jpg)\n\nOleg Kulyk,  \nScrapingAnt 创始人\n\n* 我们的团队会尽快与您联系。\n\n谢谢!您的提交已被接收!\n\n哎呀!提交表单时出现问题。\n\n![](images/illustration-speed-lines-white.svg)\n\n## 与我们一起发展您的业务\n\n[ 尝试我们的免费计划! ](https://app.scrapingant.com/signup)\n\n[\n\n## 特性\n\n](https://scrapingant.com/#features) [\n\n## 定价\n\n](https://scrapingant.com/#pricing) [\n\n## 博客\n\n](https://scrapingant.com/blog/) [\n\n## 文档\n\n](https://docs.scrapingant.com/) [\n\n## 网络抓取 API\n\n](https://scrapingant.com) [\n\n## LLM-ready 网络数据\n\n](llm-ready-data-extraction.html) [\n\n## 住宅代理\n\n](residential-proxies.html) [\n\n## 定制抓取器开发\n\n](https://scrapingant.com/custom-scraping-solution) [\n\n## 会员计划\n\n](https://scrapingant.com/legal/affiliate/) [\n\n## 免费代理\n\n](https://scrapingant.com/free-proxies/)\n\n###### 网络抓取 101  \n\n[什么是网络抓取?](https://docs.scrapingant.com/web-scraping-101/what-is-\nweb-scraping) [**网络抓取是否合法?**](https://scrapingant.com/blog/is-\nweb-scraping-legal) [**10 种主要代理\n类型**](https://scrapingant.com/blog/main-proxy-types) [数据中心与\n住宅代理](https://scrapingant.com/blog/residential-vs-datacenter-\nproxy-webscraping) [最佳代理抓取\n工具](https://scrapingant.com/blog/top-open-source-proxy-scrapers)\n[**通过网络抓取 API 克服抓取挑战**](https://scrapingant.com/blog/data-scraping-challenges) [避免 IP 速率限制](https://scrapingant.com/blog/avoid-ip-rate-limiting)\n[使用 Puppeteer 轮换代理](https://scrapingant.com/blog/how-to-use-\nrotating-proxies-with-puppeteer) [使用 Python 抓取动态网站](https://scrapingant.com/blog/scrape-dynamic-website-with-python) [使用 Python 进行网络抓取](https://scrapingant.com/blog/top-5-popular-python-\nlibraries-for-web-scraping-in-2020) [使用 Java 进行网络抓取](https://scrapingant.com/blog/web-scraping-java) [使用 NodeJS 进行网络抓取](https://scrapingant.com/blog/web-scraping-javascript) [使用 Deno 进行网络抓取](https://scrapingant.com/blog/deno-web-scraping) [**使用 R 进行网络抓取**](https://scrapingant.com/blog/r-web-scraping) [**使用 PHP 进行网络抓取**](https://scrapingant.com/blog/web-scraping-php) [**使用 Go 进行网络抓取**](https://scrapingant.com/blog/web-scraping-go)\n\n###### 用例  \n\n[**使用 Booking.com 抓取进行房地产决策**](https://scrapingant.com/blog/booking-data-scraping) [**使用网络抓取 API 收集运动鞋价格数据**](https://scrapingant.com/blog/sneakers-scraping-api) [**自由职业者的最佳网络抓取 API**](https://scrapingant.com/blog/best-web-\nscraping-api-freelance) [**通过数据收集做出智能 NFT 决策**](https://scrapingant.com/blog/nft-data-collection) [**数据收集如何改善人力资源流程**](https://scrapingant.com/blog/data-\ncollection-for-hr-processes) [**通过数据收集统治电子商务**](https://scrapingant.com/blog/data-collection-for-ecommerce)\n[**公司如何利用网络抓取获得竞争优势**](https://scrapingant.com/blog/how-companies-use-web-scraping)\n[**网络抓取对酒店业的好处**](https://scrapingant.com/blog/web-scraping-for-hospitality)\n[**网络抓取在价格监控中的应用**](https://scrapingant.com/blog/web-scraping-for-price-monitoring)\n[**网络抓取对房地产的好处**](https://scrapingant.com/blog/web-scraping-for-real-estate) [**数据科学家的网络抓取**](https://scrapingant.com/blog/web-scraping-for-\ndata-scientists) [**如何从 TikTok 收集数据**](https://scrapingant.com/blog/web-scraping-for-price-monitoring)\n\n###### 法律  \n\n[使用条款](https://scrapingant.com/legal/terms-of-use) [隐私政策](https://scrapingant.com/legal/privacy-policy) [Cookies\n政策](https://scrapingant.com/legal/cookies-policy)\n\n###### 外部链接  \n\n[Github](https://github.com/ScrapingAnt)\n[Linkedin](https://linkedin.com/company/scrapingant)\n[Facebook](https://www.facebook.com/scrapingant)\n[Twitter](https://twitter.com/ScrapingAnt)\n\n[![](images/ScrapingAnt-2.svg)](https://scrapingant.com)\n\n© Copyright ScrapingAnt  \nPowered by [DATAANT](https://scrapingant.com)\n\n![](images/lines-13-white.svg)\n\nBy browsing this site, you agree to our [Cookies\nPolicy](https://scrapingant.com/legal/cookies-policy)\n\n![](images/icon-x_1.svg)\n\n"), Document(metadata={'url': 'https://example.com/'}, page_content='# Example Domain\n\nThis domain is for use in illustrative examples in documents. You may use this\ndomain in literature without prior coordination or asking for permission.\n\n[More information...](https://www.iana.org/domains/example)\n\n')]

ScrapingAntLoader 还允许提供字典 - 抓取配置以自定义抓取请求。由于它基于 ScrapingAnt Python SDK,您可以将任何 常见参数 传递给 scrape_config 参数。

from langchain_community.document_loaders import ScrapingAntLoader

scrapingant_config = {
"browser": True, # 启用云浏览器的浏览器渲染
"proxy_type": "datacenter", # 选择代理类型(数据中心或住宅)
"proxy_country": "us", # 选择代理位置
}

scrapingant_additional_config_loader = ScrapingAntLoader(
["https://scrapingant.com/"],
api_key="<YOUR_SCRAPINGANT_TOKEN>", # 从 https://scrapingant.com/ 获取您的 API 密钥
continue_on_failure=True, # 忽略无法处理的网页并记录其异常
scrape_config=scrapingant_config, # 传递 scrape_config 对象
)
[Document(metadata={'url': 'https://scrapingant.com/'}, page_content="![](images/loader.svg)\n\n[![](images/ScrapingAnt-1.svg)](/) Features Pricing\n\nServices\n\n[Web Scraping API](/) [LLM-ready data extraction](/llm-ready-data-extraction)\n[AI data scraping](/ai-data-scraper) [Residential Proxy](/residential-proxies)\n\n[Blog](https://scrapingant.com/blog/)\n\nDocumentatation\n\n[Web Scraping API](https://docs.scrapingant.com) [Residential\nProxies](https://proxydocs.scrapingant.com)\n\nContact Us\n\n[Sign In](https://app.scrapingant.com/login)\n\n![](images/icon-menu.svg)\n\n![](images/Capterra-Rating.png)\n\n# Enterprise-Grade Scraping API.  \nAnt Sized Pricing.\n\n## Get the mission-critical speed, reliability, and features you need at a\nfraction of the cost!  \n\nGot Questions?  \n(get expert advice)\n\n[ Try Our Free Plan (10,000 API Credits) ](https://app.scrapingant.com/signup)\n\n![](images/lines-10-white.svg)![](images/lines-12-white.svg)\n\n### Proudly scaling with us\n\n![](images/_2cd6c6d09d261d19_281d72aa098ecca8.png)![](images/_bb8ca9c8d001abd4_dc29a36ce27bdee8_1_bb8ca9c8d001abd4_dc29a36ce27bdee8.png)![](images/_d84700234b61df23_9abf58d176a2d7fc.png)![](images/_ca6d37170ae5cd25_fca779750afd17ef.png)![](images/Screenshot-2024-05-22-at-23.28.16.png)\n\n### Industry Leading Pricing\n\nFrom our generous 10,000 API credit free plan to our industry leading paid\nplans, we strive to provide unbeatable bang for your buck. That's just what\nants do!  \n\u200d\n\n![](images/industry-leading-prcing--compressed.webp)\n\nCost per 1,000 API Credits - Level 1 Plan\n\n### Unparalleled Value\n\nLow cost per API credit is great, but what’s even more important is how much\ndata you can actually collect for each credit spent. Like any good Ant we\nnever waste a crumb!\n\n![](images/unparalleled-value-compressed.webp)\n\nGoogle SERP API - Cost per 1,000 Requests – Level 1 Plan\n\n![](images/Doodle-4-White.svg)![](images/Doodle-Left-1-White.svg)\n\n## Ultimate Black Box Scraping Solution\n\n### Unlimited Concurrency  \n\u200d\n\nWith unlimited parallel requests easily gather LARGE volumes of data from\nmultiple locations in record time. Available on ALL plan levels.  \n\u200d\n\n### Lightning Fast Scraping WITHOUT Getting Blocked\n\nOur proprietary algo seamlessly switches to the exact right proxy for almost\nany situation, saving you and your dev team countless hours of frustration.  \n\u200d\n\n#### What's inside?\n\n  * Chrome Page Rendering\n\n  * Low Latency Rotating Proxies  \n\n  * Javascript Execution\n\n  * Custom Cookies\n\n  * Fastest AWS & Hetzner Servers\n\n  * Unlimited Parallel Requests\n\n  * Headless Browsers  \n\n  * Residential Proxies\n\n  * Supports All Programming Languages & Proxy\n\n  * CAPTCHA Avoidance\n\n[ Try Our Free Plan (10,000 API Credits) ](https://app.scrapingant.com/signup)\n\n![](images/Doodle-3-White.svg)\n\n###### Metrics\n\n## The most reliable web scraping API\n\nOur clients have saved up to 40% of data collection budgets by integrating\nScrapingAnt API instead of self-made solutions development.\n\n99.99%\n\nUptime over the last year.\n\n85.5%\n\nAnti-scraping avoidance rate with our custom cloud browser solution\n\n![](images/icon-gallery-dark.svg)\n\n### Unlimited parallel requests\n\n![](images/icon-id-dark.svg)\n\n### 3+ million proxy servers across the world\n\n![](images/icon-switcher-white.svg)\n\n### Open your web page as in a real browser\n\n![](images/Doodle-9-Dark.svg)\n\nSimple API integration\n\n1\n\n### Choose your plan\n\nWe offer subscription plans, or you can always request custom pricing.  \n **Free for personal use!**\n\n2\n\n### Test the API\n\nScrape your target website with our **UI request executor** or generate\nscraping code for your preferred language.\n\n3\n\n### Scrape the Web\n\nBuild your data extraction pipeline using our **API** and forget about **rate\nlimits** and **blocks**.\n\n![](images/Doodle-4-White.svg)![](images/Doodle-Left-1-White.svg)\n\n###### Pricing\n\n## Industry leading pricing that scales with your business.\n\n### Enthusiast\n\n#### 100.000 API credits\n\n$19\n\n/mo\n\nIdeal for freelancers or students.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nEmail support\n\n![](images/check-small.svg)\n\nDocumentation-only integration\n\n### Startup\n\n#### 500.000 API credits\n\n$49\n\n/mo\n\nFor small to medium sized teams looking to grow.  \n  \nPopular choice!\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nExpert assistance\n\n![](images/check-small.svg)\n\nIntegration with custom code snippets\n\n### Business\n\n#### 3.000.000 API credits\n\n$249\n\n/mo\n\nFor larger teams and companies.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nLive integration calls\n\n![](images/check-small.svg)\n\nExpert guidance and integration planning\n\n![](images/check-small.svg)\n\nCustom proxy pools\n\n![](images/check-small.svg)\n\nCustom avoidances\n\n![](images/check-small.svg)\n\nDedicated manager\n\n### Business Pro\n\n#### 8.000.000 API credits\n\n$599\n\n/mo\n\nExtended volume Business plan.\n\n[ Get Started ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nPriority email support\n\n![](images/check-small.svg)\n\nLive integration calls\n\n![](images/check-small.svg)\n\nExpert guidance and integration planning\n\n![](images/check-small.svg)\n\nCustom proxy pools\n\n![](images/check-small.svg)\n\nCustom avoidances\n\n![](images/check-small.svg)\n\nDedicated manager\n\n### Custom Plan\n\n#### 10M+ API credits\n\n$699+\n\n/mo\n\nExplore custom deals and services we could provide for Enterprise level\ncustomers.\n\n[ Contact us ](https://app.scrapingant.com/signup)\n\n![](images/check-small.svg)\n\nFully customisable solution\n\n![](images/check-small.svg)\n\nResidential Proxy special prices\n\n![](images/check-small.svg)\n\nSLA\n\n[![](images/Capterra-\nRating.png)](https://www.capterra.com/p/214735/ScrapingAnt/reviews/)\n\n★ ★ ★ ★ ★\n\n![](images/5521ce5758e089d7d7f5d226a2e995c3.jpg)\n\n#### “Onboarding and API integration was smooth and clear. Everything works\ngreat. The support was excellent. **Overall a great scraper**.”\n\nIllia K., Android Software Developer\n\n★ ★ ★ ★ ★\n\n![](images/e57164aafb18d9a888776c96cf159368.jpg)\n\n#### “Great communication with co-founders helped me to get the job done.\nGreat proxy diversity and good price.”\n\nAndrii M., Senior Software Engineer\n\n★ ★ ★ ★ ★\n\n![](images/Dmytro-T..jpg)\n\n#### “This product helps me to scale and extend my business. The API is easy\nto integrate and support is really good.”\n\nDmytro T., Senior Software Engineer\n\n![](images/Doodle-7-Dark.svg)![](images/Doodle-8-Dark.svg)\n\n#### Frequently asked questions.\n\nIf you have any further questions, [Get in\ntouch](https://scrapingant.com/#contact) with our friendly team\n\n##### What is ScrapingAnt?\n\n![](images/icon-arrow-right.svg)\n\nScrapingAnt is a service that helps you to solve scraping tasks of any\ncomplexity. With using of millions proxies around the World and a whole\nheadless browser cluster we can provide you the best web harvesting and\nscraping experience.  \n  \nScrapingAnt also provides a custom software development service. Data\nharvesting, data storage or data querying - we can provide you the best and\naffordable custom solution that fits all your needs.\n\n##### **What is an API Credit?**\n\n![](images/icon-arrow-right.svg)\n\nEach subscription plan contains a particular amount of API credits per month.\nDepending on the parameters you configures your API calls it will cost you\nfrom one to several credits. By default, each request costs 10 API credits\nbecause JavaScript rendering and Standard proxies are enabled. [Learn more\nabout requests costs](https://docs.scrapingant.com/api-credits-usage).\n\n##### I'm not a developer, can you create custom scraping solutions for me?\n\n![](images/icon-arrow-right.svg)\n\nYes of course! We regularly create custom scraping scripts and projects for\nour clients. We are also partnering with several custom software development\ncompanies, so we won't never be out of resources to help with a scraping\nproject of any size. Just [Contact Us](https://scrapingant.com/#contact) and\ndescribe your needs.\n\n##### Do I need a credit cart to start the free trial?\n\n![](images/icon-arrow-right.svg)\n\nScrapingAnt provides a completely free subscription plan which contains 10.000\nAPI credits that can be consumed during month. Until you will need more - it\nis completely free and doesn't require a credit card.\n\n### “Our clients are pleasantly surprised by the response speed of our team.”\n\n![](images/oleg-cartoon-image.jpg)\n\nOleg Kulyk,  \nScrapingAnt Founder\n\n* Our team will contact you ASAP.\n\nThank you! Your submission has been received!\n\nOops! Something went wrong while submitting the form.\n\n![](images/illustration-speed-lines-white.svg)\n\n## Grow your business with us\n\n[ Try Our Free Plan! ](https://app.scrapingant.com/signup)\n\n[\n\n## Features\n\n](https://scrapingant.com/#features) [\n\n## Pricing\n\n](https://scrapingant.com/#pricing) [\n\n## Blog\n\n](https://scrapingant.com/blog/) [\n\n## Documentation\n\n](https://docs.scrapingant.com/) [\n\n## Web Scraping API\n\n](https://scrapingant.com) [\n\n## LLM-ready web data\n\n](llm-ready-data-extraction.html) [\n\n## Residential Proxy\n\n](residential-proxies.html) [\n\n## Custom Scraper Development\n\n](https://scrapingant.com/custom-scraping-solution) [\n\n## Affiliate program\n\n](https://scrapingant.com/legal/affiliate/) [\n\n## Free proxies\n\n](https://scrapingant.com/free-proxies/)\n\n###### Web Scraping 101  \n\n[What is Web Scraping?](https://docs.scrapingant.com/web-scraping-101/what-is-\nweb-scraping) [**Is Web Scraping Legal?**](https://scrapingant.com/blog/is-\nweb-scraping-legal) [**10 Main Proxy\nTypes**](https://scrapingant.com/blog/main-proxy-types) [Datacenter vs\nResidential Proxies](https://scrapingant.com/blog/residential-vs-datacenter-\nproxy-webscraping) [Best Proxy Scraping\nTools](https://scrapingant.com/blog/top-open-source-proxy-scrapers)\n[**Overcoming scraping challenges with Web Scraping\nAPI**](https://scrapingant.com/blog/data-scraping-challenges) [IP rate-\nlimiting avoidance](https://scrapingant.com/blog/avoid-ip-rate-limiting)\n[Rotating proxies with Puppeteer](https://scrapingant.com/blog/how-to-use-\nrotating-proxies-with-puppeteer) [Scraping Dynamic Website with\nPython](https://scrapingant.com/blog/scrape-dynamic-website-with-python) [Web\nScraping with Python](https://scrapingant.com/blog/top-5-popular-python-\nlibraries-for-web-scraping-in-2020) [Web Scraping with\nJava](https://scrapingant.com/blog/web-scraping-java) [Web Scraping with\nNodeJS](https://scrapingant.com/blog/web-scraping-javascript) [Web Scraping\nwith Deno](https://scrapingant.com/blog/deno-web-scraping) [**Web Scraping\nwith R**](https://scrapingant.com/blog/r-web-scraping) [**Web Scraping with\nPHP**](https://scrapingant.com/blog/web-scraping-php) [**Web Scraping with\nGo**](https://scrapingant.com/blog/web-scraping-go)\n\n###### Use Cases  \n\n[**Real estate decisions with Booking.com\nscraping**](https://scrapingant.com/blog/booking-data-scraping) [**Sneaker\nPrice Data Collection with Web Scraping\nAPI**](https://scrapingant.com/blog/sneakers-scraping-api) [**Best Web\nScraping APIs For Freelancers**](https://scrapingant.com/blog/best-web-\nscraping-api-freelance) [**Smart NFT Decisions with Data\nCollection**](https://scrapingant.com/blog/nft-data-collection) [**How Data\nCollection Can Improve HR Processes**](https://scrapingant.com/blog/data-\ncollection-for-hr-processes) [**Rule eCommerce with Data\nCollection**](https://scrapingant.com/blog/data-collection-for-ecommerce)\n[**How companies use Web Scraping to gain a Competitive\nEdge**](https://scrapingant.com/blog/how-companies-use-web-scraping)\n[**Benefits of Web Scraping for\nHospitality**](https://scrapingant.com/blog/web-scraping-for-hospitality)\n[**Uses of Web Scraping for Price\nMonitoring**](https://scrapingant.com/blog/web-scraping-for-price-monitoring)\n[**Benefits of Web Scraping for Real\nEstate**](https://scrapingant.com/blog/web-scraping-for-real-estate) [**Web\nScraping for Data Scientists**](https://scrapingant.com/blog/web-scraping-for-\ndata-scientists) [**How to Collect Data from\nTikTok**](https://scrapingant.com/blog/web-scraping-for-price-monitoring)\n\n###### Legal  \n\n[Terms of Use](https://scrapingant.com/legal/terms-of-use) [Privacy\nPolicy](https://scrapingant.com/legal/privacy-policy) [Cookies\nPolicy](https://scrapingant.com/legal/cookies-policy)\n\n###### External Links  \n\n[Github](https://github.com/ScrapingAnt)\n[Linkedin](https://linkedin.com/company/scrapingant)\n[Facebook](https://www.facebook.com/scrapingant)\n[Twitter](https://twitter.com/ScrapingAnt)\n\n[![](images/ScrapingAnt-2.svg)](https://scrapingant.com)\n\n© Copyright ScrapingAnt  \nPowered by [DATAANT](https://scrapingant.com)\n\n![](images/lines-13-white.svg)\n\nBy browsing this site, you agree to our [Cookies\nPolicy](https://scrapingant.com/legal/cookies-policy)\n\n![](images/icon-x_1.svg)\n\n")]

加载

使用 load 方法抓取网页并获取提取的 markdown 内容。

# Load documents from URLs as markdown
documents = scrapingant_loader.load()

print(documents)

懒加载

使用 'lazy_load' 方法来抓取网页并懒惰地获取提取的 Markdown 内容。

# Lazy load documents from URLs as markdown
lazy_documents = scrapingant_loader.lazy_load()

for document in lazy_documents:
print(document)

API 参考

此加载器基于 ScrapingAnt Python SDK。有关更多配置选项,请参见 通用参数

相关


此页面是否有帮助?


您还可以留下详细的反馈 在 GitHub 上