Bring web data extraction into your IDE or coding assistant.
Access, extract, and structure data as part of your normal workflow.
Bare Claude knows Python. It doesn't know Scrapy the way Zyte engineers do. The plugin teaches it — async patterns, scrapy-poet page objects, the right libraries, real tests, anti-bot wired in. The spider that comes out of the prompt actually runs. Describe data. Get data.
Claude on its own
Generic AI coding assistant
With Zyte skills
Domain-expert AI system
A spider that runs first time
Get a runnable Scrapy spider for a new site without correction turns.
Gets modern Scrapy patterns wrong
Writes simple & non testable code, and fails to include key addons.
Spider runs first time.
Correct async Scrapy patterns, scrapy-poet page objects, targeted selectors, testable code with pytest.
Code that holds up to review
Extraction shaped like a senior scraping engineer would write it.
Best-guess extraction
Often decides your data shape for you, and hallucinates selectors, leaving you with questionable quality.
Built to Zyte's quality bar
Code shaped like the spiders Zyte engineers ship — accurate fields, clean patterns, scalable production spiders.
Anti-bot handled out of the box
No 403s and no unblocking layer to write yourself.
No anti-bot solution
You wire one in by hand or watch the spider fail at the first ban.
Zyte API wired in by default
Bring your own if you prefer, but it works out of the box.
Testable code generation
pytest to guarantee your code.
Doesn't know how to create HTML fixtures.
No tests, or tests built.
Tests scaffolded with the spider
pytest tests scaffolded against saved HTML fixtures. Tests fail? Prompt a fix and run again.
Scrapy stack awareness
The right library at the right version, wired through correctly with our most important addons.
Produces generic Scrapy.
Doesn't know scrapy-poet, zyte-common-items, scrapy-zyte-api, or the addons system.
Reaches for the right library and version
Wires them through the addons system, not legacy middleware.
See how the Zyte Claude Code plugin turns a website URL into a working, tested Scrapy scraper in under 15 minutes — ready to deploy and run.
Web Scraping Copilot is built to reduce friction between idea and extraction. No context switching. No black boxes. No hidden automation.
Get started with Zyte plugins, from setup to building production-ready scraping workflows.
Explore opinions, lessons, and emerging approaches to extracting and managing web data at scale.

Built for developers and data teams who need reliable, production-ready web extraction fast. It's especially useful if you want Scrapy + web-poet scaffolding without writing boilerplate from scratch, a reusable extraction spec before committing to any specific implementation, or a "data-first" workflow where you validate what you're extracting before generating code.
Three things: Claude Code (CLI or desktop app), uv Python package manager (the plugin will prompt you to install it if it's missing), and a Zyte API key — recommended for fetching and rendering pages, but not strictly required to run the plugin.
The process has two clearly separated stages. First, the extraction spec: the agent fetches pages, discovers fields, and iterates with you until you approve the schema and expected values. Second, code generation: the approved spec is used to produce a complete Scrapy project (page objects, test fixtures, spider).
This stage does not crawl the web — it works entirely from the outputs of stage 1. Crawling only happens when you run the generated spider locally, or deploy and run it on Scrapy Cloud.
You get two artifacts: a .scrape/{site-name}/ folder containing the extraction spec (schema, saved pages, and approved expected values), and a fully generated Scrapy project with web-poet page objects, pytest fixtures, and a wired-up spider ready to run — including an option to deploy to Scrapy Cloud.
Not yet — Claude Code is the only supported agent today. Support for additional agents is coming soon.
Yes. The spec is a technology-neutral artifact that isn't tied to any framework. It can feed a plain-Python implementation, direct Zyte API extraction, or a fully manual process. The generated Scrapy project is one possible consumer — not the only one.
The spec is the durable part of your project. When a site changes, you re-run the spec stage (or update field selectors in the page objects) without discarding the rest of your project structure. The deliberate separation of spec from generated code is what makes the system resilient to change.
Yes. The plugin delegates page fetching and rendering to the Zyte API, so JavaScript-rendered content is handled transparently — no extra configuration needed.
No, the plugin can run without one. However, we recommend using the Zyte API for fetching and extraction — it handles JavaScript rendering, anti-bot measures, and reliability, giving you significantly better results on complex sites.
Yes, the plugin is Free to use. However, we recommend using the Zyte API for fetching and extraction — it handles JavaScript rendering, anti-bot measures, and reliability, giving you significantly better results on complex sites.
G2.com