You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
FTR: I've recently created a middleware similar to deltafetch, but which is more explicit: https://github.com/TeamHG-Memex/scrapy-crawl-once. It does a similar thing, but in a less automatic way - user needs to set request.meta['crawl_once'] = True. I considered contributing to scrapy-deltafetch instead, but implementations have almost nothing in common (sqlite vs bsddb, items vs meta keys, different options).
@redapple , Hi I am new here. I have a project in which i used deltafetch,
is there a way, we can specify main url and some sub urls that the spider visit again ?
I am using my spider to scrape data periodically. Can you please help me ?
See scrapinghub/scrapylib#45 (comment) for motivation.
It can be counter-intuitive for newcomers that the middleware will let the spider revisit pages if they did not produce any item.
The text was updated successfully, but these errors were encountered: