I am learning Scrapy a web crawling framework.
by default it does not crawl duplicate urls or urls which scrapy have already crawled.
How to make Scrapy to crawl duplicate urls or urls which have already crawled?
I tried to find out on internet but could not find relevant help.
I found DUPEFILTER_CLASS = RFPDupeFilter
and SgmlLinkExtractor
from Scrapy - Spider crawls duplicate urls but this question is opposite of what I am looking
See Question&Answers more detail:
os 与恶龙缠斗过久,自身亦成为恶龙;凝视深渊过久,深渊将回以凝视…