Skip to content

grainrain/BasicCrawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 

Repository files navigation

scrapy-fake-useragent 依赖 fake-useragent 生成用户代理,但 fake-useragent==2.0.0 开始对其内部实现做了较大改动,可能导致兼容性问题。
如果 scrapy-fake-useragent 版本是 1.4.4,需要使用fake-useragent==1.1.0。

pip install scrapy==2.11.2

pip install fake-useragent==1.1.0

pip install scrapy_fake_useragent==1.4.4

pip install pipreqs

pipreqs . --encoding=utf8 --force

scrapy crawl BasicCrawlerSpider -o basic.json

scrapy crawl BasicCrawlerSpider -o basic.csv -s CLOSESPIDER_ITEMCOUNT=90

scrapy crawl BasicCrawlerSpider -o basic.xml

scrapy crawl BasicCrawlerSpider -s FEED_URI='d:\basic.csv' -s FEED_FORMAT=csv

scrapy crawl BasicCrawlerSpider -s FEED_URI='d:\basic.json' -s FEED_FORMAT=json

About

BasicCrawler

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages