项目作者: Shraddhaduwal

项目描述 :
This is a data-crawler made in Scrapy. The data from local newspapers is crawled for data processing purpose. There are a number of spiders through which we can start the data crawling. The crawled data is stored in csv files. The files can also be stored in other formats like .json, .txt etc as per the convenience.
高级语言: Python
项目地址: git://github.com/Shraddhaduwal/scrapy-crawler.git
创建时间: 2020-03-02T07:28:46Z
项目社区:https://github.com/Shraddhaduwal/scrapy-crawler

开源协议:

下载


scrapy-crawler

This is a data-crawler made in Scrapy. The data from local newspapers is crawled for data processing purpose. There are a number of spiders through which we can start the data crawling. The crawled data is stored in csv files. The files can also be stored in other formats like .json, .txt etc as per the convenience.

Start Scrapy Project

scrapy startproject <project_name>

Run Spider

After navigating to the parent folder <project_name> :

  • scrapy crawl <spider_name>

Save the crawled data to different files

  • scrapy crawl <spider_name> -o <filename>.csv
  • scrapy crawl <spider_name> -o <filename>.json