Skip to content

cuyu/scrapy-lianjia

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

scrapy lianjia

This a project using scrapy to crawl data from 链家网.

How to use

  1. Change directory to the root directory of this project.
  2. Use scrapy crawl lianjia -o outputs.csv to save the crawled data into a .csv(also support .json and .xml format) file.

Some scrapy commands for debugging

  • scrapy shell URL can be used for debugging specific url
  • scrapy crawl lianjia -s LOG_FILE=scrapy.log will save the logging info into a file

TODO list

  • Add analysis for different houses of the same time
  • Add analysis for the same house of different time
  • Add function of scheduling the script
  • Use proxy to crawl data more quickly

About

No description or website provided.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages