All Projects → nstapelbroek → estate-crawler

nstapelbroek / estate-crawler

Licence: BSD-3-Clause license
Scraping the real estate agencies for up-to-date house listings as soon as they arrive!

Programming Languages

python
139335 projects - #7 most used programming language
Makefile
30231 projects
Dockerfile
14818 projects
shell
77523 projects

Projects that are alternatives of or similar to estate-crawler

Python3 Spider
Python爬虫实战 - 模拟登陆各大网站 包含但不限于:滑块验证、拼多多、美团、百度、bilibili、大众点评、淘宝,如果喜欢请start ❤️
Stars: ✭ 2,129 (+10545%)
Mutual labels:  scrapy, scrapy-crawler
IMDB-Scraper
Scrapy project for scraping data from IMDB with Movie Dataset including 58,623 movies' data.
Stars: ✭ 37 (+85%)
Mutual labels:  scrapy, scrapy-crawler
Weibospider
This is a sina weibo spider built by scrapy [微博爬虫/持续维护]
Stars: ✭ 2,408 (+11940%)
Mutual labels:  scrapy
Scrapy Splash
Scrapy+Splash for JavaScript integration
Stars: ✭ 2,666 (+13230%)
Mutual labels:  scrapy
Gerapy
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js
Stars: ✭ 2,601 (+12905%)
Mutual labels:  scrapy
Scrapydweb
Web app for Scrapyd cluster management, Scrapy log analysis & visualization, Auto packaging, Timer tasks, Monitor & Alert, and Mobile UI. DEMO 👉
Stars: ✭ 2,385 (+11825%)
Mutual labels:  scrapy
Ruiji.net
crawler framework, distributed crawler extractor
Stars: ✭ 220 (+1000%)
Mutual labels:  scrapy
Weibospider sentimentanalysis
借助Python抓取微博数据,并对抓取的数据进行情绪分析
Stars: ✭ 173 (+765%)
Mutual labels:  scrapy
Spider job
招聘网数据爬虫
Stars: ✭ 234 (+1070%)
Mutual labels:  scrapy
Py Elasticsearch Django
基于python语言开发的千万级别搜索引擎
Stars: ✭ 207 (+935%)
Mutual labels:  scrapy
Spiderkeeper
admin ui for scrapy/open source scrapinghub
Stars: ✭ 2,562 (+12710%)
Mutual labels:  scrapy
Github Spider
Github 仓库及用户分析爬虫
Stars: ✭ 190 (+850%)
Mutual labels:  scrapy
Livetv mining
直播网站数据采集
Stars: ✭ 188 (+840%)
Mutual labels:  scrapy
City Scrapers
Scrape, standardize and share public meetings from local government websites
Stars: ✭ 220 (+1000%)
Mutual labels:  scrapy
Marmot
💐Marmot | Web Crawler/HTTP protocol Download Package 🐭
Stars: ✭ 186 (+830%)
Mutual labels:  scrapy
Filesensor
Dynamic file detection tool based on crawler 基于爬虫的动态敏感文件探测工具
Stars: ✭ 227 (+1035%)
Mutual labels:  scrapy
Wenshu spider
🌈Wenshu_Spider-Scrapy框架爬取中国裁判文书网案件数据(2019-1-9最新版)
Stars: ✭ 177 (+785%)
Mutual labels:  scrapy
News spider
新闻抓取(微信、微博、头条...)
Stars: ✭ 190 (+850%)
Mutual labels:  scrapy
Stealer
抖音、快手、火山、皮皮虾,视频去水印程序
Stars: ✭ 217 (+985%)
Mutual labels:  scrapy
Awesome crawl
腾讯新闻、知乎话题、微博粉丝,Tumblr爬虫、斗鱼弹幕、妹子图爬虫、分布式设计等
Stars: ✭ 246 (+1130%)
Mutual labels:  scrapy

Estate Crawler

Finding a place to live can be tough. If for any reason you are unable to buy a house, the search for a rental apartment can cost you a lot of time and money. Real estate agencies will often charge you before you've even visited your first potential "object". The subscriptions they force upon you are unfair and do not guarantee success of finding a place to live.

This project aims to relieve some pain by giving you insights of the offers at several agencies. By collecting and comparing this data, you can pick an agency that is best suited for your needs. Saving you time and potentially money.

Usage

You can run the estate crawler by executing the crawler.py file. Below is an example with Python3 and pipenv installed on your OS:

pipenv install && pipenv run ./crawler.py --region amsterdam,rotterdam,arnhem

If you have the luxury of Docker available on your system, you can run the crawler without installing any dependencies like so:

docker run --rm -v $(pwd):/app/build docker.io/nstapelbroek/estate-crawler --region amsterdam

Results of your crawl run are available in ./build/results.json after a successful run. You can change this path by passing a --output-file argument. Note that the output file does not contain valid json, only valid json lines.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].