All Projects β†’ covidatlas β†’ Coronadatascraper

covidatlas / Coronadatascraper

Licence: bsd-2-clause
COVID-19 Coronavirus data scraped from government and curated data sources.

Projects that are alternatives of or similar to Coronadatascraper

instagram explorer
πŸ“· An app to scrap instagram posts and analyze data.
Stars: ✭ 17 (-95.43%)
Mutual labels:  scraping
Edu Mail Generator
Generate Free Edu Mail(s) within minutes
Stars: ✭ 301 (-19.09%)
Mutual labels:  scraping
Autoscraper
A Smart, Automatic, Fast and Lightweight Web Scraper for Python
Stars: ✭ 4,077 (+995.97%)
Mutual labels:  scraping
Apify Js
Apify SDK β€” The scalable web scraping and crawling library for JavaScript/Node.js. Enables development of data extraction and web automation jobs (not only) with headless Chrome and Puppeteer.
Stars: ✭ 3,154 (+747.85%)
Mutual labels:  scraping
Clean Text
🧹 Python package for text cleaning
Stars: ✭ 284 (-23.66%)
Mutual labels:  scraping
Linkedin
Linkedin Scraper using Selenium Web Driver, Chromium headless, Docker and Scrapy
Stars: ✭ 309 (-16.94%)
Mutual labels:  scraping
jazz
The Scripting Engine that Combines Speed, Safety, and Simplicity
Stars: ✭ 132 (-64.52%)
Mutual labels:  scraping
Comic Dl
Comic-dl is a command line tool to download manga and comics from various comic and manga sites. Supported sites : readcomiconline.to, mangafox.me, comic naver and many more.
Stars: ✭ 365 (-1.88%)
Mutual labels:  scraping
Sasila
δΈ€δΈͺ灡活、友ε₯½ηš„ηˆ¬θ™«ζ‘†ζžΆ
Stars: ✭ 286 (-23.12%)
Mutual labels:  scraping
Tinking
🧢 Extract data from any website without code, just clicks.
Stars: ✭ 331 (-11.02%)
Mutual labels:  scraping
Gopa
[WIP] GOPA, a spider written in Golang, for Elasticsearch. DEMO: http://index.elasticsearch.cn
Stars: ✭ 277 (-25.54%)
Mutual labels:  scraping
Scrapy Crawlera
Crawlera middleware for Scrapy
Stars: ✭ 281 (-24.46%)
Mutual labels:  scraping
Spidermon
Scrapy Extension for monitoring spiders execution.
Stars: ✭ 309 (-16.94%)
Mutual labels:  scraping
schedule-tweet
Schedules tweets using TweetDeck
Stars: ✭ 14 (-96.24%)
Mutual labels:  scraping
Socialreaper
Social media scraping / data collection library for Facebook, Twitter, Reddit, YouTube, Pinterest, and Tumblr APIs
Stars: ✭ 338 (-9.14%)
Mutual labels:  scraping
facebook-discussion-tk
A collection of tools to (semi-)automatically collect and analyze data from online discussions on Facebook groups and pages.
Stars: ✭ 33 (-91.13%)
Mutual labels:  scraping
Elixir Scrape
Scrape any website, article or RSS/Atom Feed with ease!
Stars: ✭ 306 (-17.74%)
Mutual labels:  scraping
Post Tuto Deployment
Build and deploy a machine learning app from scratch πŸš€
Stars: ✭ 368 (-1.08%)
Mutual labels:  scraping
Katana
A Python Tool For google Hacking
Stars: ✭ 355 (-4.57%)
Mutual labels:  scraping
Social Media Profiles Regexs
πŸ“‡ Extract social media profiles and more with regular expressions
Stars: ✭ 324 (-12.9%)
Mutual labels:  scraping

THIS PROJECT IS BEING REPLACED BY Li

This project is being replaced by Li, the next-generation serverless crawler for COVID-19 data.

The reasons for the switchover are documented in Issue 782.

We are not actively accepting PRs for this repository, but are still using this repo to track issues.

Scraper code written for this project is not compatible with Li, see that project for examples. We have some helpers to assist in migrating code, see this document.

Thanks very much!


coronadatascraper

A crawler that scrapes COVID-19 Coronavirus data from government and curated data sources.

This project exists to scrape, de-duplicate, and cross-check county-level data on the COVID-19 coronavirus pandemic.

Every piece of data includes GeoJSON and population data, cites the source from which the data was obtained, and includes a rating of the source's technical quality (completeness, machine readability, best practices -- not accuracy).

Where's the data?

https://coronadatascraper.com/

How often is it updated?

We upload fresh data every day at around 9PM PST.

How do I use this data?

Read the Data Fields documentation for details on exactly what each field in the dataset means.

How can I run the crawler locally?

Check out our Getting Started guide to help get our project running on your local machine.

Contributing

NOTE: This project is being replaced by Li, the next-generation serverless crawler for COVID-19 data.

You can contribute to this project in two big ways:

Contribute to the project core

Check the Issues for any task we need to get done. If you are new to open-source, look for the label Good first issue

Contribute a source

Contributions for any place in the world are welcome. See the community-curated list of verified data sources to find a new datasource to add, and be sure to update the "Scraped?" column when you do.

To help you contribute a new source, please read the Sources and Scrapers guide before you start!

Send a pull request with your scraper, and be sure to run the scraper first with the instructions specified in the guide to make sure the data is valid.

License

This project is licensed under the permissive BSD 2-clause license.

The data produced by this project is public domain.

This project uses data from ISO-3166 Country and Dependent Territories Lists with UN Regional Codes under the Creative Commons Attribution-ShareAlike 4.0 International License.

Attribution

Please cite this project if you use it in your visualization or reporting.

Data obtained from Corona Data Scraper

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].