All Projects → reichlab → Covid19 Forecast Hub

reichlab / Covid19 Forecast Hub

Licence: other
Projections of COVID-19, in standardized format

Projects that are alternatives of or similar to Covid19 Forecast Hub

Covid19 Dashboard
A site that displays up to date COVID-19 stats, powered by fastpages.
Stars: ✭ 1,212 (+254.39%)
Mutual labels:  github-pages, jupyter-notebook, analytics
Suspeitando
Projeto de análise de contratos com suspeita de superfaturamento e má qualidade na prestação de serviços.
Stars: ✭ 76 (-77.78%)
Mutual labels:  jupyter-notebook, analytics
Av Ltfs Data Science Finhack Ml Hackathon
L&T Financial Services & Analytics Vidhya presents ‘DataScience FinHack’ organised by Analytics Vidhya
Stars: ✭ 48 (-85.96%)
Mutual labels:  jupyter-notebook, analytics
Design Of Experiment Python
Design-of-experiment (DOE) generator for science, engineering, and statistics
Stars: ✭ 143 (-58.19%)
Mutual labels:  jupyter-notebook, analytics
Agile data code 2
Code for Agile Data Science 2.0, O'Reilly 2017, Second Edition
Stars: ✭ 413 (+20.76%)
Mutual labels:  jupyter-notebook, analytics
Sciblog support
Support content for my blog
Stars: ✭ 694 (+102.92%)
Mutual labels:  jupyter-notebook, analytics
Interactive machine learning
IPython widgets, interactive plots, interactive machine learning
Stars: ✭ 140 (-59.06%)
Mutual labels:  jupyter-notebook, analytics
Labml
🔎 Monitor deep learning model training and hardware usage from your mobile phone 📱
Stars: ✭ 508 (+48.54%)
Mutual labels:  jupyter-notebook, analytics
Kmeans pytorch
kmeans using PyTorch
Stars: ✭ 98 (-71.35%)
Mutual labels:  github-pages, jupyter-notebook
Web Database Analytics
Web scrapping and related analytics using Python tools
Stars: ✭ 175 (-48.83%)
Mutual labels:  jupyter-notebook, analytics
Stats Maths With Python
General statistics, mathematical programming, and numerical/scientific computing scripts and notebooks in Python
Stars: ✭ 381 (+11.4%)
Mutual labels:  jupyter-notebook, analytics
Introduction Datascience Python Book
Introduction to Data Science: A Python Approach to Concepts, Techniques and Applications
Stars: ✭ 275 (-19.59%)
Mutual labels:  jupyter-notebook, analytics
Ds With Pysimplegui
Data science and Machine Learning GUI programs/ desktop apps with PySimpleGUI package
Stars: ✭ 93 (-72.81%)
Mutual labels:  jupyter-notebook, analytics
Spark With Python
Fundamentals of Spark with Python (using PySpark), code examples
Stars: ✭ 150 (-56.14%)
Mutual labels:  jupyter-notebook, analytics
Fastpages
An easy to use blogging platform, with enhanced support for Jupyter Notebooks.
Stars: ✭ 2,888 (+744.44%)
Mutual labels:  github-pages, jupyter-notebook
Beeva Best Practices
Best Practices and Style Guides in BEEVA
Stars: ✭ 335 (-2.05%)
Mutual labels:  jupyter-notebook, analytics
Dl pytorch
Code for the Deep Learning with PyTorch lesson
Stars: ✭ 339 (-0.88%)
Mutual labels:  jupyter-notebook
Dsprites Dataset
Dataset to assess the disentanglement properties of unsupervised learning methods
Stars: ✭ 340 (-0.58%)
Mutual labels:  jupyter-notebook
License plate detection pytorch
A two stage lightweight and high performance license plate recognition in MTCNN and LPRNet
Stars: ✭ 337 (-1.46%)
Mutual labels:  jupyter-notebook
Numpile
A tiny 1000 line LLVM-based numeric specializer for scientific Python code.
Stars: ✭ 341 (-0.29%)
Mutual labels:  jupyter-notebook

COVID-19 Forecast Hub

Node.js CI DOI

The goal of this repository is to create a standardized set of data on forecasts from teams making projections of cumulative and incident deaths and incident hospitalizations due to COVID-19 in the United States. This repository is the data source for the official CDC COVID-19 Forecasting page. This project to collect, standardize, visualize and synthesize forecast data has been led by the CDC-funded UMass-Amherst Influenza Forecasting Center of Excellence based at the Reich Lab, with contributions from many others.

This README provides an overview of the project. Additional specific links can be found in the list below:

chart

Data license and reuse

We are grateful to the teams who have generated these forecasts. They have spent a huge amount of time and effort in a short amount of time to operationalize these important real-time forecasts. The groups have graciously and courageously made their public data available under different terms and licenses. You will find the licenses (when provided) within the model-specific folders in the data-processed directory. Please consult these licenses before using these data to ensure that you follow the terms under which these data were released.

All source code that is specific to this project, along with our d3-foresight visualization tool is available under an open-source MIT license. We note that this license does NOT cover model code from the various teams (maybe available from them under other licenses) or model forecast data (available under specified licenses as described above).

To cite the COVID-19 Forecast Hub, please use a relevant research article or preprint produced by the group and a permanent DOI for the GitHub repo (the DOI is updated by Zenodo when we create a new "release" of this GitHub repository).

What forecasts we are tracking, and for which locations

Different groups are making forecasts at different times, and for different geographic scales. The specifications below were created by consulting with collaborators at CDC and looking at what models forecasting teams were already building.

What do we consider to be "gold standard" data? We will use the daily reports containing case and death data from the JHU CSSE group as the gold standard reference data for deaths in the US. These data are the time-series version of the JHU data that do occasionally contain "revisions" of previous daily reports. Note that there are not insignificant differences (especially in daily incident death data) between the JHU data and another commonly used source, from the New York Times. The team at UTexas-Austin is tracking this issue on a separate GitHub repository.

When will forecast data be updated? We will be storing new forecasts from each group as they are either provided to us directly via pull requests. Teams are encouraged to submit data as often has they have it available, although we only support one upload for each day. In general, "updates" to forecasts will not be permitted. Teams are responsible for checking that their forecasts are ready for public viewing upon submission. This can be done locally using our interactive visualization tool.

What locations will have forecasts? Currently, forecasts may be submitted for any state and county in the US and the US at the national level.

How will probabilistic forecasts be represented? Forecasts will be represented in a standard format using quantile-based representations of predictive distributions. We encourage all groups to make available the following 23 quantiles for each distribution: c(0.01, 0.025, seq(0.05, 0.95, by = 0.05), 0.975, 0.99). One goal of this effort is to create probabilistic ensemble forecasts, and having high-resolution component distributions will provide data to create better ensembles.

What forecast targets will be stored? We will store forecasts for 1 through 20 week-ahead incident and cumulative deaths, 0 through 130 day-ahead incident hospitalizations, and 1 through 8 week-ahead incident reported cases. Please refer to the technical README for details on aligning targets with forecast dates.

Ensemble model

Every Monday at 6pm ET, we will update our COVID Forecast Hub ensemble forecast and interactive visualization using the most recent forecast from each team as long as it was submitted before 6pm ET on Monday and has a forecast_date of any day since the previous Tuesday. All models meeting the above criteria will be considered for the ensemble. For inclusion in the ensemble, we additionally require that forecasts include a full set of 23 quantiles to be submitted for each of the one through four week ahead values for forecasts of deaths, and a full set of 7 quantiles for the one through four week ahead values for forecasts of cases (see technical README for details), and that the 10th quantile of the predictive distribution for a 1 week ahead forecast of cumulative deaths is not below the most recently observed data. Before the week of July 28, we also performed manual visual inspection checks to ensure that forecasts were in alignment with the ground truth data; this step is no longer a part of our weekly ensemble generation process. Details on which models were included each week in the ensemble are available in the ensemble metadata folder.

From April 13 to July 21, the ensemble was created by taking the arithmetic average of each prediction quantile for all eligible models for a given location. Starting on the week of July 28, we have instead used the median prediction across all eligible models at each quantile level.

Forecast files

Participating teams provide their forecasts in a quantile-based format. We have developed specifications that can be used to represent all of the forecasts in a simple, long-form data format. For details about this file format specifications, please see the technical README.

Teams and models

Our list of teams whose forecasts are currently standardized and in the repository are (with data reuse license):

Participating teams must provide a metadata file.

The COVID Forecast Hub Team

Carefully curating these datasets into a standard format has taken a Herculean team effort. The following lists those who have helped out, in reverse alphabetical order:

  • Nutcha Wattanachit (ensemble model, data processing)
  • Serena Wang (data curation)
  • Nicholas Reich (project lead, ensemble model, data processing)
  • Evan Ray (ensemble model)
  • Jarad Niemi (data processing and organization)
  • Khoa Le (validation, automation)
  • Ayush Khandelwal (architecture, data curation)
  • Abdul Hannan Kanji (architecture, data curation)
  • Katie House (visualization, validation, project management)
  • Estee Cramer (data curation, ensemble model)
  • Matt Cornell (validation, Zoltar integration)
  • Andrea Brennen (metadata curation)
  • Johannes Bracher (evaluation, data processing)
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].