All Projects → sfbrigade → datasci-housing-pipeline

sfbrigade / datasci-housing-pipeline

Licence: other
Help us analyze data to understand why it takes so long to build housing in the Bay Area!

Projects that are alternatives of or similar to datasci-housing-pipeline

adopt-a-drain
A web application that allows citizens to "adopt" a storm drain in San Francisco. In use, and in development at other brigades. Looking for a maintainer or someone interested in developing further in collaboration with others across the country.
Stars: ✭ 43 (+258.33%)
Mutual labels:  code-for-sf, code-for-america
HomeUniteUs
We're working with community non-profits who have a Host Home or empty bedrooms initiative to develop a workflow management tool to make the process scalable (across all providers), reduce institutional bias, and effectively capture data.
Stars: ✭ 22 (+83.33%)
Mutual labels:  code-for-america
ballotnav
A repository for HackforLA's BallotNav project
Stars: ✭ 21 (+75%)
Mutual labels:  code-for-america
clean-slate-data
No description or website provided.
Stars: ✭ 27 (+125%)
Mutual labels:  code-for-america
CTI-website-frontend
Join a worldwide movement to catalog 
every open source 
civic tech project.
Stars: ✭ 28 (+133.33%)
Mutual labels:  code-for-america
recordexpungPDX
A project to automate the expungement of qualifying criminal records in Oregon. This project is done in conjunction with Qiu-Qiu Law.
Stars: ✭ 45 (+275%)
Mutual labels:  code-for-america
windfall-elimination
Windfall Elimination Provision Awareness Project: improving the experience of retirees around an obscure Social Security rule affecting 8+ states (previously https://ssacalculator.org)
Stars: ✭ 23 (+91.67%)
Mutual labels:  code-for-america
311-data
Empowering Neighborhood Associations to improve the analysis of their initiatives using 311 data
Stars: ✭ 48 (+300%)
Mutual labels:  code-for-america
openbudgetoakland
Visualizations of Oakland's budget data, and explanations about the budget process.
Stars: ✭ 91 (+658.33%)
Mutual labels:  code-for-america
lucky-parking
Visualization of parking data to assist in understanding of the effects of parking policies on a neighborhood by neighborhood basis in the City of Los Angeles
Stars: ✭ 24 (+100%)
Mutual labels:  code-for-america
Neighborhood-Dashboard
We're working to help neighborhoods help themselves by giving them easy access to useful data to allow them to identify and track problem areas in their neighborhood.
Stars: ✭ 19 (+58.33%)
Mutual labels:  code-for-america
heart
Heart is a project working directly with the LA City Attorney’s Homeless Engagement and Response Team. We are building a database and case management system to streamline their workflow and enable them to scale their program. Find us on the Hack for LA Slack #heart.
Stars: ✭ 27 (+125%)
Mutual labels:  code-for-america
food-oasis
Repository for the current redevelopment of the Food Oasis Los Angeles website
Stars: ✭ 33 (+175%)
Mutual labels:  code-for-america
woeip
A platform for impacted communities to understand their local air quality and advocate for environmental justice.
Stars: ✭ 26 (+116.67%)
Mutual labels:  code-for-america
members
Online portal for Code for Denver members
Stars: ✭ 15 (+25%)
Mutual labels:  code-for-america
communityconnect
No description or website provided.
Stars: ✭ 23 (+91.67%)
Mutual labels:  code-for-america
tdm-calculator
DTLA Hack for LA is partnering with Los Angeles Department of Transportation (LADOT) to develop a Traffic Demand Management (TDM) calculator tool. This tool will help planners at LADOT and real estate developers to meet the Los Angeles’s Mobility Plan goals by 2035.
Stars: ✭ 21 (+75%)
Mutual labels:  code-for-america
CampaignFinance
NC Campaign Finance Dashboard. Making NC Campaign Funds visible to all citizens
Stars: ✭ 36 (+200%)
Mutual labels:  code-for-america
migrant service map
Refugees welcome
Stars: ✭ 23 (+91.67%)
Mutual labels:  code-for-america

Housing Pipeline Analysis

This project is a part of the Data Science Working Group at Code for San Francisco. Other DSWG projects can be found at the main GitHub repo.

Project Status: Active

Project Intro/Objective

The purpose of this project is to analyze the impact of zoning laws on the development of housing in San Francisco. We use datasets provided by the city of San Francisco to analyze the initiation, progress, and completion of entitlements for the construction of housing. We want to empower policymakers and citizens with the facts and data that help explain the potential impact of housing policy decisions.

See our Doc for more info and specific objectives for the project.

Methods Used

  • Inferential Statistics
  • Data Visualization
  • Predictive Modeling

Technologies

  • Python
  • Pandas, jupyter

Project Description

We focus on utilizing our distinct skillset of data science and the SF Planning Department’s pipeline data to a) develop quantitative facts about the state of the pipeline of planned housing in SF, and b) use predictive methods to model potential impact of policies. The potential impact could take multiple forms, such as in units built, free market price, and affordable unit availability.

Guiding Principles

  • Accessible San Francisco -- We are invested in providing facts that allow other parties, some political, to make data-informed decisions that enable San Francisco to become more inclusive and therefore healthier.
  • Science is A-political -- While we do believe in an accessible San Francisco, we are not a policy advocate. Our role is to provide the most rigorous data analysis to define facts, and use sound statistical methods and scientific investigation to predict how those facts may shift given a policy change.
  • Correlation is not causation -- Hand-in-hand with being a-political is the responsibility to clearly articulate the findings and limitations of analysis. Most analysis and prediction will only be able to leverage correlative relationships, and will likely be unable to demonstrate causative relationships. Natural experiments that would give us this kind of increased conviction in causative factors are rare.
  • Equal Access to Facts -- Our work will be shared through blog format to be accessible to all stakeholders at the same time.
  • Open to Everyone -- We are an all-volunteer organization, and our team welcomes all people of all skill bases and backgrounds to join our team. There is always a way for you to contribute!

Needs of this project

The best place to get started is our list of Issues in Github. We have workstreams spread across these skillsets:

Data Scientists

Data modeling and analysis towards a better understanding of housing pipeline changes over time.

Python Programmers

Creating data cleanup tools and methods for parsing or scraping new datasets.

Econometrists, Economic Modeling

Model market incentives and assumptions, enabling prediction of impact based on changes.

Operations

Manage the volunteer network, communicate with stakeholders, and network with new people to learn new needs.

Visual Designer

Help create visualizations of facts and predictions with Data Scientists

Policy Researcher

Help us understand the policy strategies being discussed and work with the PMs on whether we should investigate them.

Marketing

Help us publicize our work and build a stronger community overall.

The Repo

  1. Raw Data is being kept here within this repo.
  2. Data processing/transformation scripts are being kept here
  3. The Jupyter notebook contains prior analyses

Getting Started

Quickstart

Step 1: Clone the repo

  1. Navigate to a folder where you want the project folder to be located
  2. Clone the repo with the following command
git clone [email protected]:sfbrigade/datasci-housing-pipeline.git

for help see this tutorial).

  1. navigate into your newly created project folder
cd datasci-housing-pipeline

Step 2: Install Dependencies

We use Pipenv for environment management, follow the installation guides below if you don't have it.

Install all project and development dependencies:

pipenv install --dev

Step 2: Enter The Development Shell

pipenv run python -m ipykernel install --user --name=`pipenv run basename '$VIRTUAL_ENV'`

Step 3: Open the Notebook

Launch Jupyter and select datasci-housing-pipeline kernel in Jupyter.

jupyter notebook

Installation Guide

macOS

Step 1: Ensure you have Python 3.7 installed

Check your currently installed version of Python 3.

python3 --version

If you don't have Python version 3.7, the Pipfile will not complete installation. Install Python 3 using Homebrew:

brew install python3

Or, upgrade Python 3 from an earlier dot version (like 3.6) using Homebrew:

brew upgrade python3

Step 2: Install Pipenv using Homebrew

Run this in your terminal:

brew install pipenv

NOTE: dependencies will only be available within the pipenv virtualenv. Enter the virtualenv with pipenv shell, or run a single command with pipenv run my-cool-command.

Step 3: Install Postgres using Homebrew

Run this in your terminal:

brew install postgresql

Postgres should start automatically. If you run into trouble, refer to this guide.

Windows

Suggestion: Install a console emulator running on ConEmu

My personal recommendation is Cmder

Step 1: Install Chocolatey, a package manager for windows

Install chocolatey

Step 2: Ensure you have Python 3.7 installed

Check your currently installed version of Python.

python --version

If you don't have Python version 3.7, install or upgrade to Python 3 using Chocolatey:

choco install python

Step 3: Install Pipenv using pip

Python3 should install pip automatically, but check for updates with the following command:

python -m pip install -U pip

Now install pipenv with a User installation:

pip install --user pipenv

NOTE: If pipenv isn't available in your console after installing and running refreshenv you will need to add the user base's binary directory to your PATH. This is relatively simple, read the Yellow Box on this tutorial page

NOTE 2: dependencies will only be available within the pipenv virtualenv. Enter the virtualenv with pipenv shell, or run a single command with pipenv run my-cool-command.

Step 4: Install Postgres using Chocolatey

Postgres requires a password parameter, so run the following command, with your own password to be assigned to the postgres user:

choco install postgresql10 --params '/Password:YOURPASSWORDHERE' --params-global

Postgres should start automatically. If you run into trouble, refer to the Postgres website.

Contributing DSWG Members

Team Leads (Contacts)

Previous Leads:

Other Members:

Name Slack Handle
Andrew Roberts @Andrew Roberts

Contact

  • If you haven't joined the SF Brigade Slack, you can do that here.
  • Our slack channel is #datasci-projectname
  • Feel free to contact team leads with any questions or if you are interested in contributing!
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].