All Projects β†’ librariesio β†’ metrics

librariesio / metrics

Licence: AGPL-3.0, AGPL-3.0 licenses found Licenses found AGPL-3.0 LICENSE AGPL-3.0 LICENSE.txt
πŸ“ˆ What to measure, how to measure it.

Projects that are alternatives of or similar to metrics

Infinite Stories with Data
This repo consists of my analysis of random datasets using various statistical and visualization techniques.
Stars: ✭ 21 (+50%)
Mutual labels:  data-analysis
iMOKA
interactive Multi Objective K-mer Analysis
Stars: ✭ 19 (+35.71%)
Mutual labels:  data-analysis
advanced-kpi
Advanced-KPI is about creating a smart KPI object that fits to 90% of the needs of Qlik Sense users.
Stars: ✭ 19 (+35.71%)
Mutual labels:  measure
python ml tutorial
A complete tutorial in python for Data Analysis and Machine Learning
Stars: ✭ 118 (+742.86%)
Mutual labels:  data-analysis
Moose
MOOSE - Platform for software and data analysis.
Stars: ✭ 110 (+685.71%)
Mutual labels:  data-analysis
elucidate
convenience functions to help researchers elucidate patterns in their data
Stars: ✭ 26 (+85.71%)
Mutual labels:  data-analysis
signal-estimator
Measure characteristics of a looped back signal.
Stars: ✭ 37 (+164.29%)
Mutual labels:  measure
tianchi-diabetes
ε€©ζ± η²Ύε‡†εŒ»η–—ε€§θ΅›β€”β€”δΊΊε·₯ζ™Ίθƒ½θΎ…εŠ©η³–ε°Ώη—…ι—δΌ ι£Žι™©ι’„ζ΅‹ 第一衛季
Stars: ✭ 20 (+42.86%)
Mutual labels:  data-analysis
mixedvines
Python package for canonical vine copula trees with mixed continuous and discrete marginals
Stars: ✭ 36 (+157.14%)
Mutual labels:  data-analysis
RepSeP
Reproducible Self-Publishing - Demo Publications in the Most Common Formats
Stars: ✭ 14 (+0%)
Mutual labels:  data-analysis
Chapter-2
Code examples for Chapter 2 of Data Wrangling with JavaScript
Stars: ✭ 16 (+14.29%)
Mutual labels:  data-analysis
ospi
Open Source Presence Infographic of Indian Startups
Stars: ✭ 25 (+78.57%)
Mutual labels:  data-analysis
dataquest-guided-projects-solutions
My dataquest project solutions
Stars: ✭ 35 (+150%)
Mutual labels:  data-analysis
tutorials
Short programming tutorials pertaining to data analysis.
Stars: ✭ 14 (+0%)
Mutual labels:  data-analysis
computational-neuroscience
Short undergraduate course taught at University of Pennsylvania on computational and theoretical neuroscience. Provides an introduction to programming in MATLAB, single-neuron models, ion channel models, basic neural networks, and neural decoding.
Stars: ✭ 36 (+157.14%)
Mutual labels:  data-analysis
DataProfiler
What's in your data? Extract schema, statistics and entities from datasets
Stars: ✭ 843 (+5921.43%)
Mutual labels:  data-analysis
advanced-pandas
Pandas is a powerful tool for data exploration and analysis (including timeseries).
Stars: ✭ 22 (+57.14%)
Mutual labels:  data-analysis
LeTourDataSet
Every cyclist and stage of the Tour de France in two CSV files.
Stars: ✭ 61 (+335.71%)
Mutual labels:  data-analysis
Fraud-Detection-in-Online-Transactions
Detecting Frauds in Online Transactions using Anamoly Detection Techniques Such as Over Sampling and Under-Sampling as the ratio of Frauds is less than 0.00005 thus, simply applying Classification Algorithm may result in Overfitting
Stars: ✭ 41 (+192.86%)
Mutual labels:  data-analysis
online-course-recommendation-system
Built on data from Pluralsight's course API fetched results. Works with model trained with K-means unsupervised clustering algorithm.
Stars: ✭ 31 (+121.43%)
Mutual labels:  data-analysis

Metrics

This repo aims to gather a diverse group of organisations, institutions and individuals to explore how best to measure, classify and otherwise infer infromation from software ecosystems and projects.

Goals

  • To define a set of useful direct and derivative measures for judging aspects of a project within a number of thematic areas (i.e. code quality, community, documentation).
  • To signpost toward sources of data.
  • To define a set of metrics that are missing from Libraries.io that are not provided by another service.

Process

Related work

A place to refrence the significant works of others tackling these issues in industry, academia or as individuals.

Questions

For me this process begins with a number of framing questions. These questions are user-centred and based on the needs of our users, as defined in our personas from there we can define measures and metrics.

Measures

Measures can be direct (a quoted metric i.e. 'released on 1st Jan 2017), derivative (information gleaned from data i.e. 'released more than a year ago') or aggregated (compiled from >1 data i.e. 'all releases we're within the last year').

Measures may be broken down into a number of areas. At Libraries.io we have been considering areas for code, community, distribution and documentation. The single, overarching measure in Libraries.io is SourceRank which is defined over in our documentation.

Data

Data can be 'harvested' (gathered automatically using APIs, data dumps, scraping etc) or 'farmed' (gathered from a community by contribution).

Libraries.io currently focusses on harvesting metrics. It is preferable for a metric to be present in many sources rather than a single source so that we can make like-for-like comparisons across supported package managers.

Sources

Sources contain metrics. They may also contain measures themselves. We think it is important not to rely on proprietary, third party services for measures.

Reproducibility is the key issue here. An inability to reproduce a measure from the source data (metrics) risks the ability to create a like for like comparison of two pieces of software and ties all users of the classifier to that service. This is unacceptable(πŸ”Š).

Contributing

While there will be no one single approach that is right for everyone our hope is that we can come to consensus about what good measures look like and what metrics they will require, so that Libraries.io can provide them.

Please feel free to fork and PR additions to any of the documents in this repo, to share your thoughts and ideas in an issue, to propose a draft specifications for areas, measures etc. as a PR or reference.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].