All Projects → bneyshabur → over-parametrization

bneyshabur / over-parametrization

Licence: other
Computing various norms/measures on over-parametrized neural networks

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to over-parametrization

paper-notes
ML/DL/RL paper notes
Stars: ✭ 21 (-54.35%)
Mutual labels:  generalization
transformer generalization
The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We significantly improve the systematic generalization of transformer models on a variety of datasets using simple tricks and careful considerations.
Stars: ✭ 58 (+26.09%)
Mutual labels:  generalization
pomdp-baselines
Simple (but often Strong) Baselines for POMDPs in PyTorch - ICML 2022
Stars: ✭ 162 (+252.17%)
Mutual labels:  generalization
Recurrent-Independent-Mechanisms
Implementation of the paper Recurrent Independent Mechanisms (https://arxiv.org/pdf/1909.10893.pdf)
Stars: ✭ 90 (+95.65%)
Mutual labels:  generalization
pgdl
Winning Solution of the NeurIPS 2020 Competition on Predicting Generalization in Deep Learning
Stars: ✭ 36 (-21.74%)
Mutual labels:  generalization
Transferlearning
Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习
Stars: ✭ 8,481 (+18336.96%)
Mutual labels:  generalization
adapt
Awesome Domain Adaptation Python Toolbox
Stars: ✭ 46 (+0%)
Mutual labels:  generalization
limit-label-memorization
Improving generalization by controlling label-noise information in neural network weights.
Stars: ✭ 34 (-26.09%)
Mutual labels:  generalization
awesome-nn-optimization
Awesome list for Neural Network Optimization methods.
Stars: ✭ 39 (-15.22%)
Mutual labels:  generalization
DiagnoseRE
Source code and dataset for the CCKS201 paper "On Robustness and Bias Analysis of BERT-based Relation Extraction"
Stars: ✭ 23 (-50%)
Mutual labels:  generalization
3DShapeGen
Code for 3D Reconstruction of Novel Object Shapes from Single Images paper
Stars: ✭ 92 (+100%)
Mutual labels:  generalization
SpinNet
[CVPR 2021] SpinNet: Learning a General Surface Descriptor for 3D Point Cloud Registration
Stars: ✭ 181 (+293.48%)
Mutual labels:  generalization
modules
The official repository for our paper "Are Neural Nets Modular? Inspecting Functional Modularity Through Differentiable Weight Masks". We develop a method for analyzing emerging functional modularity in neural networks based on differentiable weight masks and use it to point out important issues in current-day neural networks.
Stars: ✭ 25 (-45.65%)
Mutual labels:  generalization
extrapolate
generalize counter-examples of property-based testing
Stars: ✭ 13 (-71.74%)
Mutual labels:  generalization
Generalizing-Lottery-Tickets
This repository contains code to replicate the experiments given in NeurIPS 2019 paper "One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers"
Stars: ✭ 48 (+4.35%)
Mutual labels:  generalization

The role of over-parametrization in generalization of neural nets

This repository contains the code to train neural nets and compute various measures/norms reported in the following paper:

Towards Understanding the Role of Over-Parametrization in Generalization of Neural Networks

Behnam Neyshabur, Zhiyuan Li, Srinadh Bhojanapalli, Yann LeCun, Nathan Srebro

Usage

  1. Install Python 3.6 and PyTorch 0.4.1.
  2. Clone the repository:
    git clone https://github.com/bneyshabur/over-parametrization.git
    
  3. As a simple example, the following command trains a two layer fully connected feedforward network with 1000 hidden units on CIFAR10 dataset and then computes several measures/norms on the learned network:
    python main.py --dataset CIFAR10 --nunits 1000
    

Main Inputs Arguments

  • --no-cuda: disables cuda training
  • --datadir: path to the directory that contains the datasets (default: datasets)
  • --dataset: name of the dataset(options: MNIST | CIFAR10 | CIFAR100 | SVHN, default: CIFAR10). If the dataset is not in the desired directory, it will be downloaded.
  • --nunits: number of hidden units (default: 1024)

Reported Norms/Measures

After training the network, several norms/measures will be computed and reported on the trained network. Please see the file measures.py for explanation of each measure. We also compute and report the following generalization bounds:

  • VC bound: Generalization bound based on the VC dimension by Harvey et al. 2017
  • L1max bound: Generalization bound by Bartlett and Mendelson 2002
  • Fro bound: Generalization bound by Neyshabur et al. 2015
  • Spec_L1 bound: Generalization bound by Bartlett et al. 2017
  • Spec_Fro bound: Generalization bound by Neyshabur et al. 2018
  • Our bound: The Generalization bound proposed in this paper
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].