# wide-sparse-nets **Repository Path**: mirrors_google-research/wide-sparse-nets ## Basic Information - **Project Name**: wide-sparse-nets - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2021-03-19 - **Last Updated**: 2026-03-22 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Are wider nets better given the same number of parameters? This repository contains the code used for the experiments in the following paper: ["Are wider nets better given the same number of parameters?"](https://arxiv.org/abs/2010.14495)
Anna Golubeva, Behnam Neyshabur, Guy Gur-Ari.
International Conference on Learning Representations (ICLR), 2021. **Disclaimer**: this is not an official Google product. ## Getting Started Clone this repo, then install all dependencies: ``` pip install -r requirements.txt ``` The code was tested with Python 3.6.8. ## Code Organization Below is a description of the major sections of the code base. Run `python main.py --help` for a complete description of flags and hyperparameters. ### Datasets This code supports the following datasets: CIFAR-10, CIFAR-100, MNIST, SVHN. All datasets will download automatically. ### Models We consider two types of models: MLP and ResNet18. - `MLP`: MLP (i.e., fully-connected feed-forward achitecture) with 1 hidden layer for MNIST experiments - `ResNet18`: models with ResNet18 architecture from [this repo](https://github.com/kuangliu/pytorch-cifar/blob/master/models/resnet.py) for CIFAR-10, CIFAR-100 and SVHN Use the file `generate_arg.py` in the respective folder to set the experiment parameters. Calling ``` python generate_args.py ``` will print out commands to start the main script from the shell (locally). For ResNet18 experiments, it will also dump a dictionary specifying all job parameters into a json file, which is convenient to use if submitting jobs to a cluster or to the cloud. ## Citation If you use this code for your research, please cite our paper ["Are wider nets better given the same number of parameters?"](https://arxiv.org/abs/2010.14495).