Finding Nearly Everything within Random Binary Networks

Kartik Sreenivasan, Shashank Rajput, Jy Yong Sohn, Dimitris Papailiopoulos

Research output: Contribution to journalConference articlepeer-review

8 Citations (Scopus)

Abstract

A recent work by Ramanujan et al. (2020) provides significant empirical evidence that sufficiently overparameterized, random neural networks contain untrained subnetworks that achieve state-of-the-art accuracy on several predictive tasks. A follow-up line of theoretical work provides justification of these findings by proving that slightly overparameterized neural networks, with commonly used continuous-valued random initializations can indeed be pruned to approximate any target network. In this work, we show that the amplitude of those random weights does not even matter. We prove that any target network of width d and depth l can be approximated up to arbitrary accuracy ε by simply pruning a random network of binary {±1} weights that is wider and deeper than the target network only by a polylogarithmic factor of d, l and ε.

Original languageEnglish
Pages (from-to)3531-3541
Number of pages11
JournalProceedings of Machine Learning Research
Volume151
Publication statusPublished - 2022
Event25th International Conference on Artificial Intelligence and Statistics, AISTATS 2022 - Virtual, Online, Spain
Duration: 2022 Mar 282022 Mar 30

Bibliographical note

Publisher Copyright:
Copyright © 2022 by the author(s)

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Finding Nearly Everything within Random Binary Networks'. Together they form a unique fingerprint.

Cite this