We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cond-mat.dis-nn

Change to browse by:

References & Citations

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Condensed Matter > Disordered Systems and Neural Networks

Title: Entropic alternatives to initialization

Authors: Daniele Musso
Abstract: Local entropic loss functions provide a versatile framework to define architecture-aware regularization procedures. Besides the possibility of being anisotropic in the synaptic space, the local entropic smoothening of the loss function can vary during training, thus yielding a tunable model complexity. A scoping protocol where the regularization is strong in the early-stage of the training and then fades progressively away constitutes an alternative to standard initialization procedures for deep convolutional neural networks, nonetheless, it has wider applicability. We analyze anisotropic, local entropic smoothenings in the language of statistical physics and information theory, providing insight into both their interpretation and workings. We comment some aspects related to the physics of renormalization and the spacetime structure of convolutional networks.
Comments: 19 pages, 5 figures, 2 appendices; v2 added comments and references
Subjects: Disordered Systems and Neural Networks (cond-mat.dis-nn); Statistical Mechanics (cond-mat.stat-mech); Machine Learning (cs.LG); High Energy Physics - Theory (hep-th)
Cite as: arXiv:2107.07757 [cond-mat.dis-nn]
  (or arXiv:2107.07757v2 [cond-mat.dis-nn] for this version)

Submission history

From: Daniele Musso [view email]
[v1] Fri, 16 Jul 2021 08:17:32 GMT (1170kb,D)
[v2] Wed, 28 Jul 2021 06:21:32 GMT (1169kb,D)

Link back to: arXiv, form interface, contact.