Author:
Liu Yajing,Caglar Turgay,Peterson Christopher,Kirby Michael
Abstract
This paper investigates the integration of multiple geometries present within a ReLU-based neural network. A ReLU neural network determines a piecewise affine linear continuous map, M, from an input space ℝm to an output space ℝn. The piecewise behavior corresponds to a polyhedral decomposition of ℝm. Each polyhedron in the decomposition can be labeled with a binary vector (whose length equals the number of ReLU nodes in the network) and with an affine linear function (which agrees with M when restricted to points in the polyhedron). We develop a toolbox that calculates the binary vector for a polyhedra containing a given data point with respect to a given ReLU FFNN. We utilize this binary vector to derive bounding facets for the corresponding polyhedron, extraction of “active” bits within the binary vector, enumeration of neighboring binary vectors, and visualization of the polyhedral decomposition (Python code is available at https://github.com/cglrtrgy/GoL_Toolbox). Polyhedra in the polyhedral decomposition of ℝm are neighbors if they share a facet. Binary vectors for neighboring polyhedra differ in exactly 1 bit. Using the toolbox, we analyze the Hamming distance between the binary vectors for polyhedra containing points from adversarial/nonadversarial datasets revealing distinct geometric properties. A bisection method is employed to identify sample points with a Hamming distance of 1 along the shortest Euclidean distance path, facilitating the analysis of local geometric interplay between Euclidean geometry and the polyhedral decomposition along the path. Additionally, we study the distribution of Chebyshev centers and related radii across different polyhedra, shedding light on the polyhedral shape, size, clustering, and aiding in the understanding of decision boundaries.
Subject
Artificial Intelligence,Information Systems,Computer Science (miscellaneous)
Reference25 articles.
1. “Understanding deep neural networks with rectified linear units,”;Arora;International Conference on Learning Representations,2018
2. “A spline theory of deep learning,”;Balestriero;Proceedings of the 35th International Conference on Machine Learning,2018
3. Convex Optimization
4. Pconvex geometry and duality of over-parameterized neural networks;Ergen;J. Mach. Learn. Res,2021
5. Explaining and harnessing adversarial examples;Goodfellow;arXiv,2015