Neural Network
A neural network is a parametric model. It is used with weights that are fit by loss minimization.
Introduction
A visualization of a neural network is:
There is a vector of inputs (x) with dx elements, which forms the first layer, and a vector of outputs (y) with dy elements, forming the last layer. In between there are hidden layers.
Layers are composed of nodes called neurons. Each member of the input and output vectors are nodes; the number of nodes in a hidden layer is a parameter.
In every layer (except the last), each node is connected by an edge to every node in the subsequent layer. Each edge has an unknown weight. The vector of weights (w) has dw elements; the number is calculated as...
The dimensions of every layer are indexed from 0 to h.
Let d0 = dx and dh = dy.
dw = dhdh-1 + dh-1dh-2 + ... + d1d0.
Description
A neural network is a continuous mapping as
The neuromanifold (i.e., function space that can be parameterized) of ϕ is notated as ℳϕ and is defined by
Properties
ϕ is piecewise smooth so ℳϕ is a manifold with singularities.
so dim ℳϕ < dw.
Linear neural network
A linear neural network does not feature any activation functions. It can be characterized by the following:
Let r = min{d0, d1, ..., dh}.
Let M be the matrix product of all weights WhWh-1...W2W1.
ϕ(w,x) = Mx.
Properties
If r = min{d0,dh} then .
This can be interpretted as the neuromanifold not being bounded in the vector space; it has a filling architecture. On the other hand, if r < min{d0,dh}, then the neuromanifold does not form a vector space; it has a non-filling architecture and is a determinantal variety.
