Neural Network
Short Definition
Full Definition
Neural networks are computational systems modeled after the biological neural networks found in animal brains. They consist of interconnected groups of artificial neurons organized in layers: an input layer, one or more hidden layers, and an output layer. Each connection between neurons carries a weight that is adjusted during the learning process. The concept originated in the 1940s with McCulloch and Pitts, but practical applications only became feasible with increased computing power and the development of the backpropagation algorithm in the 1980s. Modern neural networks range from simple feedforward architectures to complex deep learning systems with hundreds of layers. They excel at pattern recognition, classification, regression, and generative tasks. Neural networks learn by adjusting the weights of connections based on training data, gradually improving their ability to make accurate predictions. The universal approximation theorem shows that a neural network with even a single hidden layer can approximate any continuous function, given sufficient neurons. Today, neural networks power everything from voice assistants to autonomous vehicles, medical diagnostics to financial trading systems.
Technical Explanation
A neural network computes outputs through forward propagation: each neuron calculates a weighted sum of its inputs z = sum(w_i * x_i) + b, then applies a nonlinear activation function f(z) such as ReLU, sigmoid, or tanh. During training, the network uses backpropagation to compute gradients of the loss function with respect to each weight using the chain rule. These gradients are then used by an optimization algorithm like stochastic gradient descent (SGD) or Adam to update the weights iteratively. The loss function measures the difference between predicted and actual outputs. Regularization techniques like dropout, L1/L2 penalties, and batch normalization help prevent overfitting. The architecture choice depends on the task: fully connected layers for tabular data, convolutional layers for spatial data, and recurrent layers for sequential data.
Use Cases
Advantages
Disadvantages
Schema Type
Featured Snippet Candidate
Difficulty Level