neural network
A neural network is a computational model composed of layered, interconnected units (neurons) that learn input-to-output mappings by adjusting weighted connections and applying nonlinear activation functions.
Training typically consists of a forward pass that produces predictions and a loss/objective, followed by gradient-based optimization: gradients are computed via backpropagation (automatic differentiation using the chain rule) and used to update parameters.
Architectures vary by data and task, including the following:
- Feedforward multilayer perceptrons
- Convolutional networks for images and other grid-structured data
- Recurrent networks for sequences
- Transformer-based models that use self-attention to capture long-range dependencies
- Graph neural networks for graph-structured data
Despite the biological inspiration, modern neural networks are mathematically defined, differentiable function approximators whose parameters (weights and biases) are learned from data by minimizing a loss under supervised, self-supervised, unsupervised, or reinforcement learning setups.
By Leodanis Pozo Ramos • Updated Oct. 13, 2025