• of artificial neural networks (ANNs), the neural tangent kernel (NTK) is a kernel that describes the evolution of deep artificial neural networks during...
    35 KB (5,146 words) - 10:08, 16 April 2025
  • Fisher kernel Graph kernels Kernel smoother Polynomial kernel Radial basis function kernel (RBF) String kernels Neural tangent kernel Neural network...
    13 KB (1,670 words) - 17:02, 3 August 2025
  • artificial neural networks after random initialization of their parameters, but before training; it appears as a term in neural tangent kernel prediction...
    20 KB (2,964 words) - 01:28, 19 April 2024
  • Finland. (Known as NTK Nakkila) Neural tangent kernel, a mathematical tool to describe the training of artificial neural networks NTK, a Niterra brand of...
    616 bytes (108 words) - 21:37, 30 August 2024
  • A convolutional neural network (CNN) is a type of feedforward neural network that learns features via filter (or kernel) optimization. This type of deep...
    138 KB (15,555 words) - 03:37, 31 July 2025
  • architecture and initializations hyper-parameters. The Neural Tangent Kernel describes the evolution of neural network predictions during gradient descent training...
    9 KB (869 words) - 11:20, 5 February 2024
  • Thumbnail for Neural network (machine learning)
    Clement Hongler (2018). Neural Tangent Kernel: Convergence and Generalization in Neural Networks (PDF). 32nd Conference on Neural Information Processing...
    168 KB (17,613 words) - 12:10, 26 July 2025
  • (Not to be confused with the lazy learning regime, see Neural tangent kernel). In machine learning, lazy learning is a learning method in which generalization...
    9 KB (1,102 words) - 15:40, 28 May 2025
  • Thumbnail for Feedforward neural network
    Feedforward refers to recognition-inference architecture of neural networks. Artificial neural network architectures are based on inputs multiplied by weights...
    21 KB (2,242 words) - 18:37, 19 July 2025
  • In artificial neural networks, recurrent neural networks (RNNs) are designed for processing sequential data, such as text, speech, and time series, where...
    90 KB (10,414 words) - 07:48, 4 August 2025
  • ')=e^{-\alpha \psi (\theta ,\theta ')},\alpha >0.} The sigmoid kernel, or hyperbolic tangent kernel, is defined as K ( x , y ) = tanh ⁡ ( γ x T y + r ) , x ...
    24 KB (4,346 words) - 02:00, 27 May 2025
  • Sohl-Dickstein, Jascha; Schoenholz, Samuel S. (2020). "Neural Tangents: Fast and Easy Infinite Neural Networks in Python". International Conference on Learning...
    44 KB (5,935 words) - 23:18, 5 August 2025
  • Anouar, F. (2000). "Generalized Discriminant Analysis Using a Kernel Approach". Neural Computation. 12 (10): 2385–2404. CiteSeerX 10.1.1.412.760. doi:10...
    21 KB (2,248 words) - 07:14, 18 April 2025
  • Support vector machine (category Kernel methods for machine learning)
    using the kernel trick, representing the data only through a set of pairwise similarity comparisons between the original data points using a kernel function...
    65 KB (9,071 words) - 17:00, 3 August 2025
  • Weight initialization (category Artificial neural networks)
    initialized. Similarly, trainable parameters in convolutional neural networks (CNNs) are called kernels and biases, and this article also describes these. We...
    25 KB (2,919 words) - 23:16, 20 June 2025
  • Thumbnail for Rectifier (neural networks)
    In the context of artificial neural networks, the rectifier or ReLU (rectified linear unit) activation function is an activation function defined as the...
    23 KB (3,056 words) - 00:05, 21 July 2025
  • model Kernel adaptive filter Kernel density estimation Kernel eigenvoice Kernel embedding of distributions Kernel method Kernel perceptron Kernel random...
    39 KB (3,385 words) - 07:36, 7 July 2025
  • Thumbnail for Nonlinear dimensionality reduction
    Müller, K.-R. (1998). "Nonlinear Component Analysis as a Kernel Eigenvalue Problem". Neural Computation. 10 (5). MIT Press: 1299–1319. doi:10.1162/089976698300017467...
    48 KB (6,119 words) - 04:01, 2 June 2025
  • Gated recurrent units (GRUs) are a gating mechanism in recurrent neural networks, introduced in 2014 by Kyunghyun Cho et al. The GRU is like a long short-term...
    9 KB (1,290 words) - 21:05, 2 August 2025
  • Multilayer perceptron (category Neural network architectures)
    learning, a multilayer perceptron (MLP) is a name for a modern feedforward neural network consisting of fully connected neurons with nonlinear activation...
    16 KB (1,932 words) - 03:01, 30 June 2025
  • Thumbnail for Long short-term memory
    Long short-term memory (category Neural network architectures)
    Long short-term memory (LSTM) is a type of recurrent neural network (RNN) aimed at mitigating the vanishing gradient problem commonly encountered by traditional...
    52 KB (5,822 words) - 21:03, 2 August 2025
  • Thumbnail for Activation function
    Activation function (category Artificial neural networks)
    kernels of the previous neural network layer while i {\displaystyle i} iterates through the number of kernels of the current layer. In quantum neural...
    25 KB (1,963 words) - 00:07, 21 July 2025
  • implemented as a neural network, neural ODE methods would be needed. Indeed, CNF was first proposed in the same paper that proposed neural ODE. There are...
    56 KB (9,669 words) - 02:34, 5 August 2025
  • Thumbnail for Loss functions for classification
    The Tangent loss is quasi-convex and is bounded for large negative values which makes it less sensitive to outliers. Interestingly, the Tangent loss...
    24 KB (4,212 words) - 23:53, 20 July 2025
  • Vanishing gradient problem (category Artificial neural networks)
    and later layers encountered when training neural networks with backpropagation. In such methods, neural network weights are updated proportional to...
    24 KB (3,711 words) - 14:28, 9 July 2025
  • Thumbnail for LeNet
    LeNet (category Artificial neural networks)
    1988, LeCun et al. published a neural network design that recognize handwritten zip code. However, its convolutional kernels were hand-designed. In 1989...
    31 KB (3,946 words) - 16:34, 3 August 2025
  • Wasserstein GAN (category Neural network architectures)
    1 {\displaystyle \sup _{x}|h'(x)|\leq 1} . For example, the hyperbolic tangent function h = tanh {\displaystyle h=\tanh } satisfies the requirement. Then...
    16 KB (2,884 words) - 07:23, 26 January 2025
  • Thumbnail for Isometry
    v , w {\displaystyle v,w} on M {\displaystyle M} (i.e. sections of the tangent bundle T M {\displaystyle \mathrm {T} M} ), g ( v , w ) = g ′ ( f ∗ v ...
    18 KB (2,425 words) - 02:45, 30 July 2025
  • subalgebra of kernels which can be solved in O ( n ) {\displaystyle O(n)} . neural-tangents is a specialized package for infinitely wide neural networks....
    28 KB (1,681 words) - 20:44, 23 May 2025
  • dynamic reposing and tangent distance for drug activity prediction Archived 7 December 2019 at the Wayback Machine." Advances in Neural Information Processing...
    266 KB (15,010 words) - 06:44, 12 July 2025