• Thumbnail for Attention (machine learning)
    Attention is a machine learning method that determines the importance of each component in a sequence relative to the other components in that sequence...
    35 KB (3,424 words) - 23:19, 23 May 2025
  • Thumbnail for Transformer (deep learning architecture)
    transformer is a deep learning architecture that was developed by researchers at Google and is based on the multi-head attention mechanism, which was proposed...
    106 KB (13,105 words) - 11:32, 29 May 2025
  • Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of statistical algorithms that can learn...
    140 KB (15,570 words) - 14:43, 28 May 2025
  • Self-attention can mean: Attention (machine learning), a machine learning technique self-attention, an attribute of natural cognition This disambiguation...
    165 bytes (50 words) - 22:45, 18 February 2024
  • Thumbnail for Quantum machine learning
    Quantum machine learning is the integration of quantum algorithms within machine learning programs. The most common use of the term refers to machine learning...
    78 KB (9,362 words) - 16:46, 28 May 2025
  • Thumbnail for Attention Is All You Need
    "Attention Is All You Need" is a 2017 landmark research paper in machine learning authored by eight scientists working at Google. The paper introduced...
    15 KB (3,909 words) - 20:36, 1 May 2025
  • page is a timeline of machine learning. Major discoveries, achievements, milestones and other major events in machine learning are included. History of...
    33 KB (1,764 words) - 05:08, 20 May 2025
  • translation conference Workshop on Statistical Machine Translation. Gehring et al. combined a CNN encoder with an attention mechanism in 2017, which handled long-range...
    36 KB (3,901 words) - 17:39, 23 May 2025
  • neural networks used for attention (machine learning) This disambiguation page lists articles associated with the title Attention network. If an internal...
    308 bytes (71 words) - 08:42, 27 May 2021
  • with wide support for machine learning algorithms, written in C and Lua. Attention (machine learning) Convolution Deep learning Natural-language processing...
    138 KB (15,585 words) - 21:36, 2 June 2025
  • interests and dislikes Attention (machine learning), a machine learning technique that mimics natural attention in organisms Attention (advertising), the...
    4 KB (455 words) - 16:51, 25 July 2024
  • Thumbnail for Neural network (machine learning)
    In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a computational model inspired by the structure...
    169 KB (17,645 words) - 09:43, 1 June 2025
  • Mixture of experts (category Machine learning algorithms)
    Mixture of experts (MoE) is a machine learning technique where multiple expert networks (learners) are used to divide a problem space into homogeneous...
    42 KB (5,571 words) - 19:42, 31 May 2025
  • Thumbnail for Visual temporal attention
    semantically more substantial regions in space, visual temporal attention modules enable machine learning algorithms to emphasize more on critical video frames...
    6 KB (473 words) - 08:44, 8 June 2023
  • Large language model (category Deep learning)
    A large language model (LLM) is a machine learning model designed for natural language processing tasks, especially language generation. LLMs are language...
    113 KB (11,794 words) - 03:28, 2 June 2025
  • Thumbnail for Reinforcement learning
    Reinforcement learning is one of the three basic machine learning paradigms, alongside supervised learning and unsupervised learning. Reinforcement learning differs...
    69 KB (8,193 words) - 11:38, 2 June 2025
  • Thumbnail for Vision transformer
    exaFLOPs. Transformer (machine learning model) Convolutional neural network Attention (machine learning) Perceiver Deep learning PyTorch TensorFlow Dosovitskiy...
    37 KB (4,127 words) - 20:13, 29 April 2025
  • In machine learning, diffusion models, also known as diffusion-based generative models or score-based generative models, are a class of latent variable...
    84 KB (14,123 words) - 14:30, 1 June 2025
  • Thumbnail for Seq2seq
    Seq2seq is a family of machine learning approaches used for natural language processing. Applications include language translation, image captioning, conversational...
    23 KB (2,946 words) - 06:00, 19 May 2025
  • develops and studies methods and software that enable machines to perceive their environment and use learning and intelligence to take actions that maximize...
    279 KB (28,593 words) - 13:27, 31 May 2025
  • Multimodal learning is a type of deep learning that integrates and processes multiple types of data, referred to as modalities, such as text, audio, images...
    9 KB (2,212 words) - 22:40, 1 June 2025
  • (Cochrane Developmental, Psychosocial and Learning Problems Group) (August 2018). "Amphetamines for attention deficit hyperactivity disorder (ADHD) in...
    274 KB (29,044 words) - 20:03, 29 May 2025
  • Thumbnail for Deep learning
    Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression...
    180 KB (17,772 words) - 15:04, 30 May 2025
  • In machine learning, normalization is a statistical technique with various applications. There are two main forms of normalization, namely data normalization...
    34 KB (5,289 words) - 15:56, 26 May 2025
  • Thumbnail for Attention
    a direct result of the Learning by Observing and Pitching In model. Keen attention is both a requirement and result of learning by observing and pitching-in...
    105 KB (13,010 words) - 19:41, 23 May 2025
  • Thumbnail for Rote learning
    alternatives to rote learning include meaningful learning, associative learning, spaced repetition and active learning. Rote learning is widely used in the...
    10 KB (914 words) - 02:15, 12 September 2024
  • speech processing[citation needed]. Language modeling Transformer (machine learning model) State-space model Recurrent neural network The name comes from...
    11 KB (1,159 words) - 19:42, 16 April 2025
  • field of deep learning, most notably the development of the Transformer neural network, which he co-authored in landmark paper Attention Is All You Need...
    5 KB (383 words) - 06:54, 22 May 2025
  • AI (XAI), often overlapping with interpretable AI, or explainable machine learning (XML), is a field of research within artificial intelligence (AI) that...
    71 KB (7,829 words) - 04:32, 2 June 2025
  • Artificial neural networks (ANNs) are models created using machine learning to perform a number of tasks. Their creation was inspired by biological neural...
    85 KB (8,628 words) - 13:09, 27 May 2025