• parameters. Next, the actual task is performed with supervised or unsupervised learning. Self-supervised learning has produced promising results in recent years...
    18 KB (2,047 words) - 16:20, 4 April 2025
  • Weak supervision (also known as semi-supervised learning) is a paradigm in machine learning, the relevance and notability of which increased with the advent...
    22 KB (3,038 words) - 10:40, 31 December 2024
  • perform a specific task. Feature learning can be either supervised or unsupervised. In supervised feature learning, features are learned using labelled...
    140 KB (15,513 words) - 11:41, 29 April 2025
  • Thumbnail for Feature learning
    explicit algorithms. Feature learning can be either supervised, unsupervised, or self-supervised: In supervised feature learning, features are learned using...
    45 KB (5,114 words) - 14:51, 30 April 2025
  • Unsupervised learning is a framework in machine learning where, in contrast to supervised learning, algorithms learn patterns exclusively from unlabeled...
    31 KB (2,770 words) - 08:47, 30 April 2025
  • Artificial neural networks (ANNs) are models created using machine learning to perform a number of tasks. Their creation was inspired by biological neural...
    84 KB (8,626 words) - 11:12, 27 April 2025
  • Thumbnail for Transformer (deep learning architecture)
    requiring learning rate warmup. Transformers typically are first pretrained by self-supervised learning on a large generic dataset, followed by supervised fine-tuning...
    106 KB (13,091 words) - 21:14, 29 April 2025
  • Large language model (category Deep learning)
    are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. The largest and most capable LLMs are...
    114 KB (11,942 words) - 05:35, 30 April 2025
  • Thumbnail for GPT-1
    models primarily employed supervised learning from large amounts of manually labeled data. This reliance on supervised learning limited their use of datasets...
    32 KB (1,064 words) - 01:34, 21 March 2025
  • are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. This page lists notable large language...
    64 KB (3,361 words) - 09:20, 29 April 2025
  • Thumbnail for Attention (machine learning)
    Attention is a machine learning method that determines the relative importance of each component in a sequence relative to the other components in that...
    36 KB (3,494 words) - 17:00, 1 May 2025
  • was trained using a combination of first supervised learning on a large dataset, then reinforcement learning using both human and AI feedback, it did...
    64 KB (6,206 words) - 19:48, 1 May 2025
  • Thumbnail for Reinforcement learning
    Reinforcement learning is one of the three basic machine learning paradigms, alongside supervised learning and unsupervised learning. Reinforcement learning differs...
    64 KB (7,580 words) - 08:49, 30 April 2025
  • Thumbnail for Transfer learning
    next driver of machine learning commercial success after supervised learning. In the 2020 paper, "Rethinking Pre-Training and self-training", Zoph et al...
    15 KB (1,637 words) - 03:42, 29 April 2025
  • Mamba is a deep learning architecture focused on sequence modeling. It was developed by researchers from Carnegie Mellon University and Princeton University...
    11 KB (1,159 words) - 19:42, 16 April 2025
  • 2024. "Curriculum learning with diversity for supervised computer vision tasks". Retrieved March 29, 2024. "Self-paced Curriculum Learning". Retrieved March...
    13 KB (1,367 words) - 02:58, 30 January 2025
  • Google. It learns to represent text as a sequence of vectors using self-supervised learning. It uses the encoder-only transformer architecture. BERT dramatically...
    31 KB (3,528 words) - 01:20, 29 April 2025
  • feedback, learning a reward model, and optimizing the policy. Compared to data collection for techniques like unsupervised or self-supervised learning, collecting...
    62 KB (8,615 words) - 05:24, 30 April 2025
  • Thumbnail for Generative pre-trained transformer
    models commonly employed supervised learning from large amounts of manually-labeled data. The reliance on supervised learning limited their use on datasets...
    65 KB (5,342 words) - 13:55, 1 May 2025
  • Thumbnail for Word embedding
    (and multi-lingual) corpora, also providing an early example of self-supervised learning of word embeddings. Word embeddings come in two different styles...
    29 KB (3,154 words) - 07:58, 30 March 2025
  • International Conference on Machine Learning (ICML) is a leading international academic conference in machine learning. Along with NeurIPS and ICLR, it is...
    5 KB (377 words) - 16:54, 19 March 2025
  • much more flexible structure to exist among those alternatives. Supervised learning algorithms search through a hypothesis space to find a suitable hypothesis...
    54 KB (6,794 words) - 06:02, 19 April 2025
  • of outputs via an artificial neural network. Deep learning methods, often using supervised learning with labeled datasets, have been shown to solve tasks...
    27 KB (2,929 words) - 10:33, 13 March 2025
  • The International Conference on Learning Representations (ICLR) is a machine learning conference typically held in late April or early May each year....
    4 KB (272 words) - 11:18, 10 July 2024
  • classification and regression algorithms. Hence, it is prevalent in supervised learning for converting weak learners to strong learners. The concept of boosting...
    21 KB (2,240 words) - 13:33, 27 February 2025
  • Thumbnail for Self-organizing map
    A self-organizing map (SOM) or self-organizing feature map (SOFM) is an unsupervised machine learning technique used to produce a low-dimensional (typically...
    34 KB (4,063 words) - 21:25, 10 April 2025
  • activation map use the same set of parameters that define the filter. Self-supervised learning has been adapted for use in convolutional layers by using sparse...
    138 KB (15,599 words) - 06:42, 18 April 2025
  • Learning to rank or machine-learned ranking (MLR) is the application of machine learning, typically supervised, semi-supervised or reinforcement learning...
    54 KB (4,442 words) - 00:21, 17 April 2025
  • radial basis networks, another class of supervised neural network models). In recent developments of deep learning the rectified linear unit (ReLU) is more...
    16 KB (1,932 words) - 07:03, 29 December 2024
  • are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. The largest and most capable LLMs are...
    16 KB (2,382 words) - 00:06, 17 April 2025