A large language model (LLM) is a machine learning model designed for natural language processing tasks, especially language generation. LLMs are language...
113 KB (11,798 words) - 13:02, 5 June 2025
A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language...
64 KB (3,361 words) - 16:05, 24 May 2025
Llama (Large Language Model Meta AI, formerly stylized as LLaMA) is a family of large language models (LLMs) released by Meta AI starting in February 2023...
53 KB (4,940 words) - 07:11, 13 May 2025
A language model is a model of the human brain's ability to produce natural language. Language models are useful for a variety of tasks, including speech...
16 KB (2,383 words) - 06:50, 4 June 2025
Claude is a family of large language models developed by Anthropic. The first model was released in March 2023. The Claude 3 family, released in March...
26 KB (2,263 words) - 02:33, 6 June 2025
Gemini is a family of multimodal large language models (LLMs) developed by Google DeepMind, and the successor to LaMDA and PaLM 2. Comprising Gemini Ultra...
54 KB (4,386 words) - 16:08, 29 May 2025
Generative AI applications like large language models (LLM) are common examples of foundation models. Building foundation models is often highly resource-intensive...
44 KB (4,719 words) - 15:41, 30 May 2025
Chinchilla is a family of large language models (LLMs) developed by the research team at Google DeepMind, presented in March 2022. It is named "chinchilla"...
8 KB (615 words) - 19:51, 6 December 2024
improved the state-of-the-art for large language models. As of 2020[update], BERT is a ubiquitous baseline in natural language processing (NLP) experiments...
31 KB (3,568 words) - 19:15, 25 May 2025
Reasoning language models (RLMs) are large language models that have been further trained to solve multi-step reasoning tasks. These models perform better...
24 KB (2,859 words) - 11:13, 4 June 2025
Generative pre-trained transformer (redirect from GPT (language model))
A generative pre-trained transformer (GPT) is a type of large language model (LLM) and a prominent framework for generative artificial intelligence. It...
65 KB (5,278 words) - 15:49, 30 May 2025
Large language models have been used by officials and politicians in a wide variety of ways. The Conversation described ChatGPT described as a uniquely...
17 KB (1,396 words) - 07:17, 26 April 2025
Transformer) is a series of large language models developed by Google AI introduced in 2019. Like the original Transformer model, T5 models are encoder-decoder...
20 KB (1,932 words) - 03:55, 7 May 2025
processing including language and text generation. Unlike large language models (LLMs), small language models are much smaller in scale and scope. Typically, an...
2 KB (211 words) - 03:40, 29 April 2025
Large Open-science Open-access Multilingual Language Model (BLOOM) is a 176-billion-parameter transformer-based autoregressive large language model (LLM)...
4 KB (506 words) - 02:26, 19 April 2025
Language model benchmarks are standardized tests designed to evaluate the performance of language models on various natural language processing tasks....
91 KB (9,930 words) - 20:44, 25 May 2025
A 1.58-bit Large Language Model (1.58-bit LLM, also ternary LLM) is a version of a transformer large language model with weights using only three values:...
5 KB (618 words) - 19:44, 29 May 2025
Transformer (deep learning architecture) (redirect from Transformer model)
Later variations have been widely adopted for training large language models (LLM) on large (language) datasets. The modern version of the transformer was...
106 KB (13,108 words) - 21:15, 5 June 2025
Mistral AI (section Mistral Large 2)
headquartered in Paris. Founded in 2023, it specializes in open-weight large language models (LLMs). The company is named after the mistral, a powerful, cold...
28 KB (1,775 words) - 12:38, 31 May 2025
Prompt engineering (redirect from In-context learning (natural language processing))
intelligence (AI) model. A prompt is natural language text describing the task that an AI should perform. A prompt for a text-to-text language model can be a query...
40 KB (4,472 words) - 15:35, 2 June 2025
2024 to standardize the way artificial intelligence (AI) models like large language models (LLMs) integrate and share data with external tools, systems...
19 KB (1,679 words) - 14:02, 6 June 2025
DeepSeek (category Articles containing Chinese-language text)
DeepSeek, is a Chinese artificial intelligence company that develops large language models (LLMs). Based in Hangzhou, Zhejiang. It is owned and funded by the...
63 KB (6,059 words) - 07:52, 6 June 2025
Stochastic parrot (redirect from On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?)
describe the claim that large language models, though able to generate plausible language, do not understand the meaning of the language they process. The term...
22 KB (2,364 words) - 22:11, 31 May 2025
PaLM (redirect from Pathways Language Model)
PaLM (Pathways Language Model) is a 540 billion-parameter dense decoder-only transformer-based large language model (LLM) developed by Google AI. Researchers...
13 KB (807 words) - 13:21, 13 April 2025
and distributed systems. A large number of modeling languages appear in the literature. Example of graphical modeling languages in the field of computer...
23 KB (2,902 words) - 14:18, 4 April 2025
simply Perplexity, is an American web search engine that uses a large language model to process queries and synthesize responses based on web search results...
28 KB (2,223 words) - 18:13, 5 June 2025
Meta AI (redirect from No Language Left Behind)
Overleaf. In February 2023, Meta AI launched LLaMA (Large Language Model Meta AI), a large language model ranging from 7B to 65B parameters. On April 5, 2025...
22 KB (1,923 words) - 08:37, 31 May 2025
Reflection (artificial intelligence) (redirect from Large reasoning model)
Reflection is the term used for how some large language models (specifically reasoning language models (RLMs)) share information among their input or...
9 KB (772 words) - 19:27, 25 May 2025
a vision-language model (VLM) by training it on robot trajectory data and large-scale visual language data or Internet-scale vision-language tasks. Examples...
3 KB (184 words) - 02:50, 21 May 2025
GPT-3 (redirect from GPT-3 (language model))
Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020. Like its predecessor, GPT-2, it is a decoder-only transformer model of deep neural network...
55 KB (4,923 words) - 20:03, 12 May 2025