intelligence (AI) design, AI capability control proposals, also referred to as AI confinement, aim to increase our ability to monitor and control the behavior...
25 KB (3,182 words) - 21:22, 14 February 2025
subfield of AI safety, the study of how to build safe AI systems. Other subfields of AI safety include robustness, monitoring, and capability control. Research...
132 KB (12,973 words) - 16:13, 26 April 2025
to align AI goal systems with human values, and capability control, which aims to reduce an AI system's capacity to harm humans or gain control. An example...
39 KB (4,241 words) - 18:29, 28 April 2025
Human Compatible (redirect from Human compatible AI and the problem of control)
progress in AI capability is inevitable because of economic pressures. Such pressures can already be seen in the development of existing AI technologies...
12 KB (1,133 words) - 04:12, 3 April 2025
an existential catastrophe, it is necessary to successfully solve the "AI control problem" for the first superintelligence. The solution might involve instilling...
13 KB (1,273 words) - 06:15, 3 April 2025
Claude (language model) (redirect from Claude.ai)
@AnthropicAI (October 22, 2024). "Introducing an upgraded Claude 3.5 Sonnet, and a new model, Claude 3.5 Haiku. We're also introducing a new capability in beta:...
21 KB (1,894 words) - 20:08, 19 April 2025
Existential risk from artificial intelligence (redirect from Existential risk of AI)
greater chance that human inability to control AI will cause an existential catastrophe. In 2023, hundreds of AI experts and other notable figures signed...
127 KB (13,292 words) - 18:26, 28 April 2025
Artificial intelligence (redirect from AI)
Artificial intelligence (AI) refers to the capability of computational systems to perform tasks typically associated with human intelligence, such as learning...
278 KB (28,570 words) - 12:40, 19 April 2025
from artificial general intelligence. MIRI's work has focused on a friendly AI approach to system design and on predicting the rate of technology development...
16 KB (1,149 words) - 04:07, 16 February 2025
on – AI existential risk. Effective altruists (particularly longtermists) argue that AI companies should be cautious and strive to develop safe AI systems...
23 KB (1,988 words) - 14:43, 27 April 2025
subfield of AI safety, the study of how to build safe AI systems. Other subfields of AI safety include robustness, monitoring, and capability control. Research...
87 KB (10,322 words) - 20:49, 28 April 2025
Music and artificial intelligence (redirect from AI-generated music)
feature is the capability of an AI algorithm to learn based on past data, such as in computer accompaniment technology, wherein the AI is capable of listening...
61 KB (6,981 words) - 10:36, 26 April 2025
against the hypothetical risk of an AI takeover, and have advocated for the use of AI capability control in addition to AI alignment methods. Other fields...
108 KB (10,440 words) - 00:38, 1 May 2025
first DGX-1 supercomputer to OpenAI in August 2016 to help it train larger and more complex AI models with the capability of reducing processing time from...
219 KB (19,127 words) - 19:00, 30 April 2025
The object-capability model is a computer security model. A capability describes a transferable right to perform one (or more) operations on a given object...
8 KB (1,026 words) - 05:43, 10 October 2024
refines their decision-making capability over time. While Deep learning, as opposed to rule-based methods, supports Agentic AI through multi-layered neural...
13 KB (1,373 words) - 10:08, 1 May 2025
Roman Yampolskiy (category AI safety scientists)
ISBN 978-1482234435 AI: Unexplainable, Unpredictable, Uncontrollable. Chapman & Hall/CRC Press, 2024, ISBN 978-1032576268 AI capability control AI-complete Machine...
11 KB (860 words) - 09:35, 9 February 2025
Artificial general intelligence (redirect from Hard AI)
Artificial general intelligence (AGI)—sometimes called human‑level intelligence AI—is a type of artificial intelligence capable of performing the full spectrum...
131 KB (14,472 words) - 20:27, 29 April 2025
former OpenAI researcher Paul Christiano, ARC focuses on recognizing and comprehending the potentially harmful capabilities of present-day AI models. ARC's...
7 KB (601 words) - 14:38, 25 February 2025
Mistral AI SAS is a French artificial intelligence (AI) startup, headquartered in Paris. It specializes in open-weight large language models (LLMs). The...
27 KB (1,716 words) - 03:22, 29 April 2025
Artificial Intelligence Act (redirect from General-purpose AI)
general-purpose AI, transparency requirements are imposed, with reduced requirements for open source models, and additional evaluations for high-capability models...
32 KB (3,189 words) - 06:11, 12 April 2025
Regulation of artificial intelligence (redirect from Regulation of AI)
numerous AI ethics guidelines have been published in order to maintain social control over the technology. Regulation is deemed necessary to both foster AI innovation...
134 KB (13,317 words) - 11:33, 30 April 2025
entities (including Shield AI) on its export control list, barring the export of dual-use commodities to that business. Shield AI employs machine learning...
13 KB (1,284 words) - 11:16, 6 April 2025
AI systems, as well as "tracking highly capable AI systems and large pools of computational capability" and "robust public funding for technical AI safety...
13 KB (1,412 words) - 19:48, 16 April 2025
against and showing the capability to defeat professional teams. By choosing a game as complex as Dota 2 to study machine learning, OpenAI thought they could...
23 KB (2,260 words) - 09:10, 6 April 2025
History of artificial intelligence (redirect from History of AI)
as little as 15 seconds of audio to reproduce a voice—a capability later corroborated by OpenAI in 2024. The service went viral on social media platforms...
166 KB (19,442 words) - 15:28, 29 April 2025
Superintelligence (redirect from Superhuman AI)
proposed various approaches to mitigate risks associated with ASI: Capability control – Limiting an ASI's ability to influence the world, such as through...
43 KB (4,588 words) - 22:19, 27 April 2025
Anthropic (section Constitutional AI)
intelligence (AI) startup company founded in 2021. Anthropic has developed a family of large language models (LLMs) named Claude as a competitor to OpenAI's ChatGPT...
31 KB (2,841 words) - 09:41, 26 April 2025
Neural processing unit (redirect from AI accelerator)
in-memory computing capability. As of 2024[update], a typical AI integrated circuit chip contains tens of billions of MOSFETs. AI accelerators are used...
51 KB (4,926 words) - 10:31, 10 April 2025