• intelligence (AI) design, AI capability control proposals, also referred to as AI confinement, aim to increase our ability to monitor and control the behavior...
    25 KB (3,182 words) - 21:22, 14 February 2025
  • subfield of AI safety, the study of how to build safe AI systems. Other subfields of AI safety include robustness, monitoring, and capability control. Research...
    132 KB (12,973 words) - 16:13, 26 April 2025
  • Thumbnail for AI takeover
    to align AI goal systems with human values, and capability control, which aims to reduce an AI system's capacity to harm humans or gain control. An example...
    39 KB (4,241 words) - 18:29, 28 April 2025
  • progress in AI capability is inevitable because of economic pressures. Such pressures can already be seen in the development of existing AI technologies...
    12 KB (1,133 words) - 04:12, 3 April 2025
  • an existential catastrophe, it is necessary to successfully solve the "AI control problem" for the first superintelligence. The solution might involve instilling...
    13 KB (1,273 words) - 06:15, 3 April 2025
  • Thumbnail for Claude (language model)
    @AnthropicAI (October 22, 2024). "Introducing an upgraded Claude 3.5 Sonnet, and a new model, Claude 3.5 Haiku. We're also introducing a new capability in beta:...
    21 KB (1,894 words) - 20:08, 19 April 2025
  • greater chance that human inability to control AI will cause an existential catastrophe. In 2023, hundreds of AI experts and other notable figures signed...
    127 KB (13,292 words) - 18:26, 28 April 2025
  • Artificial intelligence (redirect from AI)
    Artificial intelligence (AI) refers to the capability of computational systems to perform tasks typically associated with human intelligence, such as learning...
    278 KB (28,570 words) - 12:40, 19 April 2025
  • from artificial general intelligence. MIRI's work has focused on a friendly AI approach to system design and on predicting the rate of technology development...
    16 KB (1,149 words) - 04:07, 16 February 2025
  • Thumbnail for Effective accelerationism
    on – AI existential risk. Effective altruists (particularly longtermists) argue that AI companies should be cautious and strive to develop safe AI systems...
    23 KB (1,988 words) - 14:43, 27 April 2025
  • subfield of AI safety, the study of how to build safe AI systems. Other subfields of AI safety include robustness, monitoring, and capability control. Research...
    87 KB (10,322 words) - 20:49, 28 April 2025
  • feature is the capability of an AI algorithm to learn based on past data, such as in computer accompaniment technology, wherein the AI is capable of listening...
    61 KB (6,981 words) - 10:36, 26 April 2025
  • Thumbnail for Technology
    against the hypothetical risk of an AI takeover, and have advocated for the use of AI capability control in addition to AI alignment methods. Other fields...
    108 KB (10,440 words) - 00:38, 1 May 2025
  • first DGX-1 supercomputer to OpenAI in August 2016 to help it train larger and more complex AI models with the capability of reducing processing time from...
    219 KB (19,127 words) - 19:00, 30 April 2025
  • The object-capability model is a computer security model. A capability describes a transferable right to perform one (or more) operations on a given object...
    8 KB (1,026 words) - 05:43, 10 October 2024
  • refines their decision-making capability over time. While Deep learning, as opposed to rule-based methods, supports Agentic AI through multi-layered neural...
    13 KB (1,373 words) - 10:08, 1 May 2025
  • Thumbnail for Roman Yampolskiy
    Roman Yampolskiy (category AI safety scientists)
    ISBN 978-1482234435 AI: Unexplainable, Unpredictable, Uncontrollable. Chapman & Hall/CRC Press, 2024, ISBN 978-1032576268 AI capability control AI-complete Machine...
    11 KB (860 words) - 09:35, 9 February 2025
  • Artificial general intelligence (AGI)—sometimes called human‑level intelligence AI—is a type of artificial intelligence capable of performing the full spectrum...
    131 KB (14,472 words) - 20:27, 29 April 2025
  • former OpenAI researcher Paul Christiano, ARC focuses on recognizing and comprehending the potentially harmful capabilities of present-day AI models. ARC's...
    7 KB (601 words) - 14:38, 25 February 2025
  • Thumbnail for Mistral AI
    Mistral AI SAS is a French artificial intelligence (AI) startup, headquartered in Paris. It specializes in open-weight large language models (LLMs). The...
    27 KB (1,716 words) - 03:22, 29 April 2025
  • Thumbnail for Artificial Intelligence Act
    general-purpose AI, transparency requirements are imposed, with reduced requirements for open source models, and additional evaluations for high-capability models...
    32 KB (3,189 words) - 06:11, 12 April 2025
  • numerous AI ethics guidelines have been published in order to maintain social control over the technology. Regulation is deemed necessary to both foster AI innovation...
    134 KB (13,317 words) - 11:33, 30 April 2025
  • entities (including Shield AI) on its export control list, barring the export of dual-use commodities to that business. Shield AI employs machine learning...
    13 KB (1,284 words) - 11:16, 6 April 2025
  • DeepSeek (redirect from AI.com)
    Incentivizing Reasoning Capability in LLMs via Reinforcement Learning, arXiv:2501.12948 Gibney, Elizabeth (23 January 2025). "China's cheap, open AI model DeepSeek...
    62 KB (6,059 words) - 16:53, 1 May 2025
  • AI systems, as well as "tracking highly capable AI systems and large pools of computational capability" and "robust public funding for technical AI safety...
    13 KB (1,412 words) - 19:48, 16 April 2025
  • against and showing the capability to defeat professional teams. By choosing a game as complex as Dota 2 to study machine learning, OpenAI thought they could...
    23 KB (2,260 words) - 09:10, 6 April 2025
  • Thumbnail for History of artificial intelligence
    as little as 15 seconds of audio to reproduce a voice—a capability later corroborated by OpenAI in 2024. The service went viral on social media platforms...
    166 KB (19,442 words) - 15:28, 29 April 2025
  • proposed various approaches to mitigate risks associated with ASI: Capability control – Limiting an ASI's ability to influence the world, such as through...
    43 KB (4,588 words) - 22:19, 27 April 2025
  • intelligence (AI) startup company founded in 2021. Anthropic has developed a family of large language models (LLMs) named Claude as a competitor to OpenAI's ChatGPT...
    31 KB (2,841 words) - 09:41, 26 April 2025
  • in-memory computing capability. As of 2024[update], a typical AI integrated circuit chip contains tens of billions of MOSFETs. AI accelerators are used...
    51 KB (4,926 words) - 10:31, 10 April 2025