Generated by DeepSeek V3.2| artificial general intelligence | |
|---|---|
| Name | artificial general intelligence |
| Synonyms | strong AI, full AI, human-level AI |
| Field | Artificial intelligence, Cognitive science, Philosophy of mind |
| Key people | John McCarthy, Marvin Minsky, Ray Kurzweil, Nick Bostrom |
| Related concepts | Artificial superintelligence, Machine learning, Turing test, Chinese room |
artificial general intelligence. Artificial general intelligence (AGI) refers to a hypothetical type of artificial intelligence that possesses the ability to understand, learn, and apply its intelligence to solve any problem, much like a human being. This concept contrasts with contemporary narrow AI, which excels in specific tasks like playing chess or recognizing speech but lacks broad, adaptable understanding. The pursuit of AGI represents a central, long-term goal for many researchers in fields like computer science and neuroscience, aiming to create machines with generalized cognitive capabilities.
A core characteristic of AGI is cognitive flexibility, enabling it to transfer knowledge across vastly different domains, a feat demonstrated by humans but not by current machine learning systems. Researchers like Ben Goertzel and the team at OpenAI often describe it in terms of achieving or surpassing human performance across a wide range of economically valuable tasks. Key hypothesized attributes include reasoning, problem-solving, abstract thinking, and comprehension of complex ideas, integrating skills seen in separate narrow AI systems. The ultimate benchmark is often considered passing a fully robust version of the Turing test, requiring mastery of natural language and social intelligence.
Current systems like DeepMind's AlphaGo or IBM's Watson are paradigmatic examples of narrow AI, designed for excellence within a constrained set of parameters. While AlphaGo mastered the board game Go, it cannot apply that learning to navigate a city or write a sonnet. In contrast, AGI would not be limited to a single domain; the same system that diagnoses medical conditions could also compose music or devise a scientific theory. This fundamental difference highlights that progress in creating advanced specialized algorithms does not directly equate to progress toward the general learning and reasoning hallmarks of AGI.
Major research approaches include symbolic AI, which uses logic and knowledge representation, and connectionism, which focuses on neural networks inspired by the human brain. Institutions like the Massachusetts Institute of Technology and companies such as DeepMind and Anthropic explore reinforcement learning and large language models as potential pathways. Alternative frameworks include cognitive architectures like ACT-R and Soar, which attempt to model unified theories of cognition. Projects like the Human Brain Project in the European Union seek insights from neuroscience to inform AGI development.
If realized, AGI could drive unprecedented advances in fields like scientific research, medicine, and space exploration, potentially solving grand challenges like climate change or disease. However, prominent figures like Nick Bostrom of the Future of Humanity Institute and the late Stephen Hawking have warned of existential risks, including the difficulty of aligning AGI's goals with human values, a problem known as the AI alignment problem. The transition from AGI to artificial superintelligence, a vastly more powerful entity, could happen rapidly, posing profound control challenges as explored in works like Superintelligence: Paths, Dangers, Strategies.
The development of AGI raises urgent questions about consciousness, personhood, and rights, debated by philosophers like Daniel Dennett. Economic impacts, including widespread automation and potential technological unemployment, are a major concern for economists and policymakers at organizations like the World Economic Forum. Issues of bias, fairness, and transparency in powerful cognitive systems necessitate robust governance frameworks, as discussed by institutes like the AI Now Institute. International coordination, perhaps through bodies like the United Nations, may be required to manage geopolitical competition and ensure safe, equitable development.
The term and foundational goals were crystallized at the Dartmouth Workshop of 1956, organized by pioneers like John McCarthy and Marvin Minsky. Early optimism in the 1960s, fueled by work at places like the Stanford AI Lab and the MIT AI Lab, gave way to the "AI winter" periods of reduced funding and progress. The 21st century has seen a resurgence with advances in deep learning and computational power from companies like NVIDIA. Modern milestones include increasingly sophisticated large language models from OpenAI and Google, though experts like Rodney Brooks and Yoshua Bengio debate the timeline and feasibility of achieving true AGI. Category:Artificial intelligence Category:Emerging technologies Category:Futures studies