Jump to content

Talk:Artificial intelligence/Where did it go? 2021

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia

In the summer and fall of 2021, I copy-edited the entire article for redundancy, WP:RELEVANCE, WP:UNDUE weight, organization and citation format. Most of the material was moved to sub-articles, such as applications of AI, artificial general intelligence, history of AI and so on. Some material (marked "Not Done" below) didn't seem to fit in anywhere, or was difficult to save for one reason or another. ---- CharlesGillingham (talk) 23:16, 6 October 2021 (UTC)[reply]

From History

[edit]

 Done These have been moved to Applications of AI. All but three of these have a one sentence mention in Artificial intelligence § Appliations --- CharlesGillingham (talk) 16:31, 29 September 2021 (UTC)[reply]

 Done Moved to Artificial intelligence § Applications

By 2020, Natural Language Processing systems such as the enormous GPT-3 (then by far the largest artificial neural network) were matching human performance on pre-existing benchmarks, albeit without the system attaining commonsense understanding of the contents of the benchmarks.[18]

 Not done China's AI program is not (yet) the most important trend of the decade. Perhaps the paragraph on the 2020s will use this. ---- CharlesGillingham (talk) 18:44, 12 September 2021 (UTC)[reply]

Around 2016, China greatly accelerated its government funding; given its large supply of data and its rapidly increasing research output, some observers believe it may be on track to becoming an "AI superpower".[19][20]

From Basics

[edit]

The article had a section called Basics which was an article-within-the-article. This is very well written, well sourced and accurate, but it is completely redundant. We still need to look at the best bits and see if they aren't better than what we already have on those topics and replace what we have if that's a good idea.

 Done Moved to intelligent agent

Computer science defines AI research as the study of "intelligent agents": any device that perceives its environment and takes actions that maximize its chance of successfully achieving its goals.[a] A more elaborate definition characterizes AI as "a system's ability to correctly interpret external data, to learn from such data, and to use those learnings to achieve specific goals and tasks through flexible adaptation."[21]

 Done Moved to intelligent agent

A typical AI analyzes its environment and takes actions that maximize its chance of success.[a] An AI's intended utility function (or goal) can be simple ("1 if the AI wins a game of Go, 0 otherwise") or complex ("Perform actions mathematically similar to ones that succeeded in the past"). Goals can be explicitly defined or induced. If the AI is programmed for "reinforcement learning", goals can be implicitly induced by rewarding some types of behavior or punishing others.[b] Alternatively, an evolutionary system can induce goals by using a "fitness function" to mutate and preferentially replicate high-scoring AI systems, similar to how animals evolved to innately desire certain goals such as finding food.[22] Some AI systems, such as nearest-neighbor, instead of reason by analogy, these systems are not generally given goals, except to the degree that goals are implicit in their training data.[23] Such systems can still be benchmarked if the non-goal system is framed as a system whose "goal" is to accomplish its narrow classification task.[24]

 Not done Where? Algorithm?

AI often revolves around the use of algorithms. An algorithm is a set of unambiguous instructions that a mechanical computer can execute.[c] A complex algorithm is often built on top of other, simpler, algorithms. A simple example of an algorithm is the following (optimal for first player) recipe for play at tic-tac-toe:[25]
  1. If someone has a "threat" (that is, two in a row), take the remaining square. Otherwise,
  2. if a move "forks" to create two threats at once, play that move. Otherwise,
  3. take the center square if it is free. Otherwise,
  4. if your opponent has played in a corner, take the opposite corner. Otherwise,
  5. take an empty corner if one exists. Otherwise,
  6. take any empty square.

TODO Heuristic learning?

Many AI algorithms are capable of learning from data; they can enhance themselves by learning new heuristics (strategies, or "rules of thumb", that have worked well in the past), or can themselves write other algorithms.

 Not done Dubious.

Some of the "learners" described below, including Bayesian networks, decision trees, and nearest-neighbor, could theoretically, (given infinite data, time, and memory) learn to approximate any function,including which combination of mathematical functions would best describe the world.[citation needed] These learners could therefore derive all possible knowledge, by considering every possible hypothesis and matching them against the data.

TODO Move to Intractability (just the example)

In practice, it is seldom possible to consider every possibility, because of the phenomenon of "combinatorial explosion", where the time needed to solve a problem grows exponentially. Much of AI research involves figuring out how to identify and avoid considering a broad range of possibilities unlikely to be beneficial.[26] For example, when viewing a map and looking for the shortest driving route from Denver to New York in the East, one can in most cases skip looking at any path through San Francisco or other areas far to the West; thus, an AI wielding a pathfinding algorithm like A* can avoid the combinatorial explosion that would ensue if every possible route had to be ponderously considered.[27]

 Not done This is really good, especially the examples, but I'm not sure where to work it into the article or anywhere else in Wikipedia. The Tools section basically covers these same points in the same order. Could it work there?

The earliest (and easiest to understand) approach to AI was symbolism (such as formal logic): "If an otherwise healthy adult has a fever, then they may have influenza".
A second, more general, approach is Bayesian inference: "If the current patient has a fever, adjust the probability they have influenza in such-and-such way".
The third major approach, extremely popular in routine business AI applications, are analogizers such as SVM and nearest-neighbor: "After examining the records of known past patients whose temperature, symptoms, age, and other factors mostly match the current patient, X% of those patients turned out to have influenza".
A fourth approach is harder to intuitively understand, but is inspired by how the brain's machinery works: the artificial neural network approach uses artificial "neurons" that can learn by comparing itself to the desired output and altering the strengths of the connections between its internal neurons to "reinforce" connections that seemed to be useful. These four main approaches can overlap with each other and with evolutionary systems; for example, neural nets can learn to make inferences, to generalize, and to make analogies. Some systems implicitly or explicitly use multiple of these approaches, alongside many other AI and non-AI algorithms; the best approach is often different depending on the problem.[28][29]

 Done Moved to Machine learning

Learning algorithms work on the basis that strategies, algorithms, and inferences that worked well in the past are likely to continue working well in the future. These inferences can be obvious, such as "since the sun rose every morning for the last 10,000 days, it will probably rise tomorrow morning as well". They can be nuanced, such as "X% of families have geographically separate species with color variants, so there is a Y% chance that undiscovered black swans exist".[30]

 Done I'm not confident about where this fits into machine learning, so I can't put it anywhere myself. Sending it to Talk:Machine learning.

Learners can also work on the basis of "Occam's razor": The simplest theory that explains the data is the likeliest. Therefore, according to Occam's razor principle, a learner must be designed such that it prefers simpler theories to complex theories, except in cases where the complex theory is proven substantially better.[30]

 Done Moved to Machine learning § Limitations

The blue line could be an example of overfitting a linear function due to random noise.
Settling on a bad, overly complex theory gerrymandered to fit all the past training data is known as overfitting. Many systems attempt to reduce overfitting by rewarding a theory in accordance with how well it fits the data, but penalizing the theory in accordance with how complex the theory is.[30] Besides classic overfitting, learners can also disappoint by "learning the wrong lesson". A toy example is that an image classifier trained only on pictures of brown horses and black cats might conclude that all brown patches are likely to be horses.[31] A real-world example is that, unlike humans, current image classifiers often don't primarily make judgments from the spatial relationship between components of the picture, and they learn relationships between pixels that humans are oblivious to, but that still correlate with images of certain types of real objects. Modifying these patterns on a legitimate image can result in "adversarial" images that the system misclassifies.[d][32][33]

 Done Point is already made in Artificial intelligence § knowledge. This text appears in Commonsense reasoning.

A self-driving car system may use a neural network to determine which parts of the picture seem to match previous training images of pedestrians, and then model those areas as slow-moving but somewhat unpredictable rectangular prisms that must be avoided.

AI lacks several features of human "commonsense reasoning"; most notably, humans have powerful mechanisms for reasoning about "naïve physics" such as space, time, and physical interactions. This enables even young children to easily make inferences like "If I roll this pen off a table, it will fall on the floor". Humans also have a powerful mechanism of "folk psychology" that helps them to interpret natural-language sentences such as "The city councilmen refused the demonstrators a permit because they advocated violence" (A generic AI has difficulty discerning whether the ones alleged to be advocating violence are the councilmen or the demonstrators[34][35][36]).

This lack of "common knowledge" means that AI often makes different mistakes than humans make, in ways that can seem incomprehensible. For example, existing self-driving cars cannot reason about the location nor the intentions of pedestrians in the exact way that humans do, and instead must use non-human modes of reasoning to avoid accidents.[37][38][39]

From Goals

[edit]

From Goals/Lede

[edit]

 Not done Is this a re-invention/re-framing of symbolic vs. sub-symbolic? Perhaps it could go in symbolic AI; although I would really like to see this in a WP:SECONDARY source. ---- CharlesGillingham (talk) 04:18, 6 October 2021 (UTC)[reply]

The cognitive capabilities of current architectures are very limited, using only a simplified version of what intelligence is really capable of. For instance, the human mind has come up with ways to reason beyond measure and logical explanations to different occurrences in life. What would have been otherwise straightforward, an equivalently difficult problem may be challenging to solve computationally as opposed to using the human mind. This gives rise to two classes of models: structuralist and functionalist. The structural models aim to loosely mimic the basic intelligence operations of the mind such as reasoning and logic. The functional model refers to the correlating data to its computed counterpart.[40]

From Social Intelligence

[edit]

 Done The first source is actually about technological employment (the paradox is relevant because computers are bad at perceptual and motor tasks). Moved the source to artificial intelligence § Technological unemployment. The second source is about giving AI programs a "theory of (other) minds", which is a form of social intelligence. Added the citation to Artificial intelligence § Social intelligence ---- CharlesGillingham (talk) 22:09, 6 October 2021 (UTC)[reply]

Moravec's paradox can be extended to many forms of social intelligence.[41][42]

 Not done Too vague to be useful anywhere.

Distributed multi-agent coordination of autonomous vehicles remains a difficult problem.[43]

From General Intelligence

[edit]

 Done Cyc is covered in the article History of AI as well as Artificial general intelligence, FGCP is covered in a footnote in Artificial intelligence § History (UTC)

Historically, projects such as the Cyc knowledge base (1984–) and the massive Japanese Fifth Generation Computer Systems initiative (1982–1992) attempted to cover the breadth of human cognition. These early projects failed to escape the limitations of non-quantitative symbolic logic models and, in retrospect, greatly underestimated the difficulty of cross-domain AI.

 Done This has been moved to Applications of AI

One high-profile example is that DeepMind in the 2010s developed a "generalized artificial intelligence" that could learn many diverse Atari games on its own, and later developed a variant of the system which succeeds at sequential learning.[44][45][46]

 Done This is added to artificial intelligence § Learning

 Not done This is unclear. However, the source is perfect and the point is good. Needs layman's language (first half) and encyclopedic tone (second half)

hypothetical AGI breakthroughs could include the development of reflective architectures that can engage in decision-theoretic metareasoning, and figuring out how to "slurp up" a comprehensive knowledge base from the entire unstructured Web.[48]

 Done Moved to artificial general intelligence

Many of the problems in this article may also require general intelligence, if machines are to solve the problems as well as people do. For example, even specific straightforward tasks, like machine translation, require that a machine read and write in both languages (NLP), follow the author's argument (reason), know what is being talked about (knowledge), and faithfully reproduce the author's original intent (social intelligence). A problem like machine translation is considered "AI-complete", because all of these problems need to be solved simultaneously in order to reach human-level machine performance.

From "Approaches"

[edit]

Before 2021, the article had a section called "Approaches". This has been divided between History, Philosophy and the sub-articles. ---- CharlesGillingham (talk) 18:21, 12 September 2021 (UTC)[reply]

From Symbolic AI

[edit]

 Done I moved this section into Symbolic AI. The history Symbolic AI of is described in two paragraphs of Artificial Intelligence § History, and the weaknesses and strengths of the approach are describe in the section Artificial intelligence § Symbolic AI and its limits

During the 1960s, symbolic approaches achieved great success at simulating intelligent behavior in small demonstration programs. AI research was centered in three institutions in the 1960s: Carnegie Mellon University, Stanford, MIT and (later) University of Edinburgh. Each one developed its own style of research. Earlier approaches based on cybernetics or artificial neural networks were abandoned or pushed into the background.
Cognitive simulation

Economist Herbert Simon and Allen Newell studied human problem-solving skills and attempted to formalize them, and their work laid the foundations of the field of artificial intelligence, as well as cognitive science, operations research and management science. Their research team used the results of psychological experiments to develop programs that simulated the techniques that people used to solve problems.[49][50] This tradition, centered at Carnegie Mellon University would eventually culminate in the development of the Soar architecture in the middle 1980s.[51][52]

Logic-based

Unlike Simon and Newell, John McCarthy felt that machines did not need to simulate human thought, but should instead try to find the essence of abstract reasoning and problem-solving, regardless of whether people used the same algorithms.[e] His laboratory at Stanford (SAIL) focused on using formal logic to solve a wide variety of problems, including knowledge representation, planning and learning.[57] Logic was also the focus of the work at the University of Edinburgh and elsewhere in Europe which led to the development of the programming language Prolog and the science of logic programming.[58][59]

Anti-logic or "scruffy"

Researchers at MIT (such as Marvin Minsky and Seymour Papert)[60][61][62] found that solving difficult problems in vision and natural language processing required ad hoc solutions—they argued that no simple and general principle (like logic) would capture all the aspects of intelligent behavior. Roger Schank described their "anti-logic" approaches as "scruffy" (as opposed to the "neat" paradigms at CMU and Stanford).[63][64] Commonsense knowledge bases (such as Doug Lenat's Cyc) are an example of "scruffy" AI, since they must be built by hand, one complicated concept at a time.[65][66][67]

Knowledge based
When computers with large memories became available around 1970, researchers from all three traditions began to build knowledge into AI applications.[68][69] The knowledge revolution was driven by the realization that enormous amounts of knowledge would be required by many simple AI applications.

From Embodied Intelligence

[edit]

 Done The coverage is sufficient, and of course this definition is in the article developmental robotics. --- CharlesGillingham (talk) 03:02, 1 October 2021 (UTC)[reply]

Within developmental robotics, developmental learning approaches are elaborated upon to allow robots to accumulate repertoires of novel skills through autonomous self-exploration, social interaction with human teachers, and the use of guidance mechanisms (active learning, maturation, motor synergies, etc.).[70][71][72][73]

From Integrating the Approaches

[edit]

TODO Artificial intelligence § General intelligence mentions cognitive architectures and multi-agent systems as approaches to AGI, and the others here are mentioned in a footnote. Technically, I can't call this "Done" because our article doesn't acknowledge that these are also used as tools for particular applications. Still might need to have a (very short) section on this stuff in Tools.

From Tools

[edit]

From Logic

[edit]

 Done These points have been moved into Fuzzy logic#Applications.

Fuzzy logic is successfully used in control systems to allow experts to contribute vague rules such as "if you are close to the destination station and moving fast, increase the train's brake pressure"; these vague rules can then be numerically refined within the system. Fuzzy logic fails to scale well in knowledge bases; many AI researchers question the validity of chaining fuzzy-logic inferences.[f][78][79]

From neural networks

[edit]

 Not done I think that the author of this was trying explain that information is distributed throughout the network, rather than being stored in a specific location (as it would be with symbolic AI). however using the word "concepts" (which has a specific meaning in cognitive science) is a misleading way to describe this -- it actually confuses the issue. This is also unsourced. Perhaps someone else can figure out what the original author meant and say it better.

The neural network forms "concepts" that are distributed among a subnetwork of shared[g] neurons that tend to fire together; a concept meaning "leg" might be coupled with a subnetwork meaning "foot" that includes the sound for "foot".

 Done Moved to Artificial neural network § History

Neural networks' early successes included predicting the stock market and (in 1995) a mostly self-driving car.[h][80]: Chapter 4 

 Done This point is made Artificial intelligence § History

In the 2010s, advances in neural networks using deep learning thrust AI into widespread public consciousness and contributed to an enormous upshift in corporate AI spending;

 Not done Artificial intelligence § History already reports two excellent metrics of the uptick in AI interest 2015-2020 (total publications, corporate spending). This is not a particularly notable metric, and we can't really use it when we have better ones.

AI-related M&A in 2017 was over 25 times as large as in 2015.[81][82]

 Done Frank Rosenblatt is discussed in the History of AI, and Pitts & McCullough is mentioned there and in Artificial intelligence § History.

The study of non-learning neural networks began in the decade before the field of AI research was founded, in the work of Walter Pitts and Warren McCullouch. Frank Rosenblatt invented the perceptron, a learning network with a single layer, similar to the old concept of linear regression.

 Not done without sources, Wikipedia can't really make any assertion about their importance.

 Done Linnaimaa is credited in a footnote.

which has been around since 1970 as the reverse mode of automatic differentiation published by Seppo Linnainmaa,[83][84] and was introduced to neural networks by Paul Werbos.[85][86][87]

 Not done WP:UNDUE weight on this approach. Can't really move this to an AI sub-article either, because it's not really in use -- biologically based AI, maybe?

Hierarchical temporal memory is an approach that models some of the structural and algorithmic properties of the neocortex.[88]

 Done Similarly, this is probably WP:UNDUE weight on this approach. Moved to Artificial neural network

However, some research groups, such as Uber, argue that simple neuroevolution to mutate new neural network topologies and weights may be competitive with sophisticated gradient descent approaches[citation needed]. One advantage of neuroevolution is that it may be less prone to get caught in "dead ends".[89]

From Feedforward Networks

[edit]

 Done All this precedence is covered in Deep learning

According to one overview,[90]the expression "Deep Learning" was introduced to the machine learning community by Rina Dechter in 1986[91] and gained traction after Igor Aizenberg and colleagues introduced it to artificial neural networks in 2000.[92] The first functional deep Learning networks were published by Alexey Grigorevich Ivakhnenko and V. G. Lapa in 1965.[93] These networks are trained one layer at a time. Ivakhnenko's 1971 paper[94] describes the learning of a deep feedforward multilayer perceptron with eight layers, already much deeper than many later networks.

 Done Too much undefined WP:JARGON. Significance isn't clear. Covered in Deep learning.

In 2006, a publication by Geoffrey Hinton and Ruslan Salakhutdinov introduced another way of pre-training many-layered feedforward neural networks (FNNs) one layer at a time, treating each layer in turn as an unsupervised restricted Boltzmann machine, then using supervised backpropagation for fine-tuning.[95] Similar to shallow artificial neural networks, deep neural networks can model complex non-linear relationships.

 Done More precedence. Covered in Deep learning.

(CNNs), whose origins can be traced back to the Neocognitron introduced by Kunihiko Fukushima in 1980.[96] In 1989, Yann LeCun and colleagues applied backpropagation to such an architecture.

 Done Covered in Deep learning

In the early 2000s, in an industrial application, CNNs already processed an estimated 10% to 20% of all the checks written in the US.[97] Since 2011, fast implementations of CNNs on GPUs have won many visual pattern recognition competitions.

 Done The article has enough detail about AlphaGo. Moved to AlphaGo.

CNNs with 12 convolutional layers were used with reinforcement learning by Deepmind's "AlphaGo Lee", the program that beat a top Go champion in 2016.[98]

From Deep recurrent neural networks

[edit]

 Not done This is unsourced (but unlikely to be challenged). Still, don't think we need it, since there are more applications today.

 Done Just the source.

(RNNs)[99]

 Done Moved to Recurrent neural networks, where this fact did not appear (and thus probably not notable enough for this article).

recurrent neural network are theoretically Turing complete and can run arbitrary programs to process arbitrary sequences of inputs.[100]

 Done Kept, Edited for brevity.

The depth of an RNN is unlimited and depends on the length of its input sequence; thus, an RNN is an example of deep learning.[101]

 Done Schmidhuber's work 1991-92 is described in Recurrent neural network.

In 1992, it was shown that unsupervised pre-training of a stack of recurrent neural networks can speed up subsequent supervised learning of deep sequential problems.[102]

 Done LSTM is mentioned, with this source.

Numerous researchers now use variants of a deep learning recurrent NN called the long short-term memory (LSTM) network published by Hochreiter & Schmidhuber in 1997.[103]

 Done Undefined WP:JARGON. This is covered in Recurrent Neural Network.

LSTM is often trained by Connectionist Temporal Classification (CTC).[104]

 Done Applications of LSTM. These projects are described in Recurrent neural network § LSTM, with the same sources.

At Google, Microsoft and Baidu this approach has revolutionized speech recognition.[105] For example, in 2015, Google's speech recognition experienced a dramatic performance jump of 49% through CTC-trained LSTM. Google also used LSTM to improve machine translation,[106] language modeling,[107] and multilingual language processing.[108] LSTM combined with CNNs also improved automatic image captioning[109] and a plethora of other applications.

From Applications

[edit]

 Done Moved to Applications of AI

With social media sites overtaking TV as a source for news for young people and news organizations increasingly reliant on social media platforms for generating distribution,[110]

From Evaluating progress

[edit]

 Done Moved into Applications of AI.

AI, like electricity or the steam engine, is a general purpose technology. There is no consensus on how to characterize which tasks AI tends to excel at.[111]

 Done Moved into Applications of AI.

While projects such as AlphaZero have succeeded in generating their own knowledge from scratch, many other machine learning projects require large training datasets.[112][113]

 Done Moved into Moravec's paradox. ---- CharlesGillingham (talk) 16:41, 12 October 2021 (UTC)[reply]

Researcher Andrew Ng has suggested, as a "highly imperfect rule of thumb", that "almost anything a typical human can do with less than one second of mental thought, we can probably now or in the near future automate using AI."[114]

 Done Moravec's paradox is covered Artificial intelligence § Symbolic AI and its limits

Moravec's paradox suggests that AI lags humans at many tasks that the human brain has specifically evolved to perform well.[115]

 Done Games & AlphaGo are covered in Artificial intelligence § Applications

Games provide a well-publicized benchmark for assessing rates of progress. AlphaGo around 2016 brought the era of classical board-game benchmarks to a close.

 Done This appears in Progress in artificial intelligence.

Games of imperfect knowledge provide new challenges to AI in game theory.[116][117]

 Done This is moved to Applications of AI.

E-sports such as StarCraft continue to provide additional public benchmarks.[118][119]

 Done This has been added to Progress in artificial intelligence

Many competitions and prizes, such as the Imagenet Challenge, promote research in artificial intelligence. The most common areas of competition include general machine intelligence, conversational behavior, data-mining, robotic cars, and robot soccer as well as conventional games.[120]

 Done This appears in Progress in artificial intelligence

The "imitation game" (an interpretation of the 1950 Turing test that assesses whether a computer can imitate a human) is nowadays considered too exploitable to be a meaningful benchmark.[121] A derivative of the Turing test is the Completely Automated Public Turing test to tell Computers and Humans Apart (CAPTCHA). As the name implies, this helps to determine that a user is an actual person and not a computer posing as a human. Unlike the standard Turing test, CAPTCHA is administered by a machine and targeted to a human as opposed to being administered by a human and targeted to a machine. A computer asks a user to complete a simple test then generates a grade for that test. Computers are unable to solve the problem, so correct solutions are deemed to be the result of a person taking the test. A common type of CAPTCHA is the test that requires the typing of distorted letters, numbers or symbols that appear in an image undecipherable by a computer.[122]

 Done This appears in Progress in artificial intelligence

Proposed "universal intelligence" tests aim to compare how well machines, humans, and even non-human animals perform on problem sets that are generic as possible. At an extreme, the test suite can contain every possible problem, weighted by Kolmogorov complexity; unfortunately, these problem sets tend to be dominated by impoverished pattern-matching exercises where a tuned AI can easily exceed human performance levels.[123][124][125][126]

 Done Moved to Hardware for artificial intelligence

Since the 2010s, advances in both machine learning algorithms and computer hardware have led to more efficient methods for training deep neural networks that contain many layers of non-linear hidden units and a very large output layer.[127] By 2019, graphic processing units (GPUs), often with AI-specific enhancements, had displaced CPUs as the dominant method of training large-scale commercial cloud AI.[128] OpenAI estimated the hardware compute used in the largest deep learning projects from AlexNet (2012) to AlphaZero (2017), and found a 300,000-fold increase in the amount of compute required, with a doubling-time trendline of 3.4 months.[129][130]

From Philosophy

[edit]

 Done This is covered in Philosophy of AI

In the proposal for the Dartmouth Workshop of 1956, John McCarthy wrote "Every aspect of learning or any other feature of intelligence can be so precisely described that a machine can be made to simulate it."[131]

 Done This is covered in Philosophy of AI

Kurt Gödel,[132] John Lucas (in 1961) and Roger Penrose (in a more detailed argument from 1989 onwards) made highly technical arguments that human mathematicians can consistently see the truth of their own "Gödel statements" and therefore have computational abilities beyond that of mechanical Turing machines.[133] However, some people do not agree with the "Gödelian arguments".[134][135][136]

 Done The AI effect has been covered in the Lede and in Applications.

The AI effect claims that machines are already intelligent, but observers have failed to recognize it. For example, when Deep Blue beat Garry Kasparov in chess, the machine could be described as exhibiting intelligence. However, onlookers commonly discount the behavior of an artificial intelligence program by arguing that it is not "real" intelligence, with "real" intelligence being in effect defined as whatever behavior machines cannot do.

 Done This is covered in philosophy of AI

The artificial brain argument asserts that the brain can be simulated by machines and, because brains exhibit intelligence, these simulated brains must also exhibit intelligence − ergo, machines can be intelligent. Hans Moravec, Ray Kurzweil and others have argued that it is technologically feasible to copy the brain directly into hardware and software, and that such a simulation will be essentially identical to the original.[137]

From Future of AI

[edit]

From Singularity

[edit]

 Done Kurzweil's prediction is covered in artificial general intelligence

Ray Kurzweil has used Moore's law (which describes the relentless exponential improvement in digital technology) to calculate that desktop computers will have the same processing power as human brains by the year 2029 and predicts that the singularity will occur in 2045.[138]

From Robot Rights

[edit]

 Done Plug & Play is mentioned in the footnote

The subject is profoundly discussed in the 2010 documentary film Plug & Pray

 Not done Can't really move this into artificial intelligence in fiction because that article is tightly structured and there's no place for this topic at the moment.

and many sci fi media such as Star Trek Next Generation, with the character of Commander Data, who fought being disassembled for research, and wanted to "become human", and the robotic holograms in Voyager.

From Risks

[edit]

 Not done This is devoid of actual content about AI, and too vague to be useful in existential risk of artificial intelligence

The potential negative effects of AI and automation were a major issue for Andrew Yang's 2020 presidential campaign in the United States.[139]

 Not done Redundant. The points that Beridze is making are vague and are covered in more detail elsewhere in the article. Added this citation to a paragraph about the same concerns citing Musk, Gates and Hawkins. Also a bit vague to be useful in Existential risk of AI

Irakli Beridze, Head of the Centre for Artificial Intelligence and Robotics at UNICRI, United Nations, has expressed that "I think the dangerous applications for AI, from my point of view, would be criminals or large terrorist organizations using it to disrupt large processes or simply do pure harm. [Terrorists could cause harm] via digital warfare, or it could be a combination of robotics, drones, with AI and other things as well that could be really dangerous. And, of course, other risks come from things like job losses. If we have massive numbers of people losing jobs and don't find a solution, it will be extremely dangerous. Things like lethal autonomous weapons systems should be properly governed—otherwise there's massive potential of misuse."[140]

From technological unemployment

[edit]

 Done Redundant: Each contribution seemed to want to introduce the topic again.

  • The long-term economic effects of AI are uncertain.
  • about whether the increasing use of robots and AI will cause a substantial increase in long-term unemployment

 Done Redundant: This point was made twice, and I chose the one based on Ford. Keeping the reference.

The relationship between automation and employment is complicated. While automation eliminates old jobs, it also creates new jobs through micro-economic and macro-economic effects.[141]

 Not done These were off-topic

  • A 2017 study by PricewaterhouseCoopers sees the People's Republic of China gaining economically the most out of AI with 26.1% of GDP until 2030.[142]
  • A February 2020 European Union white paper on artificial intelligence advocated for artificial intelligence for economic benefits, including "improving healthcare (e.g. making diagnosis more precise, enabling better prevention of diseases), increasing the efficiency of farming, contributing to climate change mitigation and adaptation, [and] improving the efficiency of production systems through predictive maintenance", while acknowledging potential risks.[143]

 Done Moved to technological unemployment

Author Martin Ford and others go further and argue that many jobs are routine, repetitive and (to an AI) predictable; Ford warns that these jobs may be automated in the next couple of decades, and that many of the new jobs may not be "accessible to people with average capability", even with retraining.[144]

From Existential Risk

[edit]

 Done Kept a sentence of this. This point is also made in Existential risk of AI (in three places).

Physicist Stephen Hawking, Microsoft founder Bill Gates, history professor Yuval Noah Harari, and SpaceX founder Elon Musk have expressed concerns about the possibility that AI could evolve to the point that humans could not control it, with Hawking theorizing that this could "spell the end of the human race".[145][146][147][148]

The development of full artificial intelligence could spell the end of the human race. Once humans develop artificial intelligence, it will take off on its own and redesign itself at an ever-increasing rate. Humans, who are limited by slow biological evolution, couldn't compete and would be superseded.

 Done Kept one sentence of each of this, the whole paragraph is moved to Existential risk of AI

In his book Superintelligence, philosopher Nick Bostrom provides an argument that artificial intelligence will pose a threat to humankind. He argues that sufficiently intelligent AI, if it chooses actions based on achieving some goal, will exhibit convergent behavior such as acquiring resources or protecting itself from being shut down. If this AI's goals do not fully reflect humanity's—one example is an AI told to compute as many digits of pi as possible—it might harm humanity in order to acquire more resources or prevent itself from being shut down, ultimately to better achieve its goal.

 Done ame deal. Summary in AI, all the text moved to Existential risk of AI

Bostrom also emphasizes the difficulty of fully conveying humanity's values to an advanced AI. He uses the hypothetical example of giving an AI the goal to make humans smile to illustrate a misguided attempt. If the AI in that scenario were to become superintelligent, Bostrom argues, it may resort to methods that most humans would find horrifying, such as inserting "electrodes into the facial muscles of humans to cause constant, beaming grins" because that would be an efficient way to achieve its goal of making humans smile.[150]

 Done This is covered in Friendly AI

In his book Human Compatible AI researcher Stuart J. Russell echoes some of Bostrom's concerns while also proposing an approach to developing provably beneficial machines focused on uncertainty and deference to humans[151] possibly involving inverse reinforcement learning.[152]

 Done Kept one sentence or so from this, the entire paragraph moved to Existential risk of AI

Concern over risk from artificial intelligence has led to some high-profile donations and investments. A group of prominent tech titans including Peter Thiel, Amazon Web Services and Musk have committed $1 billion to OpenAI, a nonprofit company aimed at championing responsible AI development.[153] In January 2015, Elon Musk donated $10 million to the Future of Life Institute to fund research on understanding AI decision making. The goal of the institute is to "grow wisdom with which we manage" the growing power of technology. Musk also funds companies developing artificial intelligence such as DeepMind and Vicarious to "just keep an eye on what's going on with artificial intelligence.[154] I think there is potentially a dangerous outcome there."[155][156]

 Done This is moved to Existential risk of AI

The opinion of experts within the field of artificial intelligence is mixed, with sizable fractions both concerned and unconcerned by risk from eventual superhumanly-capable AI.[157]

 Done This is moved to Technological unemployment

Oracle CEO Mark Hurd has stated that AI "will actually create more jobs, not less jobs" as humans will be needed to manage AI systems.[158]

 Done This is in Existential risk of AI

Facebook CEO Mark Zuckerberg believes AI will "unlock a huge amount of positive things," such as curing disease and increasing the safety of autonomous cars.[159]

 Done This is in Existential risk of AI

For the danger of uncontrolled advanced AI to be realized, the hypothetical AI would have to overpower or out-think all of humanity, which a minority of experts argue is a possibility far enough in the future to not be worth researching.[160][161]

 Done This is in Existential risk of AI.

Other counterarguments revolve around humans being either intrinsically or convergently valuable from the perspective of an artificial intelligence.[162]

From Ethical machines

[edit]

 Done Everything here is either in ethics of AI or history of AI

Joseph Weizenbaum in Computer Power and Human Reason wrote that AI applications cannot, by definition, successfully simulate genuine human empathy and that the use of AI technology in fields such as customer service or psychotherapy[i] was deeply misguided. Weizenbaum was also bothered that AI researchers (and some philosophers) were willing to view the human mind as nothing more than a computer program (a position now known as computationalism). To Weizenbaum, these points suggest that AI research devalues human life.[164]

From Malevolent AI

[edit]

 Done A shortened version of this paragraph was moved up into the "weaponized A" section.

Lethal autonomous weapons are of concern. By 2015, over fifty countries were reported to be researching battlefield robots, including the United States, China, Russia, and the United Kingdom. Many people concerned about risk from superintelligent AI also want to limit the use of artificial soldiers and drones.[165]

 Done Added this citation and footnote with the quote to the "existential risk" section, because this is a response to the risk. Also added the full quote to Existential risk of AI

Leading AI researcher Rodney Brooks writes, "I think it is a mistake to be worrying about us developing malevolent AI anytime in the next few hundred years. I think the worry stems from a fundamental error in not distinguishing the difference between the very real recent advances in a particular aspect of AI and the enormity and complexity of building sentient volitional intelligence."[166]

 DoneMoved (a cut-down version of) this into "existential risk" because it is an argument that there is a risk. The remainder of this was moved into Existential risk of AI § Orthogonality thesis

Political scientist Charles T. Rubin believes that AI can be neither designed nor guaranteed to be benevolent.[167] He argues that "any sufficiently advanced benevolence may be indistinguishable from malevolence." Humans should not assume machines or robots would treat us favorably because there is no a priori reason to believe that they would be sympathetic to our system of morality, which has evolved along with our particular biology (which AIs would not share). Hyper-intelligent software may not necessarily decide to support the continued existence of humanity and would be extremely difficult to stop. This topic has also recently begun to be discussed in academic publications as a real source of risks to civilization, humans, and planet Earth.

From Regulation

[edit]

 Done All of these paragraphs (or equivalent) and their sources now appear in regulation of AI

Regulation of AI through mechanisms such as review boards can also be seen as social means to approach the AI control problem.[168]

The Global Partnership on Artificial Intelligence was launched in June 2020, stating a need for AI to be developed in accordance with human rights and democratic values, to ensure public confidence and trust in the technology, as outlined in the OECD Principles on Artificial Intelligence (2019).[169] The founding members of the Global Partnership on Artificial Intelligence are Australia, Canada, the European Union, France, Germany, India, Italy, Japan, Rep. Korea, Mexico, New Zealand, Singapore, Slovenia, the US and the UK. The GPAI Secretariat is hosted by the OECD in Paris, France. GPAI's mandate covers four themes, two of which are supported by the International Centre of Expertise in Montréal for the Advancement of Artificial Intelligence, namely, responsible AI and data governance. A corresponding centre of excellence in Paris, yet to be identified, will support the other two themes on the future of work and innovation, and commercialization. GPAI will also investigate how AI can be leveraged to respond to the COVID-19 pandemic.[169]

UNESCO will be tabling an international instrument on the ethics of AI for adoption by 192 member states in November 2021.[169]

Given the concerns about data exploitation, the European Union also developed an artificial intelligence policy, with a working group studying ways to assure confidence in the use of artificial intelligence. These were issued in two white papers in the midst of the COVID-19 pandemic. One of the policies on artificial intelligence is called A European Approach to Excellence and Trust.[170][171][172]

From Fiction

[edit]

 Not done This section is about fiction, and we only have room to cover the most popular tropes. This material below doesn't illustrate a major trope and places WP:UNDUE on this artist for this article (and is unsourced). Could not find a a place for this, as artificial intelligence in fiction has a very tight structure at this point and doesn't seem to be ready to accept discussion of random works.

In the 1980s, artist Hajime Sorayama's Sexy Robots series were painted and published in Japan depicting the actual organic human form with lifelike muscular metallic skins and later "the Gynoids" book followed that was used by or influenced movie makers including George Lucas and other creatives. Sorayama never considered these organic robots to be real part of nature but always an unnatural product of the human mind, a fantasy existing in the mind even when realized in actual form.

Citations needed for the material above

[edit]

When the material above is moved into a sub article, we will need the citation it used. You should be able to find them here. Note that the citation format of the article was all over the map. ---- CharlesGillingham (talk) 09:02, 24 September 2021 (UTC)[reply]

Notes

[edit]
  1. ^ a b AI as intelligent agents (full note in artificial intelligence
  2. ^ The act of doling out rewards can itself be formalized or automated into a "reward function".
  3. ^ Terminology varies; see algorithm characterizations.
  4. ^ Adversarial vulnerabilities can also result in nonlinear systems, or from non-pattern perturbations. Some systems are so brittle that changing a single adversarial pixel predictably induces misclassification.
  5. ^ McCarthy once said: "This is AI, so we don't care if it's psychologically real".[53] McCarthy reiterated his position in 2006 at the AI@50 conference where he said "Artificial intelligence is not, by definition, simulation of human intelligence".[54]. Pamela McCorduck writes that there are "two major branches of artificial intelligence: one aimed at producing intelligent behavior regardless of how it was accomplished, and the other aimed at modeling intelligent processes found in nature, particularly human ones."[55], Stuart Russell and Peter Norvig wrote "Aeronautical engineering texts do not define the goal of their field as making 'machines that fly so exactly like pigeons that they can fool even other pigeons.'"[56]
  6. ^ "There exist many different types of uncertainty, vagueness, and ignorance... [We] independently confirm the inadequacy of systems for reasoning about uncertainty that propagates numerical factors according to only to which connectives appear in assertions."[77]
  7. ^ Each individual neuron is likely to participate in more than one concept.
  8. ^ Steering for the 1995 "No Hands Across America" required "only a few human assists".
  9. ^ In the early 1970s, Kenneth Colby presented a version of Weizenbaum's ELIZA known as DOCTOR which he promoted as a serious therapeutic tool.[163]

Citations

[edit]
  1. ^ Markoff, John (16 February 2011). "Computer Wins on 'Jeopardy!': Trivial, It's Not". The New York Times. Archived from the original on 22 October 2014. Retrieved 25 October 2014.
  2. ^ "AlphaGo – Google DeepMind". Archived from the original on 10 March 2016.
  3. ^ "Artificial intelligence: Google's AlphaGo beats Go master Lee Se-dol". BBC News. 12 March 2016. Archived from the original on 26 August 2016. Retrieved 1 October 2016.
  4. ^ Metz, Cade (27 May 2017). "After Win in China, AlphaGo's Designers Explore New AI". Wired. Archived from the original on 2 June 2017.
  5. ^ "World's Go Player Ratings". May 2017. Archived from the original on 1 April 2017.
  6. ^ "柯洁迎19岁生日 雄踞人类世界排名第一已两年" (in Chinese). May 2017. Archived from the original on 11 August 2017.
  7. ^ "MuZero: Mastering Go, chess, shogi and Atari without rules". Deepmind. Retrieved 2021-03-01.
  8. ^ Steven Borowiec; Tracey Lien (12 March 2016). "AlphaGo beats human Go champ in milestone for artificial intelligence". Los Angeles Times. Retrieved 13 March 2016.
  9. ^ Silver, David; Hubert, Thomas; Schrittwieser, Julian; Antonoglou, Ioannis; Lai, Matthew; Guez, Arthur; Lanctot, Marc; Sifre, Laurent; Kumaran, Dharshan; Graepel, Thore; Lillicrap, Timothy; Simonyan, Karen; Hassabis, Demis (7 December 2018). "A general reinforcement learning algorithm that masters chess, shogi, and go through self-play". Science. 362 (6419): 1140–1144. Bibcode:2018Sci...362.1140S. doi:10.1126/science.aar6404. PMID 30523106.
  10. ^ Schrittwieser, Julian; Antonoglou, Ioannis; Hubert, Thomas; Simonyan, Karen; Sifre, Laurent; Schmitt, Simon; Guez, Arthur; Lockhart, Edward; Hassabis, Demis; Graepel, Thore; Lillicrap, Timothy (2020-12-23). "Mastering Atari, Go, chess and shogi by planning with a learned model". Nature. 588 (7839): 604–609. arXiv:1911.08265. Bibcode:2020Natur.588..604S. doi:10.1038/s41586-020-03051-4. ISSN 1476-4687. PMID 33361790. S2CID 208158225.
  11. ^ Tung, Liam. "Google's DeepMind artificial intelligence aces Atari gaming challenge". ZDNet. Retrieved 2021-03-01.
  12. ^ Solly, Meilan. "This Poker-Playing A.I. Knows When to Hold 'Em and When to Fold 'Em". Smithsonian. Pluribus has bested poker pros in a series of six-player no-limit Texas Hold'em games, reaching a milestone in artificial intelligence research. It is the first bot to beat humans in a complex multiplayer competition.
  13. ^ Bowling, Michael; Burch, Neil; Johanson, Michael; Tammelin, Oskari (2015-01-09). "Heads-up limit hold'em poker is solved". Science. 347 (6218): 145–149. Bibcode:2015Sci...347..145B. doi:10.1126/science.1259433. ISSN 0036-8075. PMID 25574016. S2CID 3796371.
  14. ^ Rowinski, Dan (15 January 2013). "Virtual Personal Assistants & The Future Of Your Smartphone [Infographic]". ReadWrite. Archived from the original on 22 December 2015.
  15. ^ a b Clark 2015b.
  16. ^ Heath, Nick (11 December 2020). "What is AI? Everything you need to know about Artificial Intelligence". ZDNet. Retrieved 1 March 2021.
  17. ^ Fairhead, Harry (26 March 2011) [Update 30 March 2011]. "Kinect's AI breakthrough explained". I Programmer. Archived from the original on 1 February 2016.
  18. ^ Anadiotis, George (1 October 2020). "The state of AI in 2020: Democratization, industrialization, and the way to artificial general intelligence". ZDNet. Retrieved 1 March 2021.
  19. ^ Allen, Gregory (February 6, 2019). "Understanding China's AI Strategy". Center for a New American Security. Archived from the original on 17 March 2019.
  20. ^ "Review | How two AI superpowers – the U.S. and China – battle for supremacy in the field". The Washington Post. 2 November 2018. Archived from the original on 4 November 2018. Retrieved 4 November 2018.
  21. ^ Kaplan, Andreas; Haenlein, Michael (1 January 2019). "Siri, Siri, in my hand: Who's the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence". Business Horizons. 62 (1): 15–25. doi:10.1016/j.bushor.2018.08.004.
  22. ^ Domingos 2015, Chapter 5.
  23. ^ Domingos 2015, Chapter 7.
  24. ^ Lindenbaum, M., Markovitch, S., & Rusakov, D. (2004). Selective sampling for nearest neighbor classifiers. Machine learning, 54(2), 125–152.
  25. ^ Domingos 2015, Chapter 1.
  26. ^ Domingos 2015, Chapter 2, Chapter 3.
  27. ^ Hart, P. E.; Nilsson, N. J.; Raphael, B. (1972). "Correction to "A Formal Basis for the Heuristic Determination of Minimum Cost Paths"". SIGART Newsletter (37): 28–29. doi:10.1145/1056777.1056779. S2CID 6386648.
  28. ^ Domingos 2015, Chapter 2, Chapter 4, Chapter 6.
  29. ^ "Can neural network computers learn from experience, and if so, could they ever become what we would call 'smart'?". Scientific American. 2018. Archived from the original on 25 March 2018. Retrieved 24 March 2018.
  30. ^ a b c Domingos 2015, Chapter 6, Chapter 7.
  31. ^ Domingos 2015, p. 286.
  32. ^ "Single pixel change fools AI programs". BBC News. 3 November 2017. Archived from the original on 22 March 2018. Retrieved 12 March 2018.
  33. ^ "AI Has a Hallucination Problem That's Proving Tough to Fix". WIRED. 2018. Archived from the original on 12 March 2018. Retrieved 12 March 2018.
  34. ^ "Cultivating Common Sense | DiscoverMagazine.com". Discover Magazine. 2017. Archived from the original on 25 March 2018. Retrieved 24 March 2018.
  35. ^ Davis, Ernest; Marcus, Gary (24 August 2015). "Commonsense reasoning and commonsense knowledge in artificial intelligence". Communications of the ACM. 58 (9): 92–103. doi:10.1145/2701413. S2CID 13583137. Archived from the original on 22 August 2020. Retrieved 6 April 2020.
  36. ^ Winograd, Terry (January 1972). "Understanding natural language". Cognitive Psychology. 3 (1): 1–191. doi:10.1016/0010-0285(72)90002-3.
  37. ^ "Don't worry: Autonomous cars aren't coming tomorrow (or next year)". Autoweek. 2016. Archived from the original on 25 March 2018. Retrieved 24 March 2018.
  38. ^ Knight, Will (2017). "Boston may be famous for bad drivers, but it's the testing ground for a smarter self-driving car". MIT Technology Review. Archived from the original on 22 August 2020. Retrieved 27 March 2018.
  39. ^ Prakken, Henry (31 August 2017). "On the problem of making autonomous vehicles conform to traffic law". Artificial Intelligence and Law. 25 (3): 341–363. doi:10.1007/s10506-017-9210-0.
  40. ^ a b Lieto, Antonio; Lebiere, Christian; Oltramari, Alessandro (May 2018). "The knowledge level in cognitive architectures: Current limitations and possible developments". Cognitive Systems Research. 48: 39–55. doi:10.1016/j.cogsys.2017.05.001. hdl:2318/1665207. S2CID 206868967.
  41. ^ Thompson, Derek (2018). "What Jobs Will the Robots Take?". The Atlantic. Archived from the original on 24 April 2018. Retrieved 24 April 2018.
  42. ^ Scassellati, Brian (2002). "Theory of mind for a humanoid robot". Autonomous Robots. 12 (1): 13–24. doi:10.1023/A:1013298507114. S2CID 1979315.
  43. ^ Cao, Yongcan; Yu, Wenwu; Ren, Wei; Chen, Guanrong (February 2013). "An Overview of Recent Progress in the Study of Distributed Multi-Agent Coordination". IEEE Transactions on Industrial Informatics. 9 (1): 427–438. arXiv:1207.3231. doi:10.1109/TII.2012.2219061. S2CID 9588126.
  44. ^ "The superhero of artificial intelligence: can this genius keep it in check?". The Guardian. 16 February 2016. Archived from the original on 23 April 2018. Retrieved 26 April 2018.
  45. ^ Mnih, Volodymyr; Kavukcuoglu, Koray; Silver, David; Rusu, Andrei A.; Veness, Joel; Bellemare, Marc G.; Graves, Alex; Riedmiller, Martin; Fidjeland, Andreas K.; Ostrovski, Georg; Petersen, Stig; Beattie, Charles; Sadik, Amir; Antonoglou, Ioannis; King, Helen; Kumaran, Dharshan; Wierstra, Daan; Legg, Shane; Hassabis, Demis (26 February 2015). "Human-level control through deep reinforcement learning". Nature. 518 (7540): 529–533. Bibcode:2015Natur.518..529M. doi:10.1038/nature14236. PMID 25719670. S2CID 205242740.
  46. ^ Sample, Ian (14 March 2017). "Google's DeepMind makes AI program that can learn like a human". The Guardian. Archived from the original on 26 April 2018. Retrieved 26 April 2018.
  47. ^ "From not working to neural networking". The Economist. 2016. Archived from the original on 31 December 2016. Retrieved 26 April 2018.
  48. ^ Russell & Norvig 2009, Chapter 27. AI: The Present and Future.
  49. ^ & McCorduck 2004, pp. 139–179, 245–250, 322–323 (EPAM).
  50. ^ Crevier 1993, pp. 145–149.
  51. ^ McCorduck 2004, pp. 450–451.
  52. ^ Crevier 1993, pp. 258–263.
  53. ^ Kolata 1982.
  54. ^ Maker 2006.
  55. ^ McCorduck 2004, pp. 100–101.
  56. ^ Russell & Norvig 2003, pp. 2–3.
  57. ^ McCorduck 2004, pp. 251–259.
  58. ^ Crevier 1993, pp. 193–196.
  59. ^ Howe 1994.
  60. ^ McCorduck 2004, pp. 259–305.
  61. ^ Crevier 1993, pp. 83–102, 163–176.
  62. ^ Russell & Norvig 2003, p. 19.
  63. ^ McCorduck 2004, pp. 421–424, 486–489.
  64. ^ Crevier 1993, p. 168.
  65. ^ McCorduck 2004, p. 489.
  66. ^ Crevier 1993, pp. 239–243.
  67. ^ Russell & Norvig 2003, p. 363−365.
  68. ^ McCorduck 2004, pp. 266–276, 298–300, 314, 421.
  69. ^ Russell & Norvig 2003, pp. 22–23.
  70. ^ Weng et al. 2001.
  71. ^ Lungarella et al. 2003.
  72. ^ Asada et al. 2009.
  73. ^ Oudeyer 2010.
  74. ^ Agent architectures, hybrid intelligent systems: * Russell & Norvig (2003, pp. 27, 932, 970–972) * Nilsson (1998, chpt. 25)
  75. ^ Hierarchical control system: * Albus 2002
  76. ^ Lieto, Antonio; Bhatt, Mehul; Oltramari, Alessandro; Vernon, David (May 2018). "The role of cognitive architectures in general artificial intelligence". Cognitive Systems Research. 48: 1–3. doi:10.1016/j.cogsys.2017.08.003. hdl:2318/1665249. S2CID 36189683.
  77. ^ Elkan, Charles (1994). "The paradoxical success of fuzzy logic". IEEE Expert. 9 (4): 3–49. CiteSeerX 10.1.1.100.8402. doi:10.1109/64.336150. S2CID 113687.
  78. ^ Fuzzy logic:
  79. ^ "What is 'fuzzy logic'? Are there computers that are inherently fuzzy and do not apply the usual binary logic?". Scientific American. Retrieved 5 May 2018.
  80. ^ Cite error: The named reference Domingos2005 was invoked but never defined (see the help page).
  81. ^ "Why Deep Learning Is Suddenly Changing Your Life". Fortune. 2016. Retrieved 12 March 2018.
  82. ^ "Google leads in the race to dominate artificial intelligence". The Economist. 2017. Retrieved 12 March 2018.
  83. ^ Seppo Linnainmaa (1970). The representation of the cumulative rounding error of an algorithm as a Taylor expansion of the local rounding errors. Master's Thesis (in Finnish), Univ. Helsinki, 6–7.
  84. ^ Griewank, Andreas (2012). Who Invented the Reverse Mode of Differentiation?. Optimization Stories, Documenta Matematica, Extra Volume ISMP (2012), 389–400.
  85. ^ Paul Werbos, "Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences", PhD thesis, Harvard University, 1974.
  86. ^ Paul Werbos (1982). Applications of advances in nonlinear sensitivity analysis. In System modeling and optimization (pp. 762–770). Springer Berlin Heidelberg. Online Archived 14 April 2016 at the Wayback Machine
  87. ^ Backpropagation:
  88. ^ Hawkins & Blakeslee 2005.
  89. ^ "Artificial intelligence can 'evolve' to solve problems". Science | AAAS. 10 January 2018. Retrieved 7 February 2018.
  90. ^ Schmidhuber (2015b).
  91. ^ Dechter (1986).
  92. ^ Aizenberg, Aizenberg & Vandewalle (2000).
  93. ^ Ivakhnenko (1965).
  94. ^ Ivakhnenko (1971).
  95. ^ Hinton (2007).
  96. ^ Fukushima (1980).
  97. ^ LeCun (2016).
  98. ^ Silver et. al. (2017).
  99. ^ Recurrent neural networks, Hopfield nets:
  100. ^ Hyötyniemi (1996).
  101. ^ Schmidhuber (2015a).
  102. ^ Schmidhuber (1992).
  103. ^ Hochreiter & Schmidhuber (1997).
  104. ^ Graves et al. 2006.
  105. ^ Hannun et. al. (2014); Sak, Senior & Beaufays (2014); Li & Wu (2015)
  106. ^ Sutskever, Vinyals & Le (2014).
  107. ^ Jozefowicz et. al. (2016).
  108. ^ Gillick et al. (2015).
  109. ^ Vinyals et al. (2015).
  110. ^ Wakefield, Jane (15 June 2016). "Social media 'outstrips TV' as news source for young people". BBC News. Archived from the original on 24 June 2016.
  111. ^ Brynjolfsson, Erik; Mitchell, Tom (22 December 2017). "What can machine learning do? Workforce implications". Science. pp. 1530–1534. Bibcode:2017Sci...358.1530B. doi:10.1126/science.aap8062. Retrieved 7 May 2018.
  112. ^ Sample, Ian (18 October 2017). "'It's able to create knowledge itself': Google unveils AI that learns on its own". The Guardian. Retrieved 7 May 2018.
  113. ^ "The AI revolution in science". Science | AAAS. 5 July 2017. Retrieved 7 May 2018.
  114. ^ "Will your job still exist in 10 years when the robots arrive?". South China Morning Post. 2017. Retrieved 7 May 2018.
  115. ^ "IKEA furniture anymd the limits of AI". The Economist. 2018. Retrieved 24 April 2018.
  116. ^ Borowiec, Tracey Lien, Steven (2016). "AlphaGo beats human Go champ in milestone for artificial intelligence". latimes.com. Retrieved 7 May 2018.{{cite news}}: CS1 maint: multiple names: authors list (link)
  117. ^ Brown, Noam; Sandholm, Tuomas (26 January 2018). "Superhuman AI for heads-up no-limit poker: Libratus beats top professionals". Science. pp. 418–424. doi:10.1126/science.aao1733. Retrieved 7 May 2018.
  118. ^ Ontanon, Santiago; Synnaeve, Gabriel; Uriarte, Alberto; Richoux, Florian; Churchill, David; Preuss, Mike (December 2013). "A Survey of Real-Time Strategy Game AI Research and Competition in StarCraft". IEEE Transactions on Computational Intelligence and AI in Games. 5 (4): 293–311. CiteSeerX 10.1.1.406.2524. doi:10.1109/TCIAIG.2013.2286295. S2CID 5014732.
  119. ^ "Facebook Quietly Enters StarCraft War for AI Bots, and Loses". WIRED. 2017. Retrieved 7 May 2018.
  120. ^ "ILSVRC2017". image-net.org. Retrieved 2018-11-06.
  121. ^ Schoenick, Carissa; Clark, Peter; Tafjord, Oyvind; Turney, Peter; Etzioni, Oren (23 August 2017). "Moving beyond the Turing Test with the Allen AI Science Challenge". Communications of the ACM. 60 (9): 60–64. arXiv:1604.04315. doi:10.1145/3122814. S2CID 6383047.
  122. ^ O'Brien, James; Marakas, George (2011). Management Information Systems (10th ed.). McGraw-Hill/Irwin. ISBN 978-0-07-337681-3.
  123. ^ Hernandez-Orallo, Jose (2000). "Beyond the Turing Test". Journal of Logic, Language and Information. 9 (4): 447–466. doi:10.1023/A:1008367325700. S2CID 14481982.
  124. ^ Dowe, D. L.; Hajek, A. R. (1997). "A computational extension to the Turing Test". Proceedings of the 4th Conference of the Australasian Cognitive Science Society. Archived from the original on 28 June 2011.
  125. ^ Hernandez-Orallo, J.; Dowe, D. L. (2010). "Measuring Universal Intelligence: Towards an Anytime Intelligence Test". Artificial Intelligence. 174 (18): 1508–1539. CiteSeerX 10.1.1.295.9079. doi:10.1016/j.artint.2010.09.006.
  126. ^ Hernández-Orallo, José; Dowe, David L.; Hernández-Lloreda, M.Victoria (March 2014). "Universal psychometrics: Measuring cognitive abilities in the machine kingdom". Cognitive Systems Research. 27: 50–74. doi:10.1016/j.cogsys.2013.06.001. hdl:10251/50244. S2CID 26440282.
  127. ^ Research, AI (23 October 2015). "Deep Neural Networks for Acoustic Modeling in Speech Recognition". airesearch.com. Retrieved 23 October 2015.
  128. ^ "GPUs Continue to Dominate the AI Accelerator Market for Now". InformationWeek. December 2019. Retrieved 11 June 2020.
  129. ^ Ray, Tiernan (2019). "AI is changing the entire nature of compute". ZDNet. Retrieved 11 June 2020.
  130. ^ "AI and Compute". OpenAI. 16 May 2018. Retrieved 11 June 2020.
  131. ^ Dartmouth proposal: Historical significance:
  132. ^ Gödel 1951: in this lecture, Kurt Gödel uses the incompleteness theorem to arrive at the following disjunction: (a) the human mind is not a consistent finite machine, or (b) there exist Diophantine equations for which it cannot decide whether solutions exist. Gödel finds (b) implausible, and thus seems to have believed the human mind was not equivalent to a finite machine, i.e., its power exceeded that of any finite machine. He recognized that this was only a conjecture, since one could never disprove (b). Yet he considered the disjunctive conclusion to be a "certain fact".
  133. ^ The Mathematical Objection: * Russell & Norvig 2003, p. 949 * McCorduck 2004, pp. 448–449 Making the Mathematical Objection: * Lucas 1961 * Penrose 1989 Refuting Mathematical Objection: * Turing 1950 under "(2) The Mathematical Objection" * Hofstadter 1979 Background: * Gödel 1931, Church 1936, Kleene 1935, Turing 1937
  134. ^ Graham Oppy (20 January 2015). "Gödel's Incompleteness Theorems". Stanford Encyclopedia of Philosophy. Archived from the original on 22 April 2016. Retrieved 27 April 2016. These Gödelian anti-mechanist arguments are, however, problematic, and there is wide consensus that they fail.
  135. ^ Stuart J. Russell; Peter Norvig (2010). "26.1.2: Philosophical Foundations/Weak AI: Can Machines Act Intelligently?/The mathematical objection". Artificial Intelligence: A Modern Approach (3rd ed.). Upper Saddle River, NJ: Prentice Hall. ISBN 978-0-13-604259-4. even if we grant that computers have limitations on what they can prove, there is no evidence that humans are immune from those limitations.
  136. ^ Mark Colyvan. An introduction to the philosophy of mathematics. Cambridge University Press, 2012. From 2.2.2, 'Philosophical significance of Gödel's incompleteness results': "The accepted wisdom (with which I concur) is that the Lucas-Penrose arguments fail."
  137. ^ Artificial brain arguments: AI requires a simulation of the operation of the human brain * Russell & Norvig 2003, p. 957 * Crevier 1993, pp. 271 & 279 A few of the people who make some form of the argument: * Moravec 1988 * Kurzweil 2005, p. 262 * Hawkins & Blakeslee 2005 The most extreme form of this argument (the brain replacement scenario) was put forward by Clark Glymour in the mid-1970s and was touched on by Zenon Pylyshyn and John Searle in 1980.
  138. ^ Kurzweil 2005.
  139. ^ Simon, Matt (1 April 2019). "Andrew Yang's Presidential Bid Is So Very 21st Century". Wired. Archived from the original on 24 June 2019. Retrieved 2 May 2019 – via www.wired.com.
  140. ^ "Five experts share what scares them the most about AI". 5 September 2018. Archived from the original on 8 December 2019. Retrieved 8 December 2019.
  141. ^ McGaughey 2018.
  142. ^ "Sizing the prize: PwC's Global AI Study – Exploiting the AI Revolution" (PDF). Archived (PDF) from the original on 18 November 2020. Retrieved 2020-11-11.
  143. ^ European Commission 2020, p. 1.
  144. ^ Ford & Colvin 2015.
  145. ^ Rawlinson, Kevin (2015-01-29). "Microsoft's Bill Gates insists AI is a threat". BBC News. Archived from the original on 29 January 2015. Retrieved 30 January 2015.
  146. ^ Holley, Peter (28 January 2015). "Bill Gates on dangers of artificial intelligence: 'I don't understand why some people are not concerned'". The Washington Post. ISSN 0190-8286. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  147. ^ Gibbs, Samuel (2014-10-27). "Elon Musk: artificial intelligence is our biggest existential threat". The Guardian. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  148. ^ Churm, Philip Andrew (2019-05-14). "Yuval Noah Harari talks politics, technology and migration". euronews. Archived from the original on 14 May 2019. Retrieved 2020-11-15.
  149. ^ Cellan-Jones, Rory (2014-12-02). "Stephen Hawking warns artificial intelligence could end mankind". BBC News. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  150. ^ Bostrom, Nick (2015). "What happens when our computers get smarter than we are?". TED (conference). Archived from the original on 25 July 2020. Retrieved 30 January 2020.
  151. ^ Russell 2019, p. 173.
  152. ^ Russell 2019, pp. 191–193.
  153. ^ Post, Washington. "Tech titans like Elon Musk are spending $1 billion to save you from terminators". Archived from the original on 7 June 2016.
  154. ^ "The mysterious artificial intelligence company Elon Musk invested in is developing game-changing smart computers". Tech Insider. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  155. ^ Clark 2015a.
  156. ^ "Elon Musk Is Donating $10M Of His Own Money To Artificial Intelligence Research". Fast Company. 2015-01-15. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  157. ^ Müller, Vincent C.; Bostrom, Nick (2014). "Future Progress in Artificial Intelligence: A Poll Among Experts" (PDF). AI Matters. 1 (1): 9–11. doi:10.1145/2639475.2639478. S2CID 8510016. Archived (PDF) from the original on 15 January 2016.
  158. ^ "Oracle CEO Mark Hurd sees no reason to fear ERP AI". SearchERP. Archived from the original on 6 May 2019. Retrieved 2019-05-06.
  159. ^ "Mark Zuckerberg responds to Elon Musk's paranoia about AI: 'AI is going to... help keep our communities safe.'". Business Insider. 25 May 2018. Archived from the original on 6 May 2019. Retrieved 2019-05-06.
  160. ^ "Is artificial intelligence really an existential threat to humanity?". Bulletin of the Atomic Scientists. 2015-08-09. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  161. ^ "The case against killer robots, from a guy actually working on artificial intelligence". Fusion.net. Archived from the original on 4 February 2016. Retrieved 31 January 2016.
  162. ^ "Will artificial intelligence destroy humanity? Here are 5 reasons not to worry". Vox. 2014-08-22. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  163. ^ Crevier 1993, pp. 132–144.
  164. ^ Joseph Weizenbaum's critique of AI:
  165. ^ "Stephen Hawking, Elon Musk, and Bill Gates Warn About Artificial Intelligence". Observer. 2015-08-19. Archived from the original on 30 October 2015. Retrieved 30 October 2015.
  166. ^ Brooks, Rodney (10 November 2014). "artificial intelligence is a tool, not a threat". Archived from the original on 12 November 2014.
  167. ^ Rubin, Charles (Spring 2003). "Artificial Intelligence and Human Nature". The New Atlantis. 1: 88–100. Archived from the original on 11 June 2012.
  168. ^ Sotala, Kaj; Yampolskiy, Roman V (2014-12-19). "Responses to catastrophic AGI risk: a survey". Physica Scripta. 90 (1): 018001. doi:10.1088/0031-8949/90/1/018001. ISSN 0031-8949.
  169. ^ a b c UNESCO 2021.
  170. ^ "Does This Change Everything? Coronavirus and your private data". European Investment Bank. Archived from the original on 7 June 2021. Retrieved 2021-06-07.
  171. ^ "White Paper on Artificial Intelligence – a European approach to excellence and trust | Shaping Europe's digital future". digital-strategy.ec.europa.eu. Retrieved 2021-06-07.
  172. ^ "What's Ahead for a Cooperative Regulatory Agenda on Artificial Intelligence?". www.csis.org. Archived from the original on 7 June 2021. Retrieved 2021-06-07.
Cite error: A list-defined reference named "Intelligent agents" is not used in the content (see the help page).

Sources

[edit]

.

Unused citations from the article (not needed above)

[edit]

 Not done These citations were not used in the article. Some of these could be "further reading", I suppose.