Artificial Intelligence
Artificial Intelligence
Alan Turing was the first person to conduct substantial research in the field that he
called machine intelligence.[5] Artificial intelligence was founded as an academic
discipline in 1956,[6] by those now considered the founding fathers of AI, John
McCarthy, Marvin Minksy, Nathaniel Rochester, and Claude Shannon.[7][8] The field
went through multiple cycles of optimism,[9][10] followed by periods of disappointment
and loss of funding, known as AI winter.[11][12] Funding and interest vastly increased
after 2012 when deep learning surpassed all previous AI techniques,[13] and after
2017 with the transformer architecture.[14] This led to the AI boom of the early 2020s,
with companies, universities, and laboratories overwhelmingly based in the United
States pioneering significant advances in artificial intelligence.[15]
The growing use of artificial intelligence in the 21st century is influencing a societal
and economic shift towards increased automation, data-driven decision-making, and
the integration of AI systems into various economic sectors and areas of
life, impacting job
markets, healthcare, government, industry, education, propaganda,
and disinformation. This raises questions about the long-term effects, ethical
implications, and risks of AI, prompting discussions about regulatory policies to
ensure the safety and benefits of the technology.
The various subfields of AI research are centered around particular goals and the
use of particular tools. The traditional goals of AI research
include reasoning, knowledge representation, planning, learning, natural language
processing, perception, and support for robotics.[a] General intelligence—the ability to
complete any task performable by a human on an at least equal level—is among the
field's long-term goals.[16]
To reach these goals, AI researchers have adapted and integrated a wide range of
techniques, including search and mathematical optimization, formal logic, artificial
neural networks, and methods based on statistics, operations research,
and economics.[b] AI also draws
upon psychology, linguistics, philosophy, neuroscience, and other fields.[17]
Goals
The general problem of simulating (or creating) intelligence has been broken into
subproblems. These consist of particular traits or capabilities that researchers expect
an intelligent system to display. The traits described below have received the most
attention and cover the scope of AI research.[a]
Many of these algorithms are insufficient for solving large reasoning problems
because they experience a "combinatorial explosion": They become exponentially
slower as the problems grow.[20] Even humans rarely use the step-by-step deduction
that early AI research could model. They solve most of their problems using fast,
intuitive judgments.[21] Accurate and efficient reasoning is an unsolved problem.
Knowledge representation
In classical planning, the agent knows exactly what the effect of any action will be.
[40]
In most real-world problems, however, the agent may not be certain about the
situation they are in (it is "unknown" or "unobservable") and it may not know for
certain what will happen after each possible action (it is not "deterministic"). It must
choose an action by making a probabilistic guess and then reassess the situation to
see if the action worked.[41]
In some problems, the agent's preferences may be uncertain, especially if there are
other agents or humans involved. These can be learned (e.g., with inverse
reinforcement learning), or the agent can seek information to improve its
preferences.[42] Information value theory can be used to weigh the value of
exploratory or experimental actions.[43] The space of possible future actions and
situations is typically intractably large, so the agents must take actions and evaluate
situations while being uncertain of what the outcome will be.
A Markov decision process has a transition model that describes the probability that
a particular action will change the state in a particular way and a reward function that
supplies the utility of each state and the cost of each action. A policy associates a
decision with each possible state. The policy could be calculated (e.g., by iteration),
be heuristic, or it can be learned.[44]
Game theory describes the rational behavior of multiple interacting agents and is
used in AI programs that make decisions that involve other agents.[45]
Learning
Machine learning is the study of programs that can improve their performance on a
given task automatically.[46] It has been a part of AI from the beginning.[e]
In reinforcement learning, the agent is rewarded for good responses and punished
for bad ones. The agent learns to choose responses that are classified as "good".
[51]
Transfer learning is when the knowledge gained from one problem is applied to a
new problem.[52] Deep learning is a type of machine learning that runs inputs through
biologically inspired artificial neural networks for all of these types of learning.[53]
Early work, based on Noam Chomsky's generative grammar and semantic networks,
had difficulty with word-sense disambiguation[f] unless restricted to small domains
called "micro-worlds" (due to the common sense knowledge problem[34]). Margaret
Masterman believed that it was meaning and not grammar that was the key to
understanding languages, and that thesauri and not dictionaries should be the basis
of computational language structure.
Modern deep learning techniques for NLP include word embedding (representing
words, typically as vectors encoding their meaning),[57] transformers (a deep learning
architecture using an attention mechanism),[58] and others.[59] In 2019, generative pre-
trained transformer (or "GPT") language models began to generate coherent text,[60]
[61]
and by 2023 these models were able to get human-level scores on the bar
exam, SAT test, GRE test, and many other real-world applications.[62]
Perception
Machine perception is the ability to use input from sensors (such as cameras,
microphones, wireless signals, active lidar, sonar, radar, and tactile sensors) to
deduce aspects of the world. Computer vision is the ability to analyze visual input.[63]
Social intelligence
Kismet, a robot head which was made in the 1990s; a
machine that can recognize and simulate emotions[69]
Affective computing is an interdisciplinary umbrella that comprises systems that
recognize, interpret, process, or simulate human feeling, emotion, and mood.[70] For
example, some virtual assistants are programmed to speak conversationally or even
to banter humorously; it makes them appear more sensitive to the emotional
dynamics of human interaction, or to otherwise facilitate human–computer
interaction.
However, this tends to give naïve users an unrealistic conception of the intelligence
of existing computer agents.[71] Moderate successes related to affective computing
include textual sentiment analysis and, more recently, multimodal sentiment
analysis, wherein AI classifies the affects displayed by a videotaped subject.[72]
General intelligence
A machine with artificial general intelligence should be able to solve a wide variety of
problems with breadth and versatility similar to human intelligence.[16]
Techniques
AI research uses a wide variety of techniques to accomplish the goals above.[b]
Simple exhaustive searches[76] are rarely sufficient for most real-world problems:
the search space (the number of places to search) quickly grows to astronomical
numbers. The result is a search that is too slow or never completes.[20] "Heuristics" or
"rules of thumb" can help prioritize choices that are more likely to reach a goal. [77]
Distributed search processes can coordinate via swarm intelligence algorithms. Two
popular swarm algorithms used in search are particle swarm optimization (inspired
by bird flocking) and ant colony optimization (inspired by ant trails).[82]
Logic
Formal logic is used for reasoning and knowledge representation.[83] Formal logic
comes in two main forms: propositional logic (which operates on statements that are
true or false and uses logical connectives such as "and", "or", "not" and "implies")
[84]
and predicate logic (which also operates on objects, predicates and relations and
uses quantifiers such as "Every X is a Y" and "There are some Xs that are Ys").[85]
Fuzzy logic assigns a "degree of truth" between 0 and 1. It can therefore handle
propositions that are vague and partially true.[90]
Bayesian networks[97] are a tool that can be used for reasoning (using the Bayesian
inference algorithm),[g][99] learning (using the expectation–maximization algorithm),[h]
[101]
planning (using decision networks)[102] and perception (using dynamic Bayesian
networks).[95]
Probabilistic algorithms can also be used for filtering, prediction, smoothing, and
finding explanations for streams of data, thus helping perception systems analyze
processes that occur over time (e.g., hidden Markov models or Kalman filters).[95]
Expectation–maximization clustering of Old
Faithful eruption data starts from a random guess but then successfully converges on an
accurate clustering of the two physically distinct modes of eruption
Classifiers and statistical learning methods
The simplest AI applications can be divided into two types: classifiers (e.g., "if shiny
then diamond"), on one hand, and controllers (e.g., "if diamond then pick up"), on the
other hand. Classifiers[103] are functions that use pattern matching to determine the
closest match. They can be fine-tuned based on chosen examples using supervised
learning. Each pattern (also called an "observation") is labeled with a certain
predefined class. All the observations combined with their class labels are known as
a data set. When a new observation is received, that observation is classified based
on previous experience.[50]
There are many kinds of classifiers in use. The decision tree is the simplest and
most widely used symbolic machine learning algorithm.[104] K-nearest
neighbor algorithm was the most widely used analogical AI until the mid-1990s,
and Kernel methods such as the support vector machine (SVM) displaced k-nearest
neighbor in the 1990s.[105] The naive Bayes classifier is reportedly the "most widely
used learner"[106] at Google, due in part to its scalability.[107] Neural networks are also
used as classifiers.[108]
Learning algorithms for neural networks use local search to choose the weights that
will get the right output for each input during training. The most common training
technique is the backpropagation algorithm.[109] Neural networks learn to model
complex relationships between inputs and outputs and find patterns in data. In
theory, a neural network can learn any function.[110]
In feedforward neural networks the signal passes in only one direction.[111] Recurrent
neural networks feed the output signal back into the input, which allows short-term
memories of previous input events. Long short term memory is the most successful
network architecture for recurrent networks.[112] Perceptrons[113] use only a single layer
of neurons, deep learning[114] uses multiple layers. Convolutional neural
networks strengthen the connection between neurons that are "close" to each other
—this is especially important in image processing, where a local set of neurons
must identify an "edge" before the network can identify an object.[115]
Deep learning
Deep learning[114] uses several layers of neurons between the network's inputs and
outputs. The multiple layers can progressively extract higher-level features from the
raw input. For example, in image processing, lower layers may identify edges, while
higher layers may identify the concepts relevant to a human such as digits, letters, or
faces.[116]
GPT
Generative pre-trained transformers (GPT) are large language models that are
based on the semantic relationships between words in sentences (natural language
processing). Text-based GPT models are pre-trained on a large corpus of text which
can be from the internet. The pre-training consists in predicting the next token (a
token being usually a word, subword, or punctuation). Throughout this pre-training,
GPT models accumulate knowledge about the world, and can then generate human-
like text by repeatedly predicting the next token. Typically, a subsequent training
phase makes the model more truthful, useful and harmless, usually with a technique
called reinforcement learning from human feedback (RLHF). Current GPT models
are still prone to generating falsehoods called "hallucinations", although this can be
reduced with RLHF and quality data. They are used in chatbots, which allow you to
ask a question or request a task in simple text.[127][128]
Applications
Main article: Applications of artificial intelligence
AI and machine learning technology is used in most of the essential applications of
the 2020s, including: search engines (such as Google Search), targeting online
advertisements, recommendation systems (offered by Netflix, YouTube or Amazon),
driving internet traffic, targeted advertising (AdSense, Facebook), virtual
assistants (such as Siri or Alexa), autonomous
vehicles (including drones, ADAS and self-driving cars), automatic language
translation (Microsoft Translator, Google Translate), facial recognition (Apple's Face
ID or Microsoft's DeepFace and Google's FaceNet) and image labeling (used
by Facebook, Apple's iPhoto and TikTok). The deployment of AI may be overseen by
a Chief automation officer (CAO).
Games
Main article: Game artificial intelligence
Game playing programs have been used since the 1950s to demonstrate and test
AI's most advanced techniques.[140] Deep Blue became the first computer chess-
playing system to beat a reigning world chess champion, Garry Kasparov, on 11 May
1997.[141] In 2011, in a Jeopardy! quiz show exhibition match, IBM's question
answering system, Watson, defeated the two greatest Jeopardy! champions, Brad
Rutter and Ken Jennings, by a significant margin.[142] In March 2016, AlphaGo won 4
out of 5 games of Go in a match with Go champion Lee Sedol, becoming the
first computer Go-playing system to beat a professional Go player
without handicaps. Then in 2017 it defeated Ke Jie, who was the best Go player in
the world.[143] Other programs handle imperfect-information games, such as
the poker-playing program Pluribus.[144] DeepMind developed increasingly
generalistic reinforcement learning models, such as with MuZero, which could be
trained to play chess, Go, or Atari games.[145] In 2019, DeepMind's AlphaStar
achieved grandmaster level in StarCraft II, a particularly challenging real-time
strategy game that involves incomplete knowledge of what happens on the map.
[146]
In 2021, an AI agent competed in a PlayStation Gran Turismo competition,
winning against four of the world's best Gran Turismo drivers using deep
reinforcement learning.[147]
Finance
Finance is one of the fastest growing sectors where applied AI tools are being
deployed: from retail online banking to investment advice and insurance, where
automated "robot advisers" have been in use for some years. [148]
World Pensions experts like Nicolas Firzli insist it may be too early to see the
emergence of highly innovative AI-informed financial products and services: "the
deployment of AI tools will simply further automatise things: destroying tens of
thousands of jobs in banking, financial planning, and pension advice in the process,
but I’m not sure it will unleash a new wave of [e.g., sophisticated] pension
innovation."[149]
Military
Main article: Military artificial intelligence
Various countries are deploying AI military applications.[150] The main applications
enhance command and control, communications, sensors, integration and
interoperability.[151] Research is targeting intelligence collection and analysis, logistics,
cyber operations, information operations, and semiautonomous and autonomous
vehicles.[150] AI technologies enable coordination of sensors and effectors, threat
detection and identification, marking of enemy positions, target acquisition,
coordination and deconfliction of distributed Joint Fires between networked combat
vehicles involving manned and unmanned teams.[151] AI was incorporated into military
operations in Iraq and Syria.[150]
Generative AI
Main article: Generative artificial intelligence
Ethics
Main article: Ethics of artificial intelligence
AI has potential benefits and potential risks. AI may be able to advance science and
find solutions for serious problems: Demis Hassabis of Deep Mind hopes to "solve
intelligence, and then use that to solve everything else".[157] However, as the use of AI
has become widespread, several unintended consequences and risks have been
identified.[158] In-production systems can sometimes not factor ethics and bias into
their AI training processes, especially when the AI algorithms are inherently
unexplainable in deep learning.[159]
Technology companies collect a wide range of data from their users, including online
activity, geolocation data, video and audio.[160] For example, in order to build speech
recognition algorithms, Amazon has recorded millions of private conversations and
allowed temporary workers to listen to and transcribe some of them.[161] Opinions
about this widespread surveillance range from those who see it as a necessary
evil to those for whom it is clearly unethical and a violation of the right to privacy.[162]
AI developers argue that this is the only way to deliver valuable applications. and
have developed several techniques that attempt to preserve privacy while still
obtaining the data, such as data aggregation, de-identification and differential
privacy.[163] Since 2016, some privacy experts, such as Cynthia Dwork, have begun to
view privacy in terms of fairness. Brian Christian wrote that experts have pivoted
"from the question of 'what they know' to the question of 'what they're doing with
it'."[164]
Misinformation
See also: YouTube § Moderation and offensive content
YouTube, Facebook and others use recommender systems to guide users to more
content. These AI programs were given the goal of maximizing user engagement
(that is, the only goal was to keep people watching). The AI learned that users
tended to choose misinformation, conspiracy theories, and extreme partisan content,
and, to keep them watching, the AI recommended more of it. Users also tended to
watch more content on the same subject, so the AI led people into filter
bubbles where they received multiple versions of the same misinformation.[171] This
convinced many users that the misinformation was true, and ultimately undermined
trust in institutions, the media and the government.[172] The AI program had correctly
learned to maximize its goal, but the result was harmful to society. After the U.S.
election in 2016, major technology companies took steps to mitigate the problem [citation
needed]
.
In 2022, generative AI began to create images, audio, video and text that are
indistinguishable from real photographs, recordings, films, or human writing. It is
possible for bad actors to use this technology to create massive amounts of
misinformation or propaganda.[173] AI pioneer Geoffrey Hinton expressed concern
about AI enabling "authoritarian leaders to manipulate their electorates" on a large
scale, among other risks.[174]
On June 28, 2015, Google Photos's new image labeling feature mistakenly identified
Jacky Alcine and a friend as "gorillas" because they were black. The system was
trained on a dataset that contained very few images of black people,[180] a problem
called "sample size disparity".[181] Google "fixed" this problem by preventing the
system from labelling anything as a "gorilla". Eight years later, in 2023, Google
Photos still could not identify a gorilla, and neither could similar products from Apple,
Facebook, Microsoft and Amazon.[182]
A program can make biased decisions even if the data does not explicitly mention a
problematic feature (such as "race" or "gender"). The feature will correlate with other
features (like "address", "shopping history" or "first name"), and the program will
make the same decisions based on these features as it would on "race" or "gender".
[186]
Moritz Hardt said "the most robust fact in this research area is that fairness
through blindness doesn't work."[187]
Bias and unfairness may go undetected because the developers are overwhelmingly
white and male: among AI engineers, about 4% are black and 20% are women.[181]
Lack of transparency
See also: Explainable AI, Algorithmic transparency, and Right to explanation
Many AI systems are so complex that their designers cannot explain how they reach
their decisions.[191] Particularly with deep neural networks, in which there are a large
amount of non-linear relationships between inputs and outputs. But some popular
explainability techniques exist.[192]
There are several possible solutions to the transparency problem. SHAP tried to
solve the transparency problems by visualising the contribution of each feature to the
output.[198] LIME can locally approximate a model with a simpler, interpretable model.
[199]
Multitask learning provides a large number of outputs in addition to the target
classification. These other outputs can help developers deduce what the network
has learned.[200] Deconvolution, DeepDream and other generative methods can allow
developers to see what different layers of a deep network have learned and produce
output that can suggest what the network is learning.[201]
A lethal autonomous weapon is a machine that locates, selects and engages human
targets without human supervision.[n] Widely available AI tools can be used by bad
actors to develop inexpensive autonomous weapons and, if produced at scale, they
are potentially weapons of mass destruction.[203] Even when used in conventional
warfare, it is unlikely that they will be unable to reliably choose targets and could
potentially kill an innocent person.[203] In 2014, 30 nations (including China) supported
a ban on autonomous weapons under the United Nations' Convention on Certain
Conventional Weapons, however the United States and others disagreed.[204] By
2015, over fifty countries were reported to be researching battlefield robots.[205]
There many other ways that AI is expected to help bad actors, some of which can
not be foreseen. For example, machine-learning AI is able to design tens of
thousands of toxic molecules in a matter of hours.[209]
Technological unemployment
Main articles: Workplace impact of artificial intelligence and Technological
unemployment
Economists have frequently highlighted the risks of redundancies from AI, and
speculated about unemployment if there is no adequate social policy for full
employment.[211]
In the past, technology has tended to increase rather than reduce total employment,
but economists acknowledge that "we're in uncharted territory" with AI.[212] A survey of
economists showed disagreement about whether the increasing use of robots and AI
will cause a substantial increase in long-term unemployment, but they generally
agree that it could be a net benefit if productivity gains are redistributed.[213] Risk
estimates vary; for example, in the 2010s, Michael Osborne and Carl Benedikt
Frey estimated 47% of U.S. jobs are at "high risk" of potential automation, while an
OECD report classified only 9% of U.S. jobs as "high risk".[o][215] The methodology of
speculating about future employment levels has been criticised as lacking evidential
foundation, and for implying that technology, rather than social policy, creates
unemployment, as opposed to redundancies.[211] In April 2023, it was reported that
70% of the jobs for Chinese video game illustrators had been eliminated by
generative artificial intelligence.[216][217]
From the early days of the development of artificial intelligence, there have been
arguments, for example, those put forward by Joseph Weizenbaum, about whether
tasks that can be done by computers actually should be done by them, given the
difference between computers and humans, and between quantitative calculation
and qualitative, value-based judgement.[220]
Existential risk
Main article: Existential risk from artificial general intelligence
It has been argued AI will become so powerful that humanity may irreversibly lose
control of it. This could, as physicist Stephen Hawking stated, "spell the end of the
human race".[221] This scenario has been common in science fiction, when a computer
or robot suddenly develops a human-like "self-awareness" (or "sentience" or
"consciousness") and becomes a malevolent character.[p] These sci-fi scenarios are
misleading in several ways.
Second, Yuval Noah Harari argues that AI does not require a robot body or physical
control to pose an existential risk. The essential parts of civilization are not physical.
Things like ideologies, law, government, money and the economy are made
of language; they exist because there are stories that billions of people believe. The
current prevalence of misinformation suggests that an AI could use language to
convince people to believe anything, even to take actions that are destructive.[226]
The opinions amongst experts and industry insiders are mixed, with sizable fractions
both concerned and unconcerned by risk from eventual superintelligent AI.
[227]
Personalities such as Stephen Hawking, Bill Gates, and Elon Musk,[228] as well as
AI pioneers such as Geoffrey Hinton, Yoshua Bengio, Demis Hassabis, and Sam
Altman, have expressed concerns about existential risk from AI.
In 2023, many leading AI experts issued the joint statement that "Mitigating the risk
of extinction from AI should be a global priority alongside other societal-scale risks
such as pandemics and nuclear war".[229]
Machines with intelligence have the potential to use their intelligence to make ethical
decisions. The field of machine ethics provides machines with ethical principles and
procedures for resolving ethical dilemmas.[238] The field of machine ethics is also
called computational morality,[238] and was founded at an AAAI symposium in 2005.[239]
Other approaches include Wendell Wallach's "artificial moral agents"[240] and Stuart J.
Russell's three principles for developing provably beneficial machines.[241]
Open source
Active organizations in the AI open-source community include Hugging Face,
[242]
Google,[243] EleutherAI and Meta.[244] Various AI models, such as Llama
2, Mistral or Stable Diffusion, have been made open-weight,[245][246] meaning that their
architecture and trained parameters (the "weights") are publicly available. Open-
weight models can be freely fine-tuned, which allows companies to specialize them
with their own data and for their own use-case.[247] Open-weight models are useful for
research and innovation but can also be misused. Since they can be fine-tuned, any
built-in security measure, such as objecting to harmful requests, can be trained away
until it becomes ineffective. Some researchers warn that future AI models may
develop dangerous capabilities (such as the potential to drastically
facilitate bioterrorism), and that once released on the Internet, they can't be deleted
everywhere if needed. They recommend pre-release audits and cost-benefit
analyses.[248]
Frameworks
Artificial Intelligence projects can have their ethical permissibility tested while
designing, developing, and implementing an AI system. An AI framework such as the
Care and Act Framework containing the SUM values—developed by the Alan Turing
Institute tests projects in four main areas:[249][250]
Promotion of the wellbeing of the people and communities that these technologies
affect requires consideration of the social and ethical implications at all stages of AI
system design, development and implementation, and collaboration between job
roles such as data scientists, product managers, data engineers, domain experts,
and delivery managers.[253]
The AI Safety Institute in the UK has released a testing toolset called 'Inspect' for AI
safety evaluations available under a MIT open-source licence which is freely
available on GitHub and can be improved with third-party packages. It can be used
to evaluate AI models in a range of areas including core knowledge, ability to reason,
and autonomous capabilities.[254]
Voicing concerns about issues including the existential risk from artificial general
intelligence,[255] in May 2023, Geoffrey Hinton announced his resignation from Google
in order to be able to "freely speak out about the risks of A.I." [256] He stressed that in
order to avoid the worst outcomes, establishing safety guidelines will require
cooperation among those competing in use of A.I.[257]