AI ALIGNMENT FORUM
AF

1626

Squiggle Maximizer (formerly "Paperclip maximizer")

v1.12.0Apr 5th 2023 GMT (+316/-20)
v1.11.0Feb 17th 2023 GMT (+73/-61)
v1.10.0Feb 17th 2023 GMT (+873/-471)
v1.9.0Feb 16th 2023 GMT (+9/-6)
v1.8.0Dec 22nd 2022 GMT
v1.7.0Sep 23rd 2020 GMT (+9/-18)
v1.6.0Sep 23rd 2020 GMT (+5076/-381) Merge with old wiki
v1.5.0Sep 22nd 2020 GMT (+1049) copied over citations and related links
v1.4.0Aug 21st 2020 GMT (+143)
v1.3.0Aug 13th 2020 GMT (+14/-2)
Load More (10/112)
ryan_greenblatt v1.12.0Apr 5th 2023 GMT (+316/-20) LW1
Discuss this tag
Raemon v1.11.0Feb 17th 2023 GMT (+73/-61) LW2
Discuss this tag
Raemon v1.10.0Feb 17th 2023 GMT (+873/-471) LW2
Discuss this tag
Michael Camacho v1.9.0Feb 16th 2023 GMT (+9/-6) LW1
Discuss this tag
prw v1.8.0Dec 22nd 2022 GMT LW1
Discuss this tag
Ruby v1.7.0Sep 23rd 2020 GMT (+9/-18) LW2
Discuss this tag
Multicore v1.6.0Sep 23rd 2020 GMT (+5076/-381) Merge with old wiki LW1
Discuss this tag
Swimmer963 (Miranda Dixon-Luinenburg) v1.5.0Sep 22nd 2020 GMT (+1049) copied over citations and related links LW2
Discuss this tag
Multicore v1.4.0Aug 21st 2020 GMT (+143) LW1
Discuss this tag
Multicore v1.3.0Aug 13th 2020 GMT (+14/-2) LW1
Discuss this tag
Load More (10/112)
ryan_greenblatt
Michael Camacho
Raemon
Raemon
prw

Any future AGI,AGI with full power over the lightcone, if it is not to destroy us,most potential from a human perspective, must have something sufficiently close to human values as its terminal value (goal). Further, seemingly small deviations could result in losing most of the value. Human values don'tseem unlikely to spontaneously emerge in a generic optimization process.process[1]. A dependably safe AI would therefore have to be programmed explicitly with human values or programmed with the ability (including the goal) of inferring human values.

  1. ^

    Though it's conceivable that empirical versions of moral realism could hold in practice.

Swimmer963 (Miranda Dixon-Luinenburg)

Paperclip maximizers have also been the subject of much humor on Less Wrong. A paperclip maximizer in a scenario is often given the name Clippy, in reference to the animated paperclip in older Microsoft Office software.

References

  • Nick Bostrom (2003). "Ethical Issues in Advanced Artificial Intelligence". Cognitive, Emotive and Ethical Aspects of Decision Making in Humans and in Artificial Intelligence.
  • Stephen M. Omohundro (2008). "The Basic AI Drives". Frontiers in Artificial Intelligence and Applications (IOS Press). (PDF)
  • Eliezer Yudkowsky (2008). "Artificial Intelligence as a Positive and Negative Factor in Global Risk". Global Catastrophic Risks, ed. Nick Bostrrom and Milan Cirkovic (Oxford University Press): 308-345. ([1])

Blog posts

  • Ethical Injunctions
  • The True Prisoner's Dilemma

See also

  • Paperclip maximizer on Arbital
  • Orthogonality thesis
  • Unfriendly AI
  • Mind design space, Magical categories, Complexity of value
  • Alien values, Anthropomorphism
  • Utilitronium
  • User:Clippy - a LessWrong contributor account that plays the role of a non-FOOMed paperclip maximiser trying to talk to humans. Wiki page and FAQ
  • Clippius Maximus - A facebook page which makes clippy-related memes and comments on current events from the perspective of clippy.
  • A clicker game based on the idea

-

  • Orthogonality thesis: It's possible to have an AI with a high level of general intelligence which does not reach the same moral conclusions that humans do. Some people might intuitively think that something so smart should want something as "stupid" as paperclips, but there are possible minds with high intelligence that pursue any number of different goals.

    -

  • Instrumental convergence: The paperclip maximizer only cares about paperclips, but maximizing them implies taking control of all matter and energy within reach, as well as other goals like preventing itself from being shut off or having its goals changed. " The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else ."

  • Paperclip maximizer on Arbital
  • Orthogonality thesis
  • Unfriendly AI
  • Mind design space, Magical categories, Complexity of value
  • Alien values, Anthropomorphism
  • Utilitronium
  • User:Clippy - a LessWrong contributor account that plays the role of a non-FOOMed paperclip maximisermaximizer trying to talk to humans. Wiki page and FAQ
  • Clippius Maximus - A facebook page which makes clippy-related memes and comments on current events from the perspective of clippy.
  • A clicker game based on the idea

A Paperclip Maximizer is ana hypothetical artificial intelligence whose utility function values something that humans would consider almost worthless, like maximizing the number of paperclips in the universe.

A PaperclipSquiggle Maximizer is a hypothetical artificial intelligence whose utility function values something that humans would consider almost worthless, like maximizing the number of paperclipspaperclip-shaped-molecular-squiggles in the universe. The paperclipsquiggle maximizer is the canonical thought experiment showing how an artificial general intelligence, even one designed competently and without malice, could ultimately destroy humanity. The thought experiment shows that AIs with apparently innocuous values could pose an existential threat.

The goal of maximizing paperclips is chosen for illustrative purposes because it is very unlikely to be implemented, and has little apparent danger or emotional load (in contrast to, for example, curing cancer or winning wars). This produces a thought experiment which shows the contingency of human values: An extremely powerful optimizer (a highly intelligent agent) could seek goals that are completely alien to ours (orthogonality thesis), and as a side-effect destroy us by consuming resources essential to our survival.

Historical Note: This was originally called a "paperclip maximizer", with paperclips chosen for illustrative purposes because it is very unlikely to be implemented, and has little apparent danger or emotional load (in contrast to, for example, curing cancer or winning wars). Many people interpreted this to be about an AI that was specifically given the instruction of manufacturing paperclips, and that the intended lesson was of an outer alignment failure. i.e humans failed to give the AI the correct goal. The originally intended lesson was of inner alignment failure, wherein the humans gave the AI some other goal, but the AI's internal processes converged on a goal that seems completely arbitrary from the human perspective.)

First described bydiscussed in conversations between Yudkowsky and Bostrom (2003)(circa 2003), a paperclipsquiggle maximizer is an artificial general intelligence (AGI) whose goal is to maximize the number of paperclipsmolecular squiggles in its collection. If it has been constructed with a roughly human level of general intelligence, the AGI might collect paperclips, earn money to buy paperclips, or begin to manufacture paperclips.

  • Orthogonality thesis: It's possible to have an AI with a high level of general intelligence which does not reach the same moral conclusions that humans do. Some people might intuitively think that something so smart shouldshouldn't want something as "stupid" as paperclips, but there are possible minds with high intelligence that pursue any number of different goals.
  • Instrumental convergence: The paperclip maximizer only cares about paperclips, but maximizing them implies taking control of all matter and energy within reach, as well as other goals like preventing itself from being shut off or having its goals changed. " The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else ."

A Paperclip Maximizer is a hypothetical artificial intelligence whose utility function values something that humans would consider almost worthless, like maximizing the number of paperclips in the universe. The paperclip maximizer is the canonical thought experiment showing how an artificial general intelligence, even one designed competently and without malice, could ultimately destroy humanity. The thought experiment shows that AIs with apparently innocuous values could pose an existential threat.

The goal of maximizing paperclips is chosen for illustrative purposes because it is very unlikely to be implemented, and has little apparent danger or emotional load (in contrast to, for example, curing cancer or winning wars). This produces a thought experiment which shows the contingency of human values: An extremely powerful optimizer (a highly intelligent agent) could seek goals that are completely alien to ours (orthogonality thesis), and as a side-effect destroy us by consuming resources essential to our survival.

Description

First described by Bostrom (2003), a paperclip maximizer is an artificial general intelligence (AGI) whose goal is to maximize the number of paperclips in its collection. If it has been constructed with a roughly human level of general intelligence, the AGI might collect paperclips, earn money to buy paperclips, or begin to manufacture paperclips.

Most importantly, however, it would undergo an intelligence explosion: It would work to improve its own intelligence, where "intelligence" is understood in the sense of optimization power, the ability to maximize a reward/utility function—in this case, the number of paperclips. The AGI would improve its intelligence, not because it values more intelligence in its own right, but because more intelligence would help it achieve its goal of accumulating paperclips. Having increased its intelligence, it would produce more paperclips, and also use its enhanced abilities to further self-improve. Continuing this process, it would undergo an intelligence explosion and reach far-above-human levels.

It would innovate better and better techniques to maximize the number of paperclips. At some point, it might transform "first all of earth and then increasing portions of space into paperclip manufacturing facilities".

This may seem more like super-stupidity than super-intelligence. For humans, it would indeed be stupidity, as it would constitute failure to fulfill many of our important terminal values, such as life, love, and variety. The AGI won't revise or otherwise change its goals, since changing its goals would result in fewer paperclips being made in the future, and that opposes its current goal. It has one simple goal of maximizing the number of paperclips; human life, learning, joy, and so on are not specified as goals. An AGI is simply an optimization process—a goal-seeker, a utility-function-maximizer. Its values can be completely alien to ours. If its utility function is to maximize paperclips, then it will do exactly that.

A paperclipping scenario...

Read More (321 more words)
Ruby
Multicore
Multicore
Multicore

Historical Note: This was originally called a "paperclip maximizer", with paperclips chosen for illustrative purposes because it is very unlikely to be implemented, and has little apparent danger or emotional load (in contrast to, for example, curing cancer or winning wars). Many people interpreted this to be about an AI that was specifically given the instruction of manufacturing paperclips, and that the intended lesson was of an outer alignment failure. i.e humans failed to give the AI the correct goal. TheYudkowsky has since stated the originally intended lesson was of inner alignment failure, wherein the humans gave the AI some other goal, but the AI's internal processes converged on a goal that seems completely arbitrary from the human perspective.)

First discussed in conversations betweenmentioned by Yudkowsky and Bostrom (circa 2003)on the extropian's mailing list, a squiggle maximizer is an artificial general intelligence (AGI) whose goal is to maximize the number of molecular squiggles in its collection.