lecture10-weak

Download Report

Transcript lecture10-weak

Strong vs. Weak AI

ECE 847: Digital Image Processing Stan Birchfield Clemson University

The coming takeover

Common theme: - Robots become intelligent - Robots become independent - Robots get out of control - Robots must be subdued Why all the fuss? Need we fear?

The central question

• • • •

This is not just for entertainment It has far-reaching implications Two camps:

strong AI: There is no fundamental difference between man and machine

weak AI: Only people can think, machines cannot Central question: Is there a fundamental difference between man and machine, or is it only a difference in computing power?

The question restated

Stated another way,

Can computers think?

What does it mean to think?

In favor of “Strong AI”

Strong AI argument #1: Look at what machines can do

clean (Roomba vaccuum cleaner) play soccer (Robocup)

organ (Ichiro Kato’s WABOT II reads music and plays the organ 1984)

They play music

trumpet (Toyota’s trumpet-playing robot 2008) flute (Atsuo Takanishi’s flute-playing robot) conductor (Honda’s Asimo robot conducts Detroit Symphony Orchestra 2008)

They even compose music

MySong: http://research.microsoft.com/~dan/mysong/

… and have emotions

Rity sobot (from Kim Jong-Hwan at Korea's Institute of Advanced Science and Technology)

In favor of “Strong AI”

Strong AI argument #2: Look at what people said machines will never do

Hubert Dreyfus, Berkeley philosopher: No computer will ever beat me at chess 1967: Richard Greenblatt's MacHack’s program beat him Then Dreyfus: Well, no computer will beat a nationally ranked player But it did Then Dreyfus: Well, no computer will beat a world champion 1997: Deep Blue beat Garry Kasparov Kasparov vs. Deep Blue

The clincher

• •

Many people are fond of saying, “They will never make a machine to replace the human mind --- it does many things which no machine could ever do.” J. von Neumann gave talk in Princeton (1948)

Question from audience: “But of course, a mere machine can't really think, can it?”

Answer: “You insist that there is something a machine cannot do. If you will tell me precisely what it is that a machine cannot do, then I can always make a machine which will do just that!'' [from E. T. Jaynes, Probability Theory: The Logic of Science]

“Weak AI” responses

• • • • •

Hubert Dreyfus, Berkeley philosopher:

Nonformal aspects of thinking cannot be reduced to mechanized processes

John Searle, Berkeley philosopher:

Chinese room experiment – blindly translating is not the same as thinking

Thomas Ray, Oklahoma zoologist:

carbon medium and silicon medium are fundamentally different

Roger Penrose, British physicist and mathematician:

Consciousness arises from mysterious force of quantum effects

David Chalmers, UC Santa Cruz philosopher:

Basis of consciousness may be mysterious new type that has not yet been observed

One thing in common: All appeal to materialistic (even if mysterious) explanations

An alternative view

Thesis:

“Strong AI” is fundamentally wrong

• •

There is a fundamental difference Machines can never be equivalent to humans in all respects

“Weak AI” arguments are – well – weak because they all assume materialism in their foundation

The dilemma is solved by recognizing the role of the spirit (or soul)

The “Strong AI” model

Computer Human input output deterministic algorithm (running on silicon) input output deterministic algorithm (running on carbon)

The proposed model

Computer Human immaterial spirit input output deterministic algorithm (running on silicon) input output deterministic algorithm (running on carbon)

Turing machine

Recall the Turing machine:

This simple abstract device (Universal Turing machine) can simulate the behavior of any digital computer – past, present, or future!

Benjamin Schumacher, http://physics.kenyon.edu/coolphys/thrmcmp/newcomp.htm

The proposed model

Computer Human contingency mechanism deterministic algorithm (running on silicon) deterministic algorithm (running on carbon) and immaterial decision maker

Computer

Let’s zoom in

Human

http://gs.fanshawec.ca/tlc/math270/images/2_7_Bi2.jpg

0 1 contingency mechanism – decision made by immaterial spirit affects physical outcome

How can this model be tested?

0010001010101… input device output 010111010001…

Kolmogorov complexity of output string s is the length of the shortest program that outputs s

Example: K( 22/7 ) < K(

p

)

Define: Complexity of device is the maximum Kolmogorov complexity that it can output,

when no input is given For Turing machines, K(output) ≤ K(input) + C(device)

Conservation of information

device input image lossless compression algorithm (e.g., LZW) output image number of bits ( output ) < number of bits ( input ) complexity ( output ) = complexity ( input ) Process is reversible

(Note: This complexity is entropy, not Kolmogorov complexity)

Conservation of information

device input image lossy compression algorithm (e.g., JPEG) output image number of bits ( output ) < number of bits ( input ) complexity ( output ) < complexity ( input ) Process is NOT reversible

Conservation of information

device downsample input image output image number of bits ( output ) < number of bits ( input ) complexity ( output ) < complexity ( input ) Process is NOT reversible

Conservation of information

• •

Furby (1998)

– –

speaks Furbish off-the-shelf learns English over time

How does it do this?

– –

Pre-programmed to speak English Program causes more English to be used over time

– –

Nothing is learned No new information

Information generation

• • •

We do not expect computers to generate new information Rather, they only process existing information This limitation is NOT dependent on the speed / computational power of the computer “You have illegal files on your computer!” “No officer. You see, I just bought this new processor, and it’s so powerful that it decided to create those files.”

http://digitalclonesrus.com/assets/images/happy_man_at_computer.jpg

http://www.clipartof.com/images/clipart/xsmall2/4205_motorcycle_policeman_filling_out_a_traffic_citationticket_form.jpg

Information generation

We DO expect people to generate new information:

The basic requirement for a PhD is a contribution to human knowledge

Intellectual property assumes that knowledge is created by the inventors

Plagiarism is detected when one person’s work is similar to another’s

There is a distinction between original work and derivative work

Example:

In 2005, students at MIT (Stribling et al.) wrote a computer program to generate research papers

The automatically generated paper was actually accepted for publication by the W orld Multi-Conference on Systemics, Cybernetics and Informatics (WMSCI)?

– –

Why was this such a scandal?

Why did people complain that the conference organizers had not reviewed submissions thoroughly rather than conclude that computers had now reached human intelligence?

Generating information

0010001010101… input device output 010111010001… If complexity ( output ) > complexity ( input ), then the complexity must arise from the device itself (cf. Noam Chomsky’s black box for studying children’s innate ability to learn language)

But the human brain is not complex enough. Back-of-the-envelope calculation:

Library of Congress has approx. 20 TB of written information Human genome contains 3 billion DNA rungs for a total of 6 billion bits of data 20,000,000,000,000 >> 1,000,000,000

Free will

• • • •

Contingency mechanism enables human to make free decisions This is necessary for

moral responsibility

– – –

ethical standards laws of justice (e.g., was the act intentional?) self improvement (Covey’s gap between stimulus and response) In historic Christian theology, people are defined as “rational creatures”, which implies

free will

immaterial, immortal soul

“So God created man in his own image, in the image of God created he him; male and female created he them.” Gen. 1:27

Without free will, we are either

deterministic, or

random Either way leads to irrationality

Consciousness

• • •

Common view is that consciousness arises from materialistic processes in body Why? Not because of evidence, but because of prior philosophical commitment to naturalism Kurzweil proposes to produce exact replica of brain

Then he will be automatically transported to the copy

Why should we think that, even if our brain could be copied, our consciousness should go with it?

What if the brain is copied multiple times? Will we have multiple consciousnesses?

This is a vain attempt at immortality (Salvation by computer upload)

Note that all proponents of this idea predict that the technology will conveniently be in place by the time they are 70 years old – see Pattie Maes, "Why Immortality is a Dead Idea", 1993

Creativity

• • • • •

Consider song as point in high-dimensional space Interpolating between songs may be possible (blending) Creating new songs in local neighborhood may be possible Claim: Making meaningful macro-jumps is not possible Even if it were possible, who would be the judge of quality? Computer or human?

• • • •

What’s wrong with the Turing test?

Turing test:

one computer, one person, one judge

– –

All communication via terminal Goal: Judge tries to tell which is computer and which is person Turing test can never be used to tell whether there is a fundamental difference between computer and human Reason: Judge is required to be a human In other words,

Suppose computer = human

– –

Then human judge can be replaced by computer judge But now test does not make any sense

More…

• • • • •

Chess revisited:

Sure, computers can play chess, but can they invent a new game to replace chess?

Can they invent new rules?

Artificial life started with promise, then fizzled out as hopes were not realized Captchas: Reverse Turing tests Church-turing thesis Complex specified information

One final thought

• •

Similarity between computers and animals:

Animals act by instinct

Animals do not have free will Learn the lesson of Grizzly Man (Timothy Treadwell):

He thought bears were his friends

He thought they were misunderstood

He ignored warnings about getting too close

They killed him

Is computer vision possible?

• Distinction between – information processing: the information is changed from one form to another, or is lost • algorithms change information – information generation: the information is created • natural processes cannot create information • there is no algorithm to create information • information generation requires a contingency mechanism • supernatural or metanatural process -> spirit or soul • Will a computer ever be able to

enjoy

an aesthetically pleasing painting?

if (painting_is_pretty) { printf(“I love this painting”); }