Transcript notes.

12/10/14
Exam
Wedn., 12/17/13, 2pm-4:30pm,
Baker Laboratory 200
(http://registrar.sas.cornell.edu/Sched/EXFA.html)
Material:
Cumulative. Covers all material.
Study: Slides complemented by R&N.
Study: Hwks solns. & midterm.
Allowed: 1 double-sided sheet of notes & calculator.
A Few Observations on Shimon
Edelman’s talk. 11/13/14
• Valid criticism of the claim that Deep Learning (multilayer neural nets) and Reinforcement Learning (RL)
come close to capturing full cognition/AI. The brain is
much more complex!
• Need to move beyond input-output view of machine
learning /AI. (Beyond the “Google query.”) It’s the
stimulus-response view of actions. Captures only the
most basic reflex agent behaviors.
• Need richer internal representations of the world.
Knowledge representation. Causality. Can explain eg
learning new behaviors from just one or two examples
(contrast with RL).
But, there is also some validity behind the excitement for
deep learning and RL using Big Data:
the success on vision tasks (image recognition) and
speech recognition. I.e. “sensory processing.”
These tasks have frustrated AI researchers for decades.
Famously underestimated: 1966 --- “solving image
recognition (objects etc.) was assigned as a *part time*
ugrad project at MIT!
http://projects.csail.mit.edu/films/aifilms/AIFilms.html
Finally, 5 decades later, real progress!
Can we combine these advances with advances in
the symbolic/probabilistic arena, i.e., knowledge
representation, reasoning, and search?
E.g., use deep learning and RL to get the sensory
inputs into symbolic representation and proceed
from there. This will not model the brain in any
detail but the overall performance could be
impressive nevertheless.
Systems like Watson and self-driving cars suggest
that system integration of many “weak” and
distinct components can be very powerful overall.
10/22/14
• Midterm: Material up-to-and-including Adversarial
Search (Game trees, minimax, & alpha-beta,
expectiminimax; slides #10)
1) See lecture slides on web.
2) Study sections in R&N.
3) Consider hkw problems and solns.
• Midterm: closed book but 1 two-sided sheet of notes
allowed. (typed / hand written / any way you like)
• Re: reinforcement learning, focus on slides. R&N
chapter 21 too much depth.
How do we search for “longest path”
on a map? 09/26/14
Uniform cost search with highest cost first? NO
(missing “Markovian” property of shortest path.
Other approach? General tree search. No
repeating states down a single branch (limits
depth to n nodes). What if you reach the goal?
Keep going! (Stop only after full space is searched
and return longest found path to goal.
Can we do better? If so, how? If not, why not?
Not significantly. P =/= NP
Notes Wedn. 09/10/14
• Course url:
www.cs.cornell.edu/courses/cs4700/2014fa/
Enrollment is finalized. Everyone now in CMS.
This Friday 09/12/14: 11:25am this room
CS 4701 --- Organizational meeting.
News item: CNN --- When Machines Outsmart Humans
(link on course page)
--- Superintelligence by Nick Bostrom
Notes Fri. 09/05/14
• Course url:
www.cs.cornell.edu/courses/cs4700/2014fa/
CMS will be populated by Monday.
Enrollment: how many *not* enrolled?
Hopefully, resolved by today.
[email protected] --- CS ugrad admin