A Few Observations on Shimon Edelman’s talk. 11/13/14 Valid criticism of the claim that Deep Learning (multi- layer neural nets) and Reinforcement Learning (RL) come close to capturing full cognition/AI. The brain is much more complex! Need to move beyond input-output view of machine learning /AI. (Beyond the “Google query.”) It’s the stimulus-response view of actions. Captures only the most basic reflex agent behaviors. Need richer internal representations of the world. Knowledge representation. Causality. Can explain eg learning new behaviors from just one or two examples (contrast with RL).
But, there is also some validity behind the excitement for deep learning and RL using Big Data: the success on vision tasks (image recognition) and speech recognition. I.e. “sensory processing.” These tasks have frustrated AI researchers for decades. Famously underestimated: “solving image recognition (objects etc.) was assigned as a *part time* ugrad project at MIT! Finally, 5 decades later, real progress!
Can we combine these advances with advances in the symbolic/probabilistic arena, i.e., knowledge representation, reasoning, and search? E.g., use deep learning and RL to get the sensory inputs into symbolic representation and proceed from there. This will not model the brain in any detail but the overall performance could be impressive nevertheless. Systems like Watson and self-driving cars suggest that system integration of many “weak” and distinct components can be very powerful overall.
Midterm: Material up-to-and-including Adversarial Search (Game trees, minimax, & alpha-beta, expectiminimax; slides #10) 1) See lecture slides on web. 2) Study sections in R&N. 3) Consider hkw problems and solns. Midterm: closed book but 1 two-sided sheet of notes allowed. (typed / hand written / any way you like) Re: reinforcement learning, focus on slides. R&N chapter 21 too much depth. 10/22/14
How do we search for “longest path” on a map? 09/26/14 Uniform cost search with highest cost first? NO (missing “Markovian” property of shortest path. Other approach? General tree search. No repeating states down a single branch (limits depth to n nodes). What if you reach the goal? Keep going! (Stop only after full space is searched and return longest found path to goal. Can we do better? If so, how? If not, why not? Not significantly. P =/= NP
Notes Wedn. 09/10/14 Course url: Enrollment is finalized. Everyone now in CMS. This Friday 09/12/14: 11:25am this room CS Organizational meeting. News item: CNN --- When Machines Outsmart Humans (link on course page) --- Superintelligence by Nick Bostrom
Notes Fri. 09/05/14 Course url: CMS will be populated by Monday. Enrollment: how many *not* enrolled? Hopefully, resolved by today. --- CS ugrad admin