A LLISON S EIBERT & A LEXANDRA W ARLEN Efficient Episode Recall and Consolidation E MILIA V ANDERWERF & R OBERT S TILES.

Slides:



Advertisements
Similar presentations
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 7: Scoring and results assembly.
Advertisements

PERMUTATION CIRCUITS Presented by Wooyoung Kim, 1/28/2009 CSc 8530 Parallel Algorithms, Spring 2009 Dr. Sushil K. Prasad.
Reconciling Differences: towards a theory of cloud complexity George Varghese UCSD, visiting at Yahoo! Labs 1.
Incremental Clustering Previous clustering algorithms worked in “batch” mode: processed all points at essentially the same time. Some IR applications cluster.
MIT CSAIL Vision interfaces Towards efficient matching with random hashing methods… Kristen Grauman Gregory Shakhnarovich Trevor Darrell.
Similarity Search in High Dimensions via Hashing
Presented by Arshad Jamal, Rajesh Dhania, Vinkal Vishnoi Active hashing and its application to image and text retrieval Yi Zhen, Dit-Yan Yeung, Published.
Genetic Algorithms Chapter 9, Complexity: A Guided Tour.
1 Episodic Memory for Soar Andrew Nuxoll 15 June 2005.
Self-Organizing Hierarchical Neural Network
Multiple Sequence Alignment Algorithms in Computational Biology Spring 2006 Most of the slides were created by Dan Geiger and Ydo Wexler and edited by.
1 Jun Wang, 2 Sanjiv Kumar, and 1 Shih-Fu Chang 1 Columbia University, New York, USA 2 Google Research, New York, USA Sequential Projection Learning for.
1 Episodic Memory for Soar Agents Andrew Nuxoll 11 June 2004.
Genetic Algorithms GAs are one of the most powerful and applicable search methods available GA originally developed by John Holland (1975) Inspired by.
Doug Downey, adapted from Bryan Pardo, Machine Learning EECS 349 Machine Learning Genetic Programming.
Similarity Search in High Dimensions via Hashing Aristides Gionis, Protr Indyk and Rajeev Motwani Department of Computer Science Stanford University presented.
Object Recognition Using Geometric Hashing
Intro to AI Genetic Algorithm Ruth Bergman Fall 2002.
CS5371 Theory of Computation Lecture 8: Automata Theory VI (PDA, PDA = CFG)
Intro to AI Genetic Algorithm Ruth Bergman Fall 2004.
Finding Regulatory Motifs in DNA Sequences. Motifs and Transcriptional Start Sites gene ATCCCG gene TTCCGG gene ATCCCG gene ATGCCG gene ATGCCC.
Chapter 8.  Cryptography is the science of keeping information secure in terms of confidentiality and integrity.  Cryptography is also referred to as.
Indexing Techniques Mei-Chen Yeh.
Approximation algorithms for large-scale kernel methods Taher Dameh School of Computing Science Simon Fraser University March 29 th, 2010.
Genetic Algorithms: A Tutorial
Image Based Positioning System Ankit Gupta Rahul Garg Ryan Kaminsky.
Genetic Algorithms and Ant Colony Optimisation
1 An Overview of Evolutionary Computation 조 성 배 연세대학교 컴퓨터과학과.
A Multi-Domain Evaluation of Scaling in Soar’s Episodic Memory Nate Derbinsky, Justin Li, John E. Laird University of Michigan.
Cognitive Processes PSY 334 Chapter 7 – Human Memory: Retention and Retrieval May 16, 2003.
Introduction to Genetic Algorithms and Evolutionary Computation
SOFT COMPUTING (Optimization Techniques using GA) Dr. N.Uma Maheswari Professor/CSE PSNA CET.
Evolution Strategies Evolutionary Programming Genetic Programming Michael J. Watts
B-trees and kd-trees Piotr Indyk (slides partially by Lars Arge from Duke U)
Streaming Algorithms Piotr Indyk MIT. Data Streams A data stream is a sequence of data that is too large to be stored in available memory Examples: –Network.
Genetic Algorithms Michael J. Watts
Nearest Neighbor Paul Hsiung March 16, Quick Review of NN Set of points P Query point q Distance metric d Find p in P such that d(p,q) < d(p’,q)
Genetic Algorithms Genetic algorithms imitate a natural optimization process: natural selection in evolution. Developed by John Holland at the University.
Fast Similarity Search for Learned Metrics Prateek Jain, Brian Kulis, and Kristen Grauman Department of Computer Sciences University of Texas at Austin.
Applying Genetic Algorithm to the Knapsack Problem Qi Su ECE 539 Spring 2001 Course Project.
1/27 Discrete and Genetic Algorithms in Bioinformatics 許聞廉 中央研究院資訊所.
1 Machine Learning: Lecture 12 Genetic Algorithms (Based on Chapter 9 of Mitchell, T., Machine Learning, 1997)
1 “Genetic Algorithms are good at taking large, potentially huge search spaces and navigating them, looking for optimal combinations of things, solutions.
Genetic Algorithms Siddhartha K. Shakya School of Computing. The Robert Gordon University Aberdeen, UK
Derivative Free Optimization G.Anuradha. Contents Genetic Algorithm Simulated Annealing Random search method Downhill simplex method.
How Solvable Is Intelligence? A brief introduction to AI Dr. Richard Fox Department of Computer Science Northern Kentucky University.
Future Memory Research in Soar Nate Derbinsky University of Michigan.
Random Boolean Networks. Attractors A Boolean network has 2 ^N possible states. Sooner or later it will reach a previously visited state, and thus,
Evolving the goal priorities of autonomous agents Adam Campbell* Advisor: Dr. Annie S. Wu* Collaborator: Dr. Randall Shumaker** School of Electrical Engineering.
Similarity Searching in High Dimensions via Hashing Paper by: Aristides Gionis, Poitr Indyk, Rajeev Motwani.
1 Genetic Algorithms and Ant Colony Optimisation.
Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality Piotr Indyk, Rajeev Motwani The 30 th annual ACM symposium on theory of computing.
Outline Problem Background Theory Extending to NLP and Experiment
Genetic Algorithms MITM613 (Intelligent Systems).
CS Machine Learning Instance Based Learning (Adapted from various sources)
Competence-Preserving Retention of Learned Knowledge in Soar’s Working and Procedural Memories Nate Derbinsky, John E. Laird University of Michigan.
Genetic Algorithm. Outline Motivation Genetic algorithms An illustrative example Hypothesis space search.
March 1, 2016Introduction to Artificial Intelligence Lecture 11: Machine Evolution 1 Let’s look at… Machine Evolution.
Intro to Data Structures Concepts ● We've been working with classes and structures to form linked lists – a linked list is an example of something known.
Evolution Strategies Evolutionary Programming
Data Structures (CS212D) Overview & Review.
Instance Based Learning (Adapted from various sources)
Yu-Chi Ho Jonathan T. Lee Harvard University Sep. 7, 2000
John Laird, Nate Derbinsky , Jonathan Voigt University of Michigan
Near(est) Neighbor in High Dimensions
Joseph Xu Soar Workshop 31 June 2011
Locality Sensitive Hashing
Introduction to Artificial Intelligence Lecture 11: Machine Evolution
Beyond Classical Search
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

A LLISON S EIBERT & A LEXANDRA W ARLEN Efficient Episode Recall and Consolidation E MILIA V ANDERWERF & R OBERT S TILES

Task: Hashing Episodes ●Recognition without Recall (Wallace, et.al. 2013) ●Hash Function Requirements: ●Fast ●Repeatable (S1 ^epmem E1) (E1 ^command C1) (E1 ^present-id 1) (E1 ^result R2) (S1 ^io I1) (I1 ^input-link I2) (I2 ^eater E2) (E2 ^name red) (E2 ^score 0) (E2 ^x 13) Hash Fn 0110…1 ●General ●Constant time (Wallace, et.al. 2013)

Motivation: Feature Importance New Requirement: similar episodes generate similar hash values Encoding – do I need to encode this? Storage – is it ok to forget this? Retrieval – what is the best match for this cue?

Environment 4

(S1 ^epmem E1) (E1 ^command C1) (E1 ^present-id 1) (E1 ^result R2) (S1 ^io I1) (I1 ^input-link I2) (I2 ^eater E2) (E2 ^name red) (E2 ^score 0) (E2 ^x 13) epmem E1input-link I2north E4score 0content bonusfood Hash Formula 0: epmem E1 1: input-link I2 2: north E4 3: score 0 4: content bonusfood Hash Code Size: 5 CURRENT EPISODE

Genetic Algorithm Hashing 6 1: Parents 2: Children Generation I 3: Mutations Generation II 4: Find the two best children 5: Rinse and Repeat (Holland 1992)

Folding Hash Function 7 Never gonna give you up Never gonna let you down nevergonnagive you up (Bloom 1970)

Folding Hash Function never down gonnagive you up let Never gonna give you up Never gonna let you down (Bloom 1970)

Folding Hash Function never down make lie gonna run cry and give around say hurt you and goodbye up desert tell let make a Never gonna give you up Never gonna let you down Never gonna run around and desert you Never gonna make you cry Never gonna say goodbye Never gonna tell a lie and hurt you (Bloom 1970)

Folding Hash Function never down make lie gonna run cry and give around say hurt you and goodbye up desert tell let make a Never gonna give you up Never gonna let you down Never gonna run around and desert you Never gonna make you cry Never gonna say goodbye Never gonna tell a lie and hurt you (Bloom 1970)

Locality Sensitive Hashing Dictionary Hash Formula - Code Size 5 ≈ epmem E1 north E5 content eater score 0 input-link I2 content wall input-link I2 score 0 content eater north E5 epmem E1 south N3 content wall 11 (Indyk, etl.al. 1998)

Sweet Spot Hash Function 12 GA is selecting WMEs with moderate frequency of use.

Sweet Spot Hash Permutations 13 Replacing Hash Formula never gonna give we’re no and rules in love you know the so i strangers you up let down run around { givewe’renoandrules { we’renoandrulesin never gonna give we’re no and rules in love you know the so i strangers you up let down run around

Folding Sweet Spot Hash Function 14 Never gonna give you up Never gonna let you down Never gonna run around and desert you Never gonna make you cry Never gonna say goodbye Never gonna tell a lie and hurt you NeverGonnaYou And Let Give Desert Cry Goodbye VS

15

16

● Folding SS was able to reproduce GA results with smaller hashcode size ● Folding SS relies upon having a dictionary of known features (potentially grows forever) 17 Nuggets and Coal

Citations 18 Bloom, Burton H. (1970), "Space/Time Trade-offs in Hash Coding with Allowable Errors", Communications of the ACM 13 (7): 422–426. Holland, John (1992). Adaptation in Natural and Artificial Systems. Cambridge, MA: MIT Press. ISBN Indyk, Piotr.; Motwani, Rajeev. (1998)., "Approximate Nearest Neighbors: Towards Removing the Curse of Dimensionality.". Proceedings of 30th Symposium on Theory of Computing. Wallace, Scott, Dickinson, Evan and Nuxoll, Andrew (2013) Hashing for Lightweight Episodic Recall.;In Proceedings of AAAI Spring Symposium: Lifelong Machine Learning