Probabilistic inference in human semantic memory Mark Steyvers, Tomas L. Griffiths, and Simon Dennis 소프트컴퓨팅연구실오근현 TRENDS in Cognitive Sciences vol. 10,

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Modelling the Virtual Machine in Simple Rating and Categorization Tasks Simon Dennis School of Psychology University of Adelaide.
Biointelligence Laboratory, Seoul National University
Benjamin Allred 벤자민 알레드 Contents  Questions to Think About  Definitions  Recognition Versus Recall  Single Process Models  Generate-Recognize Models.
1 Latent Semantic Mapping: Dimensionality Reduction via Globally Optimal Continuous Parameter Modeling Jerome R. Bellegarda.
Bayesian Inference for Signal Detection Models of Recognition Memory Michael Lee Department of Cognitive Sciences University California Irvine
Dimension reduction (1)
What is Statistical Modeling
Introduction  Bayesian methods are becoming very important in the cognitive sciences  Bayesian statistics is a framework for doing inference, in a principled.
Expectation Maximization Method Effective Image Retrieval Based on Hidden Concept Discovery in Image Database By Sanket Korgaonkar Masters Computer Science.
Cognitive Processes PSY 334 Chapter 8 – Problem Solving May 21, 2003.
A probabilistic approach to semantic representation Paper by Thomas L. Griffiths and Mark Steyvers.
Switch to Top-down Top-down or move-to-nearest Partition documents into ‘k’ clusters Two variants “Hard” (0/1) assignment of documents to clusters “soft”
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
ITCS 6010 Natural Language Understanding. Natural Language Processing What is it? Studies the problems inherent in the processing and manipulation of.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Retrieval Models II Vector Space, Probabilistic.  Allan, Ballesteros, Croft, and/or Turtle Properties of Inner Product The inner product is unbounded.
Baysian Approaches Kun Guo, PhD Reader in Cognitive Neuroscience School of Psychology University of Lincoln Quantitative Methods 2011.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Statistics in Bioinformatics May 12, 2005 Quiz 3-on May 12 Learning objectives-Understand equally likely outcomes, counting techniques (Example, genetic.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Theory (and Application) Learning: A change in human performance or performance potential that results from practice or other experience and endures over.
1 A Presentation of ‘Bayesian Models for Gene Expression With DNA Microarray Data’ by Ibrahim, Chen, and Gray Presentation By Lara DePadilla.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Speech Analysing Component in Automatic Tutoring Systems Presentation by Doris Diedrich and Benjamin Kempe.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Introduction to Bayesian statistics Yves Moreau. Overview The Cox-Jaynes axioms Bayes’ rule Probabilistic models Maximum likelihood Maximum a posteriori.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
Katrin Erk Vector space models of word meaning. Geometric interpretation of lists of feature/value pairs In cognitive science: representation of a concept.
Shane T. Mueller, Ph.D. Indiana University Klein Associates/ARA Rich Shiffrin Indiana University and Memory, Attention & Perception Lab REM-II: A model.
Emergence of Semantic Knowledge from Experience Jay McClelland Stanford University.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Pattern Recognition with N-Tuple Systems Simon Lucas Computer Science Dept Essex University.
WHEN THE PROBE IS WILLING BUT THE MIND IS WEAK This research was supported by an NSERC operating grant For additional information please contact Marty.
Michael Isard and Andrew Blake, IJCV 1998 Presented by Wen Li Department of Computer Science & Engineering Texas A&M University.
Learning to Share Meaning in a Multi-Agent System (Part I) Ganesh Padmanabhan.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
CS246 Latent Dirichlet Analysis. LSI  LSI uses SVD to find the best rank-K approximation  The result is difficult to interpret especially with negative.
Link Distribution on Wikipedia [0407]KwangHee Park.
10.0 Latent Semantic Analysis for Linguistic Processing References : 1. “Exploiting Latent Semantic Information in Statistical Language Modeling”, Proceedings.
OBJECT TRACKING USING PARTICLE FILTERS. Table of Contents Tracking Tracking Tracking as a probabilistic inference problem Tracking as a probabilistic.
Result 1: Effect of List Length Result 2: Effect of Probe Position Prediction by perceptual similarity Prediction by physical similarity Subject
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
From NARS to a Thinking Machine Pei Wang Temple University.
Memory. Hopfield Network Content addressable Attractor network.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Plan for Today’s Lecture(s)
Computer vision: models, learning and inference
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
A Hierarchical Bayesian Look at Some Debates in Category Learning
Michal Rosen-Zvi University of California, Irvine
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Topic Models in Text Processing
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Models of memory Nisheeth April 4th 2019.
Volume 23, Issue 21, Pages (November 2013)
CS249: Neural Language Model
Presentation transcript:

Probabilistic inference in human semantic memory Mark Steyvers, Tomas L. Griffiths, and Simon Dennis 소프트컴퓨팅연구실오근현 TRENDS in Cognitive Sciences vol. 10, no. 7, 2006

Overview Relational models of memory The Retrieving Effectively from Memory(REM) Model High-dimensional semantic spaces Probabilistic topic models Modeling semantic memory at the sentence level Conclusion 1

Relational Models of Memory Encoding much knowledge stored in memory using language –Most of the stimuli used in memory experience are linguistic –How the statistics of language influence human memory? Rational analysis –A framework for developing computational models of cognition –Making the working assumption human cognition approximates an optimal response to the computational problems posed by the environmental The role of probabilistic inference in memory –A history factor The occurrence pattern of the item over time( Recent or frequent ) –A context factor The association between items (Similar) 2

The Retrieving Effectively from Memory(REM) Stronger assumptions –Encoding process –Representation of information Emphasizing the role of probabilistic inference in explaining human memory Recognition memory task(Application) –Discriminating between old items and new items –Assumptions Words : Vectors of Features(noisy and incomplete) –A calculation of the likelihood ratio Balancing the evidence for and ‘old’ decision against a ‘new’ decision Degree of match between the memory probe and contents of memory 3

Example of Comparing a Test 5 Features Green box : Matches Red box : Mismatches Empty Positions : Missing Features 3 rd memory trace : the most evidence that test item is old Data ‘D’ : Each trace 4

Posterior odds Posterior Odds –Posterior : 증거 (Evidence) 이후 –Odds : 두 배반적 가설의 확률 비 Prior odds is set at 1  # of old == # of new An old decision –Exceeding some criterion(usually set at 1) 5 An example distribution of log posterior odds

Mirror Effects Increase in hit rate –Hit rate : correctly recognizing an old item as old decrease in false alarm rates –False alarm rates : falsely recognizing a new item as old Example –Low frequency words –More rare features are stored for low frequency words relative to high frequency words –How encoding conditions that improves the diagnosticity of feature matches for low frequency target items –lower the probability of chance matches by low frequency distractor items 6

The Hyperspace Analog to Language(HAL) Each word by a vector where each element of the vector corresponds to a weighted co-occurrence value of that word with some other word 7

The Latent Semantic Analysis(LSA) The co-occurrence information between words and passages 8 Applying matrix decomposition techniques –to reduce the dimensionality of the original matrix to a much smaller size –preserving as much as possible the covariation structure of words and documents

The Word Association Space(WAS) Input a set of association norms –Formed by asking subjects to produce the first word that comes to mind in response to a given cue 9

Probabilistic Topic Models The Focus on prediction as a central problem of memory Emphasizing the role of context in guiding predictions The latent structure of words using topics –Documents : mixtures of topics –A topic : A Probability distributions over words Topic Model is a generative model for documents 10

Distributions for Topics Different weight to thematically related words P(z) : The distribution over topics z in a particular document P(W|z) : The probability distribution over words w given topic z P( =j) : the probability that the ‘j’th topic was sampled ‘i’th word P( | =j) : the probability of word under topic j T : The number of topics 11

Using Bayes’ Rule The problem is to predict the conditional probability of word (the response word) given the cue word Applying Bayes’s rule 12

The same word with two different meaning 13

Predictions about which words are likely to appear next in a document or conversation, based on the previous words A rational model of how context should influence memory –The focus on the role of history in earlier models –The effects of context being appropriately modulated by the statistics of the environment 14

The Syntagmatic Paradigmatic(SP) Model Modeling Semantic Memory at the sentence level Specifying a simple probabilistic model of knowledge and allowing representational content –to emerge response to the structure of the environment Syntagmatic Associations –Capturing structural and propositional knowledge between words that follow each other –Structural traces that correspond to individual sentence Paradigmatic Associations –between words that fit in the same slots across sentences –combined to form relational(or propositional) traces that correspond to the same sentences 15

Sampling structural and relational exemplars Mapping the two sentences to each other in a one to one fashion Change : Sampras-Kuerten / Agassi-Roddick Match : defeated Delete : Pete – ㅡ (symbol of deletion, empty words) Priority –Match > change > insertion or deletion 16

The Architecture of SP Model 4 and 5 traces in Sequential LTM # containing the pattern {Roddick, Costa} {Sampras} defeat {Agassi}  Agassi need to align ‘#’ symbol The pattern {Agassi, Roddick, Costa} 17

Conclusion Probabilistic inference –A natural way to address problems of reasoning under uncertainty A search for richer representations of the structure of linguistic stimuli Contribution to the question –How linguistic stimuli might be represented in memory Rational Analysis –A natural framework to understand the tight coupling between behavior and environmental staticstics –Linguistic stimuli form an excellent testing ground for rational model of memory 18