Presentation on theme: "Semantics and Context in Natural Language Processing (NLP) Ari Rappoport The Hebrew University."— Presentation transcript:
Semantics and Context in Natural Language Processing (NLP) Ari Rappoport The Hebrew University
Form vs. Meaning Solid NLP progress with statistical learning OK for classification, search, prediction – Prediction: language models in speech to text But all this is just form (lexicon, syntax) Language is for expressing meanings (semantics) – (i) Lexical, (ii) sentence, (iii) interaction
(i) Lexical Semantics Relatively context independent Sensory feature conjunction: house, hungry, guitar – Non-linguistic semantic machine learning: face identification in photographs Categories: is-a: a smartphone is a (type of) product, iPhone is a (type of) smartphone Configurations: part-of: engine:car, places
Generic Relationships Medicine : Illness – Hunger : Thirst – Love : Treason – Law : Anarchy – Stimulant : Sensitivity You use X in a way W, to do V to some Z, at a time T and a place P, for the purpose S, because of B, causing C, …
Flexible Patterns X w v Y : countries such as France – Davidov & Rappoport (ACL, EMNLP, COLING, etc) Content words, High frequency words Meta-patterns: CHHC, CHCHC, HHC, etc. Fully unsupervised, general Efficient hardware filtering, clustering Categories, SAT exams, geographic maps, numerical questions, etc.
Ambiguity Relative context independence does not solve ambiguity Apple: fruit, company Madrid: Spain, New Mexico Which one is more relevant? Context must be taken into account – Language use is always contextual
(ii) Sentence Semantics The basic meaning expressed by (all) languages: argument structure scenes Dynamic or static relations between participants; elaborators; connectors; embeddings: – John kicked the red ball – Paul and Anne walked slowly in the park – She remembered Johns singing
Several Scenes Linkers: cause, purpose, time, conditionality – He went there to buy fruits, Before they arrived, the party was very quiet, If X then Y Ground: referring to the speech situation – In my opinion, machine learning is the greatest development in computer science since FFT [and neither were done by computer scientists] Career, Peace, Democracy
Sentence Semantics in NLP Mostly manual: FrameNet, PropBank Unsupervised algorithms – Arg. identification, Abend & Rappoport (ACL 2010) Question Answering – Bag of words (lexical semantics) Machine Translation – Rewriting of forms (alignment, candidates, target language model)
Extreme Semantic Application Tweet Sentiment Analysis – Schwartz (Davidov, Tsur) & Rappoport 2010, 2013 Coarse semantics: 2 categories (40) Short texts, no words lists; fixed patterns
(iii) Interaction Semantics Understanding means having enough information to DO something – The brains main cycle Example: human-system interaction Full context dependence – Relevance to your current situation
Interaction Examples Searching Argo, did you mean – The plot? Reviews? Where and/or when to watch? Chinese restaurant – The best in the country? In town? The nearest to you? The best deal? There are hints: – Location (regular, irregular); time (lunch?)
Interaction Directions Extending flexible patterns: – Include Text-Action H and C items (words, actions) Action: – represented as User Interface operations Shortcut: bag of words (lexical semantics) + current context. Ignore sentence semantics Noise, failure (Siri, maps,…)
Summary Lexical, sentence, and interaction semantics Applications are possible using all levels As relevance to life grows, so do requirements from algorithms Both sentence and interaction semantics necessary for future smart applications Current focus: sentence semantics
Your consent to our cookies if you continue to use this website.