1
Representation / 2 Representation / 2
Semantic Nets - Schemes: Frames, Scripts Semantic Nets - Schemes: Frames, Scripts
Laurea in INFORMATICA Laurea in INFORMATICA
MAGISTRALE MAGISTRALE
Corso di
ARTIFICIAL INTELLIGENCE Stefano Ferilli
Categorization
●
Procedure to organize objects into classes or categories
●
Allows defining properties and making assumptions on the entire category rather than on single elements
●
Crucial in the conceptualization and knowledge representation phase
– Choice of categories determines what can be represented
●
Common in technical-scientific fields
– Leads to the definition of taxonomies
●Biology: living species
●Chemistry: elements
●Archives: subject fields
●...
●
Humans naturally organize concepts in 3 hierarchical levels
●
Basic, Subordinate, Superordinate
●[Rosch & Mervis (1975): “Family resemblance: studies in the internal structures of categories”, Cognitive Psychology, 573- 605]
– Basic concepts : the natural way to categorize objects and entities that make up our world
●The first categories learned by humans
●Other concepts acquired later as an evolution of basic categories
–Superordinate concepts -> generalization –Subordinate concepts -> specialization – E.g.:
●“chair” basic concept
●“furniture” superordinate with respect to chair
●“rocking chair” subordinate
Categorization
Categorization
●
Once formed, categories tend to have a structure that
●
stresses the similarity among members of the same category
●
maximizes the differences between members in different categories
●
This allows reasoning with prototypes
●
Prototype
– A specimen of the category near the center of the space of features for that category
Categories and Objects
●
Much reasoning happens at the category level, rather than on individuals
●
If knowledge is organized into (sub-)categories, it is sufficient to classify an object, by its perceived properties, to infer the properties of the category/- ies it belongs to
●
Simplifies the knowledge base, using INHERITANCE as a form of inference
– Placing an attribute high in the taxonomy one may allow many instances to inherit that property
Inheritance
●
Property Inheritance
●
Objects are grouped into classes because they share some properties
– => there exist inference mechanisms that work on the structure of the representation
●
Characterization of a property A given at the class level but interpreted as a property of all instances of the class
– ∀x ( is_a(x, Class) A(x) )⇒ A(x) )
●
Implementation: GRAPH SEARCH
– Starting from the given node, bottom-up search in the taxonomy for a node defining its properties
Inheritance
●
Example
●
“Peter breathes air” can be inferred from the fact that Peter is a Person, Person is a subclass of Mammal, and Mammals breath Air
Mammal
Person
Air
Peter breathes
instance_of subclass
Prototypes
●
Reasoning with prototypes
– What typically characterizes a concept
●Necessary conditions –Bird(x) Vertebrate(x) –Bird(x) Bipede (x)
●Typically necessary conditions (default) –Bird(x) Tip Flies(x)
–Bird(x) Tip Feathered(x)
●Sufficient Conditions (criterials) –Canary(x) Bird(x) –Ostrich(x) Bird(x)
●Typically sufficient conditions –Flies(x) Tweets(x) Tip Bird(x) –Feathered(x) Tip Bird(x) – Source of non-monotonicity
Knowledge Representation Schemes
●
Semantic Nets
●
A method to represent knowledge using graphs made up of nodes and arcs
– Nodes objects
– Arcs relationships between objects
●
Frames
●
A data structure to represent concepts, objects and properties through stereotypes
●
Scripts
●
Describe sequences of typical events
Graphs for Knowledge Representation
●
Needs:
●
Analyze knowledge in terms of low-level primitives
●
Organize it in high-level structures
●
Basic units to represent common-sense knowledge using graphs
●
Concepts = nodes
●
Relationships = arcs
– Used to connect concepts to each other in the graph
Sowa’s Conceptual Graphs (GCs)
●
A method to represent mental models
●
Able to systematically and formally detect how what we think about a concept can be described in terms of its relationships with other kinds of concepts
●
2 kinds of nodes
●
Concepts (Concrete or abstract)
●
Relationships (partof, object of, characteristic of...) Concept 2 Relationship
The Relationship node explains the dependence between Concept1 and Concept2 Concept 1
Conceptual Graphs
●
Most appropriate assuming that
knowing something = ability of building mental models that accurately represent both that thing and the actions that can be taken through and/or on that thing
●
Power: able to formalize
●
Both concrete and abstract concepts
●
Hierarchies of concepts
– The base of the hierarchy defines concrete concepts on which progressilvely more abstract concepts rely
Semantic Nets
●
So called because initially used to represent the meaning of natural language utterances
●
(Quillian, 1966, Semantic Memory)
●
A more appropriate name would be Associative nets
– More neutral with respect to the application domain
●
Peculiarity related to cognitive economy consisting in the identification of fundamental structural relationships (system relationships) very frequent in reality
●
Focus on the meaning to give to sentences expressed in natural language
Semantic Nets
●
From Digraphs to Semantic Nets
●
Based on the graphical representation of the relationships existing between elements in a given domain, Semantic Nets adopt the metaphor that
– objects are nodes in a graph and
– such nodes are related by arcs representing binary relationships
●
Like all formalisms based on graphical methods, can be easily and immediately understood
Semantic Nets
●
From Digraphs to Semantic Nets
●
A Semantic Net is a Labeled Directed Digraph used to describe relationships between objects, situations or actions
– Powerful representation
– Can be reduced to a tractable symbolic representation (Logics)
●
Graph notation by itself has a little advantage over logical notation
– It allows one to represent relationships between objects and define inferences through links
Semantic Nets
●
An example
A B
C
A
B
C floor
above below
to_right smaller bigger
bigger smaller above below
to_l
tef htig_rto erggbi rlleasm above below
to_left
Semantic Nets
●
Representation of constraints on features
●
How to capture semantics in representation
Digraph Representation●
Note: without the constraint that pillars “do not touch”, the representation of “arch” and “not arch”
are exactly the same
Architrave ArchitraveArch Not arch
Left Pillar Right Pillar Left Pillar Right Pillar
PillarLeft Right
Pillar supported bysupported by
architrave
left_of right_of not touch
Semantic Nets
●
A systematic use when building knowledge- based systems due to the possibility of distinguishing, among nodes that represent concepts, token-nodes and type-nodes
●
Tokens = individual physical symbols
●
Types = sets of symbols
●
Properties of a token-node derive from the type-node to which it is linked
18
Event Action
Unpremeditated Murder
Premeditated Murder Murder Pistorius’
Murder
Semantic Nets
●
Quillian’s initial work defined most of the Semantic Net formalism
●
Labeled arcs and links, hierarchical inheritance and inferences along associative links, ...
●
but representation unsuitable for complex domains
●
Most links represent very general relationships, unable to capture the complex semantics of real world problems
Semantic Nets
●
Subsequent focus : definition of a richer set of labeled links useful to model the semantics of natural language
●
Key : identification of semantic primitives for natural language (Schank, 1974)
●
Semantic primitives of a language for Semantic Net processing correspond to elementary concepts that can be expressed through the language
– An interpreter can handle them because it is programmed since the beginning to understand them
Semantic Nets
●
Basic idea
●
Concept meaning determined by relationship to other objects
●
Information stored by interconnecting nodes (entities) by labeled edges (relationships)
– Example: Physical attributes of a person
●Translation in logics:
–isa(person, mammal), instance(Mike-Hall, person), team(Mike-Hall, Cardiff)
Semantic Nets
●
Typical Relationships:
– is_a
●supertype-type (superclass-class)
●type-subtype (class-subclass)
●subtype-instance (subclass-instance) – part_of
●supertype-type (superclass-class)
●type-subtype (class-subclass) – has
●object-property – value
●property-value – linguistic
●Examples: likes, owns, travel, made_of
Semantic Nets
●
Example: consider sentence
●
John gave Mary the book.
– Different aspects of one event
Semantic Nets
●
Inference
●
Basic mechanism: a form of spreading activation of the nodes in the net
●Already proposed by Quillian in his association nets devised for natural language processing
– Given as input an unknown proposition, the system should be able to find out a representation of its meaning based on the definitions that are available in the net
●More limited performance: the system would only be able to compare the meaning of two lexical entries
●
Comparison carried out by starting from parent/ancestor nodes of the words to be
compared and progressively visiting and activating
neighbor (type or token) nodes through association
arcs. This happens for the two propositions looking
for a common area of activated nodes
Semantic Nets
●
Inference
●
General mechanism: “follow the relationships between nodes”
●
1. Search for Intersection [proposed by Quillian]
– Procedure: spreading activation on two starting nodes
●Start by labeling a set of source nodes (the concepts in the semantic net) with weights/tags (“activations”)
●Then iteratively propagate them out of the source nodes towards the nodes linked to them
●Relationships among objects found by expanding the activation from the two nodes and finding their intersection. This is obtained by assigning a special tag to each visited node
Semantic Nets
●
Inference
●
2. (obviously) Inheritance
– Based on isa and instance links
– Leverages transitivity of isa
– Easily implemented as link traversal
Person
Parent
Mother IS-A IS-A
hasChild
Semantic Nets
●
Inference
●
What about n-ary relationships (n > 2)?
– Case structure representation technique
●
Example
gives(John,Mary,Book)
– John gives Mary a book
give-events
E1
book-4
John Mary
hasAgent
hasObjecthasRecipient
Semantic Nets
●
Inference
●
Inheritance also provides a means to carry out default reasoning
– Applicable unless exceptions
– Example:
●Emus are birds
●Typically birds fly and have wings
●Emus run (??)
Semantic Nets
●
Critique: lack of semantics!
– Ambiguities and inconsistencies in the use of nodes and arcs
●Woods [75] and others
– Semantics sometimes unclear or can be found out only by manipulation programs
– Examples of confusion
●isa for “belongs” and “subset of”
●Canonical instance or class of objects?
●Different meaning of relationships (among classes, between classes and objects, among objects)
●
What about logics?
– Semantic nets are a comfortable notation for a part of FOL and nevertheless can be cast as a logic formalism
●Well, not entirely...
Semantic Nets
●
Translation in logics
●Note: classes in uppercase, individuals in lowercase
A B
B IS-A
a
A R b
A R B
INSTANCE
●
x : A(x) B(x)
●
B(a)
●
x : x A R(x,b)
●
x : x A y B : R(x,y)
Semantic Nets
●
Sample Translation
– x Mammal(x) Animal(x)
– x Mammal(x) HasNPaws(x, 4)
– x Elephant(x) Mammal(x)
– x Elephant(x) HasColor(x, gray)
– Elephant(Clyde)
●
It is possible to deduce:
– Animal(Clyde)
– Mammal(Clyde)
– HasNPaws(Clyde, 4)
– HasColor(Clyde, gray)
Mammal
Elephant
Clyde Animal
4
gray HasColor HasNPaws
INSTANCE IS-A
IS-A
Semantic Nets
●
Exceptions
– Sample Translation●x Mammal(x) HasNPaws(x, 4)
●x Dolphin(x) Mammal(x)
●x Dolphin(x) HasNPaws(x, ?)
●Dolphin(Flipper) – It is possible to deduce:
●HasNPaws(Flipper, 4) ...but also
●HasNPaws(Flipper, ?)
●
Modeling default reasoning requires non-monotonic logics
Flipper Mammal
HasNPaws
Dolphin ?
HasNPaws 4
INSTANCE IS-A
From Semantic Nets to Frames
●
70s-80s: need for wider structures than simple
“conceptual nodes”
●
Schemes, Frames, Scripts
●
Concept of “schema” rediscovered
●
Proposed to explain some behavior of human memory [Bartlett, 1932]
– Tendence to recall worse at each recall
– Better recall of important propositions
– Omissions, rationalization, search for a sense
●
And also...
– Theory of linguistic prototypes [Fillmore, 1968]
– Frames in sociology [Goffman, 1974]
– Conceptualizations of Natural Language [Schank, 1973]
Schemes, Frames, Scripts
●
Common features
[Schank & Abelson, Rumelhart, Bower]
●
Structures used by humans to organize knowledge
●
Concern objects/events/situations
●
Useful for understanding
– Create expectations/predictions
●
General structures
●
Affect the way in which we interpret and recall objects and events
– By embedding information on instances or specific events
Frames
●
Originally conceived in Psychology
●
Knowledge structures to represent stereotyped situations and typical objects
●
“framing” = an inevitable process of selective influence on the perception of meanings that an individual attributes to words or sentences
●
Used to define “interpretation schemes”
allowing individuals or groups to “position, perceive, identify and classify” events and facts, this way structuring the meaning, organizing experiences, driving actions [Goffman]
Frames
●
In semantics and computational linguistics
●
A schematization of a situation, state or event
●E.g., “trade” frame
– Using lexical units that recall the situation or event
●E.g., words “buy”, “sell”, “cost”, etc.
– Using semantic roles
●E.g., “buyer”, “seller”, “money”, “good”, “transaction”, etc.
●
Generally invoked by a verb in a sentence
●
Allow to (manually or automatically) annotate the
sentence with the corresponding semantic roles
●
Idea of invoking a known situation to give meaning to a sentence or situation already used in natural language understanding
●
Example: Sentences
– 1. Tom went to the restaurant
– A 2. Asked the waiter a steak
– 3. Paid his bill and went away
– 1. Tom went to the restaurant
– B 2. Asked the dwarf a mouse
– 3. Gave him a coin and went away
●similar in syntactic structure
●semantically consistent due to using the same primitives
●but B has no meaning, while A is understandable because it refers to the Restaurant frame
Frames
●
Example: Frame for concept “cinema”
– Describes the most common stereotype of cinema
●An entrance with ticket counter, a space for waiting, a projection hall
– May describe, in the form of a script, the most common sequence of events that happen in a cinema
●Buying a ticket, waiting for the show to begin, commercials before the movie, then the actual movie, last exiting the place
Expectation
●
Very important phenomenon
– Many things can be explained based on the hypothesis that interaction of intelligent systems with the world is driven by a rich set of expectations of many kinds
●
Each frame establishes a set of expectations (activated by the perceived signals) that can be confirmed or dismissed
●
Aim
– Representing in algorithmic form
●so as to provide statements for a computer
– the set of implicit knowledge and expectations allowing a human to understand
●by making inferences
– sequences of events that can be retraced from consistent narrations
Expectation
●
Examples
– Four metal tips appearing from under a napkin on a furnished table are immediately recognized as a fork
●Seeing the same tips appearing from a book in a library would be different!
– While the concept of “paying the ticket” does not belong to the logical definition of “cinema”
●Not in the same way as the concept of “trunk” belongs to the definition of “tree”
it is well-known, based on experience, that usually going to the cinema involves paying a ticket
Frames
●
Typical features of human intelligence and understanding, to be considered when building programs aimed at simulating them
●
Quick recall of appropriate frames from long-term memory
●
Ability of
– Activating many frames and subframes to understand complex situations
– Finding plausible justifications for situations not corresponding to the expectations of the activated frames
●
Possibility of integrating, modifying or replacing frames that do not fit current situations
Frames
●
Theory of Frames in AI (1975)
●
Proposed as a paradigm to represent knowledge from the real world so as to be exploitable by a computer
●
Formalized by Minsky, inspired by the proposals of research in cognitive psychology, sociology, linguistics
●
Aim: allowing to build a database containing the huge amount of knowledge needed by a system aimed at reproducing “common sense reasoning”
in humans
Frames
●
Frame :
●
An organization of data in memory, a set of inter- related knowledge, useful to represent complex situations or stereotypical events
– E.g., a typical museum, a typical birthday party
●
The result of a conceptualization activity
●
Components:
– Name
– Set of slots (attribute-value pairs)
●Attribute = slot name
●Value = slot filler
Frames
Reasoning Mechanisms
●
When facing a new situation, select a frame from memory
●
Reference to be adapted so as to fit reality, changing its details as needed
●
Each frame associated to several types of information
●
How to use the frame
●
What one may expect to happen later
●
What to do if such expectations are not confirmed
Frames
Reasoning Mechanisms
●
Added value compared to other formalisms:
●
Organization of knowledge similar to that used by humans to acquire and maintain up-to-date knowledge based on their everyday experience
●
Note: the data in a frame might be represented using other knowledge structures
– E.g., semantic nets or production rules
●Actually, frame-based systems often use such formalisms
Frames
●
Usually represented as graphs
●
but Minsky never explicitly refers to semantic nets or is_a categories
●
Frames & Semantic Nets
– Name = Node
– Slot names = Names of outgoing arcs from the node
– Slot Fillers = Nodes at the other end of such arcs
Frames
●
A frame can be thought of as a net of nodes and relationships
●
Top levels are fixed, represent things that are always true about the supposed situation
●
Lower levels have as terminals the slots
– To be filled by assigning specific data values
●Each terminal may specify the conditions under which such assignements must take place
– May be typed or anyway constrained
– Values may be known or unknown
– May have default values
Frames
●
Primitives for handling frames
●
Invocation
– Initial consideration of a frame
●
Determination
– Decide if there is enough evidence to infer an invoked frame
●
Processing
– Filling a slot of a given frame
●
Termination
– Inferring that a given frame is not relevant anymore
Frames
●
When a frame is chosen, an evaluation process takes place to determine if it is appropriate
●
Matching used as a form of inference
– A frame ‘instance’ can be considered as an ‘instance’ of another frame if the former matches the latter
●
Example
– John Smith
●Instance of frame Man – Is Dog_owner
– If there is a matching with an instance of frame Dog_owner
●e.g., with Owner Name
Frames
●
When a frame is not suitable to a situation, it can be
●
‘transformed’
– Perturbation procedures for small changes
●May fix the description
●
or ‘replaced’
– Replacement procedures for significant changes
●Look for new frames whose terminal slots correspond to a sufficient number of terminal slots of the original frame
●
Tasks in the mental process carried out to give meaning to sentences
●
Recognition
– Based on reference to situations that are stereotyped, known, or anyhow that can be reduced to everybody’s experience
– Implies accessing the proper high-level structure
●Encodes the possible interpretations (or predictions)
●
Interpretation
– Implies a simple processing of that structure aimed at retrieving predictions
●
Prediction
●
A kind of loop [Schank]
●
Similar to that of an Intelligent Agent
– 1. John took out the coins
– 2. Put them in the slot
●3a. Dialed the number
●3b. Started the game
●3c. Picked the cup with coffee
– Several schemes and interpretations can be invoked based on the first two sentences
– Misrecognition
●Selection of the stereotyped model of interpretation usually happens at the occurrence of the first event the selection of the schema may fail
– Reinterpretation
●When new facts occur, other schemes are selected until the only compatible scheme occurs
Frames
●
Slot
●
A way to represent knowledge
– Relationship between an object and a value explicitly named, rather than implied by position
●Example: Tom gave Mary the book –give(Tom, book, Mary, past)
–give(subj=Tom, obj=book, obj2=Mary, time=past)
●
Allow to embed relationships between different frames
– This allows mechanisms for linking frames
Frames
●
Theory of Frames is more than just a proposal for a data structure
●
Frame System
●
A schema such that “if there is access to a frame, indexes are automatically created to select other potentially relevant frames” [Minsky]
– Knowledge encoded in packets (Frames)
organized in a networked structure (Frame System) to make retrieval easier,
so that if a frame is invoked, links to other potentially connected frames are activated
Example of a Frame-based System
superclass:
vehicle reg. number producer model owner
truck class: vehicle reg. number
producer model owner tonnage part of basket car
class: vehicle reg. number producer model owner number of
doors 4
horse-power
John’s car class: car
reg. number LV97
producer BMW
model 520
owner John
number of
doors 2
horse-power 150
basket
size 2*3*1.5
material tin
John
age 22
length of driving 2
Frames
●
Frame Systems
●
In addition to providing default values, frames can
– Include knowledge in the form of rules and inherit it from other frames
– Hook procedural information
●Computed through programs
●
Effects of important actions reflected by transformations between the frames of a system
Frames
●
Allow to represent in a single place information collected in different times, places and from different viewpoints
●
Example: Visual scene analysis
– Minsky (“A framework for representing knowledge” AI Memo 306, MIT AI Lab, 1975) to better highlight the features of this representation
– Different frames
●describe a scene from different perspectives
●may have the same terminal slots
–Common slots are to be interpreted as the same physical feature, seen from different perspectives
A B
E Region-of
Transformations from a frame to another amount to
moving from a point to another
A B E
B C E
Right Right
B E
Parallelogram
A E B
Right Above
Left
Is-a
Above
D A B C
Left Left Left
Left Visual Frames
Cube Cube Cube
Spatial frames when moving towards the right
A B E
B C E
Right Right
B E
Frames
●
Possible terminal slots filler expressions
●
Frame name
●
Frame relationships to other frames
●
Slot actual value
– Symbolic, numeric, boolean
●
Slot default value
●
Slot range of values (or restriction on values)
●
Procedural information
Frames
●
Possible terminal slots filler expressions
●
Example: frame Chair
– Specialization-of = Furniture
– No. legs = an integer (DEFAULT = 4)
– Back style = high, low, straight, padded, ...
– No. arms = 0, 1, 2
– Type = normal, wheel, beach, toy, electric, ...
Frames
●
Specialization slot A-kind-of used to create a property inheritance hierarchy among frames
●
The set of frames and slots provides a description of a situation
●
When considering a specific situation,
– a copy of the frame (instantiation) is made
– and added to the short-term memory,
– filling the slots with specifications of the particular situation
Frames
●
Slots and property inheritance
●[ FRAME : Event IS-A : Thing
SLOTS : (Time (A specific time – Default 21st century)) (Place (A place – Default Italy)) ]
●[ FRAME : Action IS-A : Event
SLOTS : (Actor (A person)) ]
●[ FRAME : Walk IS-A : Action SLOTS : (Start (A place))
(Destination (A place)) ]
●[ FRAME : Election IS-A : Event
SLOTS : (Place (A (Country (Headed by (A President))))) ] – Action and Walk inherit default values (21st century and
Italy) from Event
– This is impossible for Election, that requires a specific place on its own, with a type specification that might make the inherited default value illegal
Frames
●
Procedural information in slots
●
Procedures attached to slots may drive the reasoning process
– “IF ADDED (CHANGED) DEMONS” or “TRIGGERS”
●Activated every time the value of a slot has been changed – “IF NEEDED DEMONS”
●Activated on request if the value of a slot “must” be determined
●
Demon
– Procedure that checks if some condition becomes true and, in case, activates an associated process
●IF-THEN structure
– Sometimes, the associated process is also called Demon
Frames
●
Procedural information in slots
●
Example
Temperature sensors Name
Unknown Critical
Value
Unknown Value Unknown
Status Unknown
Get-Value (Self.Name)
IF Self.Status = Alert THEN Sound-Alert Value If-Needed Method
Status If-Added Method
IF Self.Value > Self.CriticalValue THEN
Self.Status = Alert Value If-Added Method
Frames
●
Slots in frames and predicate calculus
●
Frame instance = an object in a domain
– Frame = Predicate
– Slots = Functions that create terms
●
Reasoning
–
If a slot has
associated an●IF-ADDED -> forward inference rule
●IF-NEEDED -> backward inference rule
Frames
●
Slots in frames and predicate calculus
●
Example
– FRAME: Family
SLOTS: (Mother-of (APerson)) (Father-of (APerson)) (Child-of (APerson))
– A specific instance of Family, denoted by identifier Smith-Family, can be asserted using expression
●(Family Smith-Family)
– Slot identifiers can be applied as functions
●(=(Mother-of (Smith-Family) Mary)
●(=(Father-of (Smith-Family) Tom)
●(=(Child-of (Smith-Family) John)
Frames
●
Frame-based development environments often provide an extension to the slot-filler structure through the application of Facets
●
Facet
●
A way to provide extended knowledge about a frame attribute
●
Used to
– Determine the value of an attribute
●VALUE FACETS – Check user queries
●PROMPT FACETS
– Tell the inference engine how to process the attribute
●INFERENCE FACETS
Frames
●
Tools
– Frame Languages
●OWL (Scolovits – MIT 1977)
●FRL (Roberts – Goldstein MIT 1977)
●KRL (Bobrow – Winograd 1977)
●NETL (Fahlman MIT 1979)
–Usually provide record structures with typed fields, embedded in Is-a hierarchies
– Hybrid systems
–Frame systems are sometimes adapted to create rich descriptions or definitions, rather than encoding assertions
●KL-ONE (Brachman 1982)
●KRYPTON (Brachman & Fikes 1983)
●FRAIL (Charniak 1983)
●KODIAK (Wilensky 1984)
●UNIFRAME (Maida 1984) – Shells
●KEE (Knowledge Engineering Environment, Intellicorp 1983)
Frames
●
FrameNet [Lowe, Baker, Fillmore]
●
Resource made up of collections of sentences, syntactically and semantically annotated, organized in frames
●
Frame-based semantics
– The meaning of words stems from the role they play in the conceptual structure of sentences
●
Knowledge structured in 16 general domains
– time, space, communications, cognition, health …
●
6000 lexical elements; 130.000 annotated sentences
●
http://framenet.icsi.berkeley.edu/fndrupal/
Conceptual Dependency
●
Conceptual Dependency [Schank, 1973]
●
A theory on how to represent knowledge about events usually contained in natural language sentences
●
A mechanism to represent and reason about events
●
Objective: representing knowledge so as to
– Ease inference starting from sentences
– Be independent on the original language
– Differently from semantic nets, provide both a structure and a set of primitives to build representations
Conceptual Dependency
●
Representations of actions built starting from a set of primitive acts
– ATRANS Abstract transfer relationship (give)
– PTRANS Transfer of an object’s physical position
– PROPEL Application of a force to an object
– MOVE Movement of a part of the body by its owner
– GRASP Act of grasping an object by an actor
– INGEST Ingestion of an object by an animate being
– EXPEL Expulsion of an object by an animate being
– MTRANS Mental information transfer
– MBUILD Building of new information
– SPEAK Production of sounds
– ATTEND Focus on sensorial stimuli
Conceptual Dependency
●
Conceptualizations representing events may be modified in many ways to provide the
information usually conveyed by tense and mode of a verbal form
●
Set of tenses proposed by Schank
– p past
– f future
– t transition
– ts transition started
– tf transition finished
– k ongoing
– ? interrogative
– / negative
Conceptual Dependency
●
Example
●
I gave John a book
– Arrows = dependence direction
– Double arrow = bidirectional link between agent and action
– p = tense
– ATRANS = primitive action
– o = relation object
– R = receiver
p o R
to
from
I ATRANS book
I John
Scripts
●
Script [Schank & Abelson, 1977]
●
Structure made up of a set of slots
– Each may be associated with
●Information about the ‘types’ of values they may take
●Default values
Scripts
●
Knowledge macrostructures:
Scripts and Human memory
●
Psychologists observed that
– Subjects remember details not present but compliant to the script (e.g., restaurant)
– The adopted script leads to focusing on different information (e.g., Thief vs Restaurant owner)
– Recall with same or different perspective: if change of perspective, recall increases
– Adopted script affects understanding and recall
– Elements compliant with the script more easily recalled
●E.g., PhD students: recall of objects in the Professor’s office Tendency to infer the presence of unseen objects
●E.g., books
and to forget that of little relevant ones
●E.g., umbrella, number of windows, room orientation, etc.
Scripts
●
Knowledge macrostructures
●
Scripts for social situations
– E.g.: going to restaurant
●Sit down, look at menu, order, eat, pay, exit – (actions taken by 73%
of subjects)
●
What are scripts for?
– Create expectations, notice deviations from script
●E.g., he left 100 Euros tip – Cognitive economy
●Not storing all new information
Scripts
●
In reality, event sequences have a logic
●
Causal, temporal, etc.
– Example:
●Entry Conditions
cause
●Events taking place
represent
●Conditions
cause
●Events taking place
cause
●Set of Results
●
A causal chain among events may be represented through a script
Scripts
●
Components
– Entry conditions
●Conditions to be satisfied before the events described in the script can take place
– Result
●Conditions true after the events described in the script took place
– Props
●Objects involved in the events – Roles
●Subjects and individuals involved in the events – Trace
●The variant, particular accepted meaning of a more general scheme (script)
– Scenes
●The actual sequences of events that take place
SCRIPT: Restaurant TRACE: Trattoria PROPS: Tables
Menu F = Food Control Money ROLES: S = Client
W = Waiter C = Cook M = Cashier O = Owner
ENTRY CONDITIONS S is hungry S has money
RESULTS S has less money O has more money S is not hungry S is satisfied (optional)
What might be the primitive actions to represent a scene?
SCENE 1: Entering
S PTRANS S into restaurant S ATTEND look tables S MBUILD where to sit S PTRANS S to table S MOVE S sit
SCENE 2: Ordering
W PTRANS W to table W ATRANS Menu to S S PTRANS Menu to S
S MBUILD choice of F S MTRANS signal to W W PTRANS W to table
S MTRANS (want F) to W
… S MTRANS signals to W
SCENE 3: Eating
C ATRANS F A W
W ATRANS F A S
S INGEST F
SCENE 4: Exiting
W PTRANS W A S
W ATRANS bill A S
S ATRANS money A W
or . . .
Scripts
●
Two ways for activating a script
●
1. Temporary scripts
– Mentioned, reference, but not fundamental
●Sufficient storing a pointer to the script so as to be able to access it subsequently, if necessary
– Example
●“While going to the museum, Susan passed before the restaurant. She liked very much the Picasso exhibition”
–Script associated to restaurant not central: it might be activated later through the pointer
Scripts
●
Two ways for activating a script
●
2. Non-temporary scripts
– Central in description, must be completely activated
●Useful to completely activate them, trying to fill the slots with observed instances
– Script headers
●Preconditions, roles, events, places, props used to indicate that the script should be activated
●E.g.: situations involving at least n elements of the script header
– Example
●“Susan, while going to the museum, passed before a bar. She was hungry and went in. She saw few tables and went there.”
–Presence of trace Bar and precondition hungry enough to activate script Restaurant
Scripts
●
Allow to predict events not explicitly observed
●
Example
– (Suppose script Restaurant has been activated)
– “Yesterday John went to the restaurant. He ordered a steak. Then he asked for the bill and paid. Finally he went home”.
– Question: “Did John eat yesterday?”
– Answer: “Yes”
●Even if the fact is not explicitly expressed
Scripts
●
Provide a way for building a consistent interpretation, given a set of observed facts
●
Example
– “Yesterday John went out for lunch. After sitting in the place, he called the waiter. The waiter brought him a menu. John ordered his lunch”.
– Question: “Why did the waiter bring a menu?”
– The script provides two possible answers:
●John asked for the menu
●John had to decide what to eat
Scripts
●
Allow to focus on exceptional situations (unusual events), highlighting where the description of observed facts departs from the standard sequence of events
●
When the script is ‘broken’ it cannot be used anymore to predict events
– Example
●“John went to the restaurant, he sat at a table. He waited for the waiter for a long time. The waiter did not come. John stood up and went away.”
●Cannot say if John consulted a menu or not
●
Systems using scripts
●
SAM (Cullingford, 1981)
Further Readings
●
N.J. Nilsson “Artificial Intelligence”
●
Ch. 18
●
E. Rich, K. Knight “Artificial Intelligence”, McGraw-Hill
●
Ch. 9-10
●
Marvin Minsky “A Framework for Representing Knowledge”, MIT-AI Laboratory Memo 306, June, 1974
●
Reprinted in P. Winston (Ed.) “The Psychology of Computer Vision”, McGraw-Hill, 1975
– Shorter versions in
●J. Haugeland (Ed.) “Mind Design”, MIT Press, 1981
●Allan Collins and Edward E. Smith (Eds.) “Cognitive Science”, Morgan-Kaufmann, 1992