1 / 63

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

INTRODUCTION TO ARTIFICIAL INTELLIGENCE. Massimo Poesio LECTURE 7 Psychological theories of concepts: Prototype theory, Image schemas. Evidence for the taxonomic theory of concepts. Collins & Quillian, 1969, 1970

zariel
Télécharger la présentation

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo PoesioLECTURE 7Psychological theories of concepts: Prototype theory, Image schemas

  2. Evidence for the taxonomic theory of concepts • Collins & Quillian, 1969, 1970 • Evidence: When people do a semantic verification task, you see evidence of a hierarchy (response times are correlated with the number of links).

  3. Semantic networks • Collins & Quillian, 1970: Collins & Quillian (1970, p. 306; reporting data from Collins & Quillian, 1969)

  4. Semantic network models (Collins & Quillian, 1970) • Concepts are organized as a set of nodes and links between those nodes. • Nodes hold concepts. For example, you would have a node for RED, FIRE, FIRETRUCK, etc. Each concept you know would have a node. • Links connect the nodes and encode relationships between them. Two kinds: • Superset/subset: Essentially categorize. For example, ROBIN is a subset of BIRD. • Property: Labeled links that explain the relationship between various properties and nodes. For example, WINGS and BIRD would be connected with a “has” link.

  5. Semantic network models Collins & Quillian (1970, p. 305)

  6. Semantic network models • The nodes are arranged in a hierarchy and distance (number of links traveled) matters. The farther apart two things are, the less related they are. This is going to be important in predicting performance. • The model has a property called cognitive economy. Concepts are only stored once at the highest level to which they apply (e.g., BREATHES is stored at the ANIMAL node instead of with each individual animal). This was based on the model’s implementation on early computers and may not have a brain basis. Also, not necessarily crucial for the model.

  7. PROBLEMS FOR THE CLASSICAL THEORY OF CONCEPTS • The view dominant since Aristotle that knowledge is organized around concepts whose definitions provide necessary and sufficient conditions in terms of genus and differentia (ie in terms of a taxonomic organization) has been challenged first from within Philosophy, then from within Psychology

  8. PLATO’S PROBLEM For many concepts, there simply aren’t any definitions” (LM p.14) A theory that correctly describes the behavior of perhaps three hundred words has been asserted to correctly describe the behavior of the tens of thousands of general names (Putnam)

  9. WITTGENSTEIN’s EXAMPLE: ‘GAME’ • What is common to all games? • Are they all ‘amusing’? • Cfr. chess • Or is there always winning and losing? • Counterex: child throwing his ball at the wall • Look at the parts played by skill and luck • “I can think of no better expression that FAMILY RESEMBLANCE” • ‘games form a family’

  10. PROBLEMATIC EVIDENCE FROM PSYCHOLOGY • Typicality effects • Is a tomato a vegetable or a fruit? • ‘Is this art?’ • Evidence that IS-A may have different properties – eg., failures of transitivity • If A is a B and B is a C, is A a C? • Evidence that categorization may not be based on necessary and sufficient conditions • Basic level effects

  11. Typicality effects • Rips, Shoben & Smith (1973): Typicality seems to influence responses. • Technically, “a robin is a bird” and “a chicken is a bird” are each one link. But, the robin is more typical. What is that in the network? Shows up in reaction times as well.

  12. Typicality ratings Smith, Shoben, & Rips (1974, p. 218)

  13. Typicality effects • With the hierarchy: “A horse is an animal” is faster than “A horse is a mammal” which violates the hierarchy. • A chicken is more typical of animal than a robin, a robin is more typical of bird than a chicken. How can a network account for this (Smith, Shoben, & Rips, 1974)?

  14. Typicality effects • Answering “no”: You know the answer to a questions is “no” (e.g., “a bird has four legs”) because the concepts are far apart in the network. But, some “no” responses are really fast (e.g., “a bird is a fish”) and some are really slow (e.g., “a whale is a fish”). The reason for this isn’t obvious in the model. • “Loosely speaking, a bat is a bird” is true, but how does a network model do it (Smith, Shoben, & Rips, 1974)?

  15. Agreement on typicality judgments • Agreement on typicality judgments • (‘think of a fish, any fish’) • Rosch (1975): very high correlation (.97) between subjects’s typicality rankings for 10 categories

  16. TYPICALITY JUDGMENTS 1 chair 1 sofa 3 couch 3 table 5 easy chair 6 dresser 6 rocking chair 8 coffee table 9 rocker 10 love seat 11 chest of drawers 12 desk 13 bed ... ... 22 bookcase 27 cabinet 29 bench 31 lamp 32 stool 35 piano 41 mirror 42 tv 44 shelf 45 rug 46 pillow 47 wastebasket 49 sewing machine 50 stove 54 refrigerator 60 telephone

  17. Other typicality effects • Learning: typical items learned before atypical ones (Rosch Simpson & Miller 1976) • Learning is faster if subjects are taught on typical items • Typicality affects speed of inference • Rips 1975: • Garrod & Sanford 1977: faster reading time for “The bird came in through the front door” when ROBIN than when GOOSE

  18. Feature models • Smith, Shoben, & Rips, 1974: • Concepts are clusters of semantic features. There are two kinds: • Distinctive features: Core parts of the concept. They must be present to be a member of the concept, they’re the defining features. For example, WINGS for BIRD. • Characteristic features: Typically associated with the concept, but not necessary. For example, CAN FLY for BIRD.

  19. Feature models Smith, Shoben, & Rips (1974, p. 216)

  20. Feature models • Some examples:

  21. Feature models • Why characteristic features? Various evidence, such as hedges: Smith, Shoben, & Rips (1974, p. 217)

  22. Feature models • Why characteristic features? Various evidence, such as hedges: • OK: • A robin is a true bird. • Technically speaking, a chicken is a bird. • Feels wrong: • Technically speaking, a robin is a bird. • A chicken is a true bird. • The answer depends on the kinds of feature overlap.

  23. Feature models • Answering a semantic verification question is a two-step process. • Compare on all features. If there is a lot of overlap it’s an easy “yes.” If there is almost no overlap, it’s an easy “no.” In the middle, go to step two. • Compare distinctive features. This involves an extra stage and should take longer.

  24. Smith, Shoben, & Rips (1974, p. 222)

  25. Feature models • Some questions:

  26. Feature models • Can account for: • Typicality effects: One step for more typical members, two steps for less typical members, that explains the time difference. • Answering “no”: Why are “no” responses different? Depends on the number of steps (feature overlap). • Hierarchy: Since it isn’t a hierarchy but similarity, we can understand why different types of decisions take different amounts of time.

  27. Feature models • Problem: How do we get the distinctive features? • What makes something a game? What makes someone a bachelor? How about cats? • How many of the features of a bird can you lose and still have a bird? • The distinction between defining and characteristic features addresses this somewhat, but it is still a problem in implementation.

  28. FEATURE NORMS • Psychologists have been collecting concept features from subjects at least since Rosch and Mervis (1975) • Different methodologies used (from free association to very tightly controlled) • Three such databases currently available • Garrard et al (2001) - GA • Vinson and Vigliocco (2004) - VV • McRae et al (2005) – MCRA - the largest, also classified

  29. SPEAKER-GENERATED FEATURES (VINSON AND VIGLIOCCO)

  30. FEATURE NORMS (GARRARD)

  31. What makes an item typical? Rosch & Mervis 1975 • Items are typical when they have HIGH FAMILY RESEMBLANCE with members of the category: • Typical items have many of the attributes of members • Do not have properties of nonmembers • Irrespective of frequency: ORIOLE vs CHICKEN • Evidence 1: checked that subjects agree on typicality for several natural categories • Asked subjects to list attributes (actually, check) • Weighed each attribute by how many items it occurred with within the category • ‘SCORE’ indicates how many common features • Found that score highly predictive of typicality (.84-.91) Five most typical ‘furniture’ (CHAIR, SOFA, TABLE, DRESSER, DESK) have 13 features in common Five least typical (CLOCK, PICTURE, CLOSET, VASE, TELEPHONE) had 2 attributes in common

  32. Rosch and Mervis 1975 (2) • Evidence 2: non-typical elements have more features in common with other categories • Evidence 3: speed of learning with artificial stimuli belonging to 2 classes • Items with more features in common with family easier to learn • Items with more features in common with contrast category harder to learn

  33. ‘Fuzzy’ or ‘graded’ categorization • A necessary and sufficient definition should pick up all the category members and none of the non-members • But this is not what happens: • Hampton (1979): no clear division between members and non-members of 8 categories • Kitchen utensils: SINK? SPONGE? • Vegetables: TOMATOES? GOURDS?

  34. CATEGORIES AS CLUSTERS CHICKEN GOOSE ORIOLE ROBIN OSTRICH

  35. PROTOTYPE THEORY • The dominant theory of concepts in Psychology, developed by E. Rosch and collaborators in the ’70s

  36. Prototype theory • Wittgenstein’s examination of game • Generally necessary that all games be amusing, not sufficient since many things are amusing • Board games, ball games, card games, etc. have different objectives, call on different skills and motor routines - categories normally not definable in terms of necessary and sufficient features

  37. Prototypes and Multidimensional Concept Spaces • A Concept is represented by a prototypical item = central tendency (e.g. location P below) A new exemplar is classified based on its similarity to the prototype

  38. Is this a “cat”? Is this a “chair”? Is this a “dog”?

  39. Prototype theory in a nutshell • Certain members of a category are prototypical – or instantiate the prototype • Categories form around prototypes; new members added on basis of resemblance to prototype • No requirement that a property or set of properties be shared by all members • Features/attributes generally gradable • Category membership a matter of degree • Categories do not have clear boundaries

  40. Prototype theory • Certain members of a category are prototypical – or instantiate the prototype Category members are not all equal a robin is a prototypical bird, but we may not want to say it is the prototype, rather it instantiates (manifests) the prototype or ideal -- it exhibits many of the features that the abstract prototype does “It is conceivable that the prototype for dog will be unspecified for sex; yet each exemplar is necessarily either male or female.” (Taylor)

  41. Prototypes and typicality Effects • typical • is robin a bird? • is dog a mammal? • is diamond a precious stone? • atypical • is ostrich a bird? • is a whale a mammal? • is turquoise a precious stone? slower verification times for atypical items

  42. Prototype theory 3. No requirement that a property or set of properties be shared by all members -- no criterial attributes • Category where a set of necessary and sufficient attributes can be found is the exception rather than the rule • Labov household dishes experiment • Necessary that cups be containers, not sufficient since many things are containers • Cups can’t be defined by material used, shape, presence of handles or function • Cups vs. bowls is graded and context dependent

  43. Graded Structure • Typical items are similar to a prototype • Typicality effects are naturally predicted atypical P typical

  44. Problem with Prototype Models • All information about individual exemplars is lost • category size • variability of the exemplars • correlations among attributes (e.g., only small birds sing)

  45. Exemplar Models • category representation consists of storage of a number of category members • New exemplars are compared to known exemplars – most similar item will influence classification the most dog ?? cat dog dog cat dog cat

  46. Exemplar Models • Model can explain • Prototype classification effects • Prototype is similar to most exemplars from a category • Graded typicality • How many exemplars is new item similar to? • Effects of variability • Overall, compared to prototype models, exemplar models better explain data from categorization experiments (Storms et al., 2000)

  47. THE BASIC LEVEL HYPOTHESIS • Recent psychological evidence also challenges the taxonomic view developed in work on ontologies in another respect: • Not all levels the same

  48. Superordinate level Superordinate Furniture Preferred level BASIC LEVEL Basic Chair Subordinate level Subordinate Windsor

  49. NOT ALL LEVELS ARE THE SAME • Brown (1958): parents would call a sheepdog DOG or DOGGIE rather than ANIMAL or SHEEPDOG • Roschet al (1976): • Subjects name more attributes for basic concepts (PANTS) than for superordinate (CLOTHING) or subordinate (JEANS) • Subjects are faster at verifying that a picture was a TABLE than it was a PIECE OF FURNITURE of a KITCHEN TABLE

  50. Basic Level and Expertise Dog and bird experts identifying dogs and birds at different levels Experts make subordinate as quickly as basic categorizations

More Related