Type of Document Dissertation Author Pangburn, Brian Edward Author's Email Address firstname.lastname@example.org URN etd-1011102-170733 Title Experience-Based Language Acquisition: A Computational Model of Human Language Acquisition Degree Doctor of Philosophy (Ph.D.) Department Computer Science Advisory Committee
Advisor Name Title S. Sitharama Iyengar Committee Chair Doris L. Carver Committee Member Jianhua Chen Committee Member Robert C. Mathews Committee Member Michael Hegarty Dean's Representative Keywords
- computer vision
- cross-situational learning
- language acquisition
- event recognition
- natural language processing
- computational linguistics
Date of Defense 2002-10-04 Availability unrestricted AbstractAlmost from the very beginning of the digital age, people have sought better ways to communicate with computers. This research investigates how computers might be enabled to understand natural language in a more humanlike way. Based, in part, on cognitive development in infants, we introduce an open computational framework for visual perception and grounded language acquisition called Experience-Based Language Acquisition (EBLA). EBLA can “watch” a series of short videos and acquire a simple language of nouns and verbs corresponding to the objects and object-object relations in those videos. Upon acquiring this protolanguage, EBLA can perform basic scene analysis to generate descriptions of novel videos.
The general architecture of EBLA is comprised of three stages: vision processing, entity extraction, and lexical resolution. In the vision processing stage, EBLA processes the individual frames in short videos, using a variation of the mean shift analysis image segmentation algorithm to identify and store information about significant objects. In the entity extraction stage, EBLA abstracts information about the significant objects in each video and the relationships among those objects into internal representations called entities. Finally, in the lexical acquisition stage, EBLA extracts the individual lexemes (words) from simple descriptions of each video and attempts to generate entity-lexeme mappings using an inference technique called cross-situational learning. EBLA is not primed with a base lexicon, so it faces the task of bootstrapping its lexicon from scratch.
The performance of EBLA has been evaluated based on acquisition speed and accuracy of scene descriptions. For a test set of simple animations, EBLA had average acquisition success rates as high as 100% and average description success rates as high as 96.7%. For a larger set of real videos, EBLA had average acquisition success rates as high as 95.8% and average description success rates as high as 65.3%. The lower description success rate for the videos is attributed to the wide variance in entities across the videos.
While there have been several systems capable of learning object or event labels for videos, EBLA is the first known system to acquire both nouns and verbs using a grounded computer vision system.
Filename Size Approximate Download Time (Hours:Minutes:Seconds)
28.8 Modem 56K Modem ISDN (64 Kb) ISDN (128 Kb) Higher-speed Access Pangburn_dis.pdf 1.82 Mb 00:08:26 00:04:20 00:03:47 00:01:53 00:00:09