Language learning with some negative information

Ganesh Baliga, John Case, Sanjay Jain

Research output: Contribution to journalArticlepeer-review

20 Scopus citations

Abstract

Gold-style language learning is a formal theory of learning from examples by algorithmic devices called learning machines. Originally motivated by child language learning, it features the algorithmic synthesis (in the limit) of grammars for formal languages from information about those languages. In traditional Gold-style language learning, learning machines are not provided with negative information, i.e., information about the complements of the input languages. We investigate two approaches to providing small amounts of negative information and demonstrate in each case a strong resulting increase in learning power. Finally, we show that small packets of negative information also lead to increased speed of learning. This result agrees with a psycholinguistic hypothesis of McNeill correlating the availability of parental expansions with the speed of child language development.

Original languageEnglish (US)
Pages (from-to)273-285
Number of pages13
JournalJournal of Computer and System Sciences
Volume51
Issue number2
DOIs
StatePublished - Oct 1995
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • General Computer Science
  • Computer Networks and Communications
  • Computational Theory and Mathematics
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Language learning with some negative information'. Together they form a unique fingerprint.

Cite this