Paper

Learning Phonotactics from Linguistic Informants

Authors
  • Canaan Breiss (University of Southern California)
  • Alexis Ross (Massachusetts Institute of Technology)
  • Amani Maina-Kilaas (Massachusetts Institute of Technology)
  • Roger Levy (Massachusetts Institute of Technology)
  • Jacob Andreas (Massachusetts Institute of Technology)

Abstract

We propose an interactive approach to language learning that utilizes linguistic acceptability judgments from an informant (a competent language user) to learn a grammar. Given a grammar formalism and a framework for synthesizing data, our model iteratively selects or synthesizes a data-point according to one of a range of information-theoretic policies, asks the informant for a binary judgment, and updates its own parameters in preparation for the next query. We demonstrate the effectiveness of our model in the domain of phonotactics, the rules governing what kinds of sound-sequences are acceptable in a language, and carry out two experiments, one with typologically-natural linguistic data and another with a range of procedurally-generated languages. We find
that the information-theoretic policies that our model uses to select items to query the informant achieve sample efficiency comparable to, and sometimes greater than, fully supervised approaches.

Keywords: interactive learning, data synthesis, phonotactics, vowel harmony, language learning

How to Cite:

Breiss, C., Ross, A., Maina-Kilaas, A., Levy, R. & Andreas, J., (2024) “Learning Phonotactics from Linguistic Informants”, Society for Computation in Linguistics 7(1), 20–31. doi: https://doi.org/10.7275/scil.2126

Downloads:
Download PDF

236 Views

61 Downloads

Published on
24 Jun 2024
Peer Reviewed