Phonotactic learning with neural language models
- Connor Mayer (University of California, Los Angeles)
- Max Nelson (University of Massachusetts, Amherst)
Abstract
Computational models of phonotactics share much in common with language models, which assign probabilities to sequences of words. While state of the art language models are implemented using neural networks, phonotactic models have not followed suit. We present several neural models of phonotactics, and show that they perform favorably when compared to existing models. In addition, they provide useful insights into the role of representations on phonotactic learning and generalization. This work provides a promising starting point for future modeling of human phonotactic knowledge.
Keywords: phonology, phonotactics, neural networks, sonority sequencing
How to Cite:
Mayer, C. & Nelson, M., (2020) “Phonotactic learning with neural language models”, Society for Computation in Linguistics 3(1), 149-159. doi: https://doi.org/10.7275/g3y2-fx06
Downloads:
Download PDF