Gaja JaroszJoe PaterPrickett, Brandon2024-04-262024-04-262021-052021-0510.7275/jchq-cm44https://hdl.handle.net/20.500.14394/18517This dissertation tests sequence-to-sequence neural networks to see whether they can simulate human phonological learning and generalization in a number of artificial language experiments. These experiments and simulations are organized into three chapters: one on opaque interactions, one on computational complexity in phonology, and one on reduplication. The first chapter focuses on two biases involving interactions that have been proposed in the past: a bias for transparent patterns and a bias for patterns that maximally utilize all of the processes in a language. The second chapter looks at harmony patterns of varying complexity to see whether both Formal Language Theory and the sequence‑to‑sequence network correctly predict which kinds of patterns humans most easily learn. Finally, the third chapter investigates reduplication—a pattern that involves copying all or part of a word. These simulations focus on the model’s ability to generalize reduplication to novel words and compares these results to past experiments involving humans. The conclusions drawn from these three chapters suggest that the kind of language‑specific representations and explicit biases used in past models are not necessary to capture the behavior of humans in the relevant experiments. Instead, the ability of the network to capture these various behaviors is attributed to two characteristics of its architecture: its recurrent connections, which provide a limited memory through time, and the fact that it is made up of two separate mechanisms (an encoder and a decoder) which require forms to be processed into an intermediate representation.phonologycomputationalexperimentalopacityharmonyreduplicationComputational LinguisticsPhonetics and PhonologyLEARNING PHONOLOGY WITH SEQUENCE-TO-SEQUENCE NEURAL NETWORKSdissertationhttps://orcid.org/0000-0001-9217-2130