File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Bottom-up learning of a phonetic system using an autoencoder
Title | Bottom-up learning of a phonetic system using an autoencoder |
---|---|
Authors | |
Issue Date | 26-May-2023 |
Citation | Hanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023), 2023 How to Cite? |
Abstract | This study investigates an autoencoder model’s effectiveness in learning phonemes and distinctive features from unsegmented, non-transcribed wave data, mirroring infants’ early language acquisition stages. The results of the experiment on Mandarin and English show that features can be learned through repeated projection of sounds to a hidden space and reconstruction back to their original form without prior knowledge of segmentation. The acquired knowledge is represented by different distributions in the hidden space. The model was capable of clustering segments of the same phoneme and projecting different phonemes to separate regions in the hidden space. However, the model struggled to cluster allophones closely together, which may indicate the boundary between the contributions of bottom-up and top-down information. Overall, the study suggests that sound knowledge can be learned to a certain degree through unsupervised learning techniques that do not require labeled data or prior knowledge of segmentation, and that this approach may provide insights into the early stages of human language acquisition. |
Persistent Identifier | http://hdl.handle.net/10722/329225 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Tan, Frank Lihui | - |
dc.contributor.author | Do, Youngah | - |
dc.date.accessioned | 2023-08-09T02:19:19Z | - |
dc.date.available | 2023-08-09T02:19:19Z | - |
dc.date.issued | 2023-05-26 | - |
dc.identifier.citation | Hanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023), 2023 | - |
dc.identifier.uri | http://hdl.handle.net/10722/329225 | - |
dc.description.abstract | <p>This study investigates an autoencoder model’s effectiveness in learning phonemes and distinctive features from unsegmented, non-transcribed wave data, mirroring infants’ early language acquisition stages. The results of the experiment on Mandarin and English show that features can be learned through repeated projection of sounds to a hidden space and reconstruction back to their original form without prior knowledge of segmentation. The acquired knowledge is represented by different distributions in the hidden space. The model was capable of clustering segments of the same phoneme and projecting different phonemes to separate regions in the hidden space. However, the model struggled to cluster allophones closely together, which may indicate the boundary between the contributions of bottom-up and top-down information. Overall, the study suggests that sound knowledge can be learned to a certain degree through unsupervised learning techniques that do not require labeled data or prior knowledge of segmentation, and that this approach may provide insights into the early stages of human language acquisition.</p> | - |
dc.language | eng | - |
dc.relation.ispartof | Hanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023) (26/05/2023-27/05/2023, Hanyang University, Seoul, Korea) | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.title | Bottom-up learning of a phonetic system using an autoencoder | - |
dc.type | Conference_Paper | - |