File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: What Should Embeddings Embed? Autoregressive Models Represent Latent Generating Distributions

TitleWhat Should Embeddings Embed? Autoregressive Models Represent Latent Generating Distributions
Authors
Issue Date2025
Citation
Transactions on Machine Learning Research, 2025, v. July-2025 How to Cite?
AbstractAutoregressive language models have demonstrated a remarkable ability to extract latent structure from text. The embeddings from large language models have been shown to capture aspects of the syntax and semantics of language. But what should embeddings represent? We show that the embeddings from autoregressive models correspond to predictive sufficient statistics. By identifying settings where the predictive sufficient statistics are interpretable distributions over latent variables, including exchangeable models and latent state models, we show that embeddings of autoregressive models encode these explainable quantities of interest. We conduct empirical probing studies to extract information from transformers about latent generating distributions. Furthermore, we show that these embeddings generalize to out-of-distribution cases, do not exhibit token memorization, and that the information we identify is more easily recovered than other related measures. Next, we extend our analysis of exchangeable models to more realistic scenarios where the predictive sufficient statistic is difficult to identify by focusing on an interpretable subcomponent of language, topics. We show that large language models encode topic mixtures inferred by latent Dirichlet allocation (LDA) in both synthetic datasets and natural corpora.
Persistent Identifierhttp://hdl.handle.net/10722/367863

 

DC FieldValueLanguage
dc.contributor.authorZhang, Liyi-
dc.contributor.authorLi, Michael Y.-
dc.contributor.authorThomas Mccoy, R.-
dc.contributor.authorSumers, Theodore R.-
dc.contributor.authorZhu, Jian Qiao-
dc.contributor.authorGriffiths, Thomas L.-
dc.date.accessioned2025-12-19T08:00:03Z-
dc.date.available2025-12-19T08:00:03Z-
dc.date.issued2025-
dc.identifier.citationTransactions on Machine Learning Research, 2025, v. July-2025-
dc.identifier.urihttp://hdl.handle.net/10722/367863-
dc.description.abstractAutoregressive language models have demonstrated a remarkable ability to extract latent structure from text. The embeddings from large language models have been shown to capture aspects of the syntax and semantics of language. But what should embeddings represent? We show that the embeddings from autoregressive models correspond to predictive sufficient statistics. By identifying settings where the predictive sufficient statistics are interpretable distributions over latent variables, including exchangeable models and latent state models, we show that embeddings of autoregressive models encode these explainable quantities of interest. We conduct empirical probing studies to extract information from transformers about latent generating distributions. Furthermore, we show that these embeddings generalize to out-of-distribution cases, do not exhibit token memorization, and that the information we identify is more easily recovered than other related measures. Next, we extend our analysis of exchangeable models to more realistic scenarios where the predictive sufficient statistic is difficult to identify by focusing on an interpretable subcomponent of language, topics. We show that large language models encode topic mixtures inferred by latent Dirichlet allocation (LDA) in both synthetic datasets and natural corpora.-
dc.languageeng-
dc.relation.ispartofTransactions on Machine Learning Research-
dc.titleWhat Should Embeddings Embed? Autoregressive Models Represent Latent Generating Distributions-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-105011731541-
dc.identifier.volumeJuly-2025-
dc.identifier.eissn2835-8856-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats