Disentangling Latent Representations for Interpretability and Controllability: Summary

Written by textmodels | Published 2024/06/01
Tech Story Tags: llm-natural-supervision | llm-self-supervision | llm-language-pretraining | disentangling-semantics | sentence-representations | syntactic-exemplar | latent-representations | latent-variable-model

TLDRIn this study, researchers disentangle latent representations using naturally-occurring structures of paired data.via the TL;DR App

Author:

(1) Mingda Chen.

Table of Links

5.3 Summary

In this chapter, we demonstrated the utility of our proposed latent-variable framework in the context of representation learning (Section 5.1) and controllable generation (Section 5.2). In both cases, we leveraged the structures of paired data to disentangle semantics and syntax in sentence representations. We found that the syntactic and semantic latent variables showed desirable characteristics. For controlled generation, we provided human-annotated evaluation sets to promote future research in this direction. In addition, in a follow-up work, we showed that the multi-task, latent-variable framework can generalize to bilingual text corpora (Chen et al., 2020b).

This paper is available on arxiv under CC 4.0 license.


Written by textmodels | We publish the best academic papers on rule-based techniques, LLMs, & the generation of text that resembles human text.
Published by HackerNoon on 2024/06/01