Unsupervised Learning of Structured Representations via Closed-Loop Transcription
Shengbang Tong\textsuperscript{\rm 1 \quad Xili Dai\rm 1,2 * \quad Yubei Chen\rm 3 \quad Mingyang Li\rm 5 \quad Zengyi Li\rm 1 \quad Brent Yi\rm 1 \quad }, Yann LeCun\textsuperscript{\rm 3,4\quad Yi Ma\rm 1,5 }, \textsuperscript{\rm 1University of California, Berkeley \quad \rm 2Hong Kong University of Science and Technology (Guangzhou)}, \textsuperscript{\rm 3Center for Data Science, New York University\quad \rm 4Courant Inst., New York University }, % \textsuperscript{\rm 4Courant Inst., New York University\quad \rm 5Tsinghua-Berkeley Shenzhen Institute (TBSI) \quad }
Abstract
-2mm This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes. While most existing unsupervised learning approaches focus on a representation for only one of these two goals, we show that a unified representation can enjoy the mutual benefits of having both. Such a representation is attainable by generalizing the recently proposed closed-loop transcription framework, known as CTRL, to the unsupervised setting. This entails solving a constrained maximin game over a rate reduction objective that expands features of all samples while compressing features of augmentations of each sample. Through this process, we see discriminative low-dimensional structures emerge in the resulting representations. Under comparable experimental conditions and network complexities, we demonstrate that these structured representations enable classification performance close to state-of-the-art unsupervised discriminative representations, and conditionally generated image quality significantly higher than that of state-of-the-art unsupervised generative models. Source code can be found at https://github.com/Delay-Xili/uCTRL{{https://github.com/Delay-Xili/uCTRL}}.
Unsupervised Representation Learning via Closed-Loop Transcription
Shengbang Tong 1 ∗ Xili Dai 1,2 * † Yubei Chen 3 Mingyang Li 5 Zengyi Li 1 Brent Yi 1 Yann LeCun 3,4 Yi Ma 1,5
1 University of California, Berkeley 2 Hong Kong University of Science and Technology (Guangzhou)
3 Center for Data Science, New York University 4
5
Courant Inst., New York University Tsinghua-Berkeley Shenzhen Institute (TBSI)
This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes. While most existing unsupervised learning approaches focus on a representation for only one of these two goals, we show that a unified representation can enjoy the mutual benefits of having both. Such a representation is attainable by generalizing the recently proposed closed-loop transcription framework, known as CTRL, to the unsupervised setting. This entails solving a constrained maximin game over a rate reduction objective that expands features of all samples while compressing features of augmentations of each sample. Through this process, we see discriminative low-dimensional structures emerge in the resulting representations. Under comparable experimental conditions and network complexities, we demonstrate that these structured representations enable classification performance close to state-of-the-art unsupervised discriminative representations, and conditionally generated image quality significantly higher than that of state-of-the-art unsupervised generative models. Source code can be found at https://github.com/Delay-Xili/uCTRL.
Introduction
In the past decade, we have witnessed an explosive development in the practice of machine learning, particularly with deep learning methods. A key driver of success in practical applications has been marvelous engineering endeavors, often focused on fitting increasingly large deep networks to input data paired with task-specific sets of labels. Brute-force approaches of this nature, however, exert tremendous demands on hand-labeled data for supervision and computational resources for training and inference. As a result, an increasing amount of attention has been directed toward using selfsupervised or unsupervised techniques to learn representations that can not only learn without human annotation effort, but also be shared across downstream tasks.
Discriminative versus Generative. Tasks in unsupervised learning are typically separated into two categories. Discriminative ones frame high-dimensional observations as inputs, from which lowdimensional class or latent information can be extracted, while generative ones frame observations as generated outputs, which should often be sampled given some semantically meaningful conditioning.
Unsupervised learning approaches targeted at discriminative tasks are mainly based on a key idea: to pull different views from the same instance closer while enforcing a non-collapsed representation by either contrastive learning techniques (Chen et al., 2020b; He et al., 2020; Grill et al., 2020a), covariance regularization methods (Bardes et al., 2021; Zbontar et al., 2021), or using architecture design (Chen & He, 2020; Grill et al., 2020b). Their success is typically measured by the accuracy of a simple classifier (say a shallow network) trained on the representations that they produce, which have progressively improved over the years. Representations learned from these approaches, however, do not emphasize much about the intrinsic structure of the data distribution, and have not demonstrated success for generative purposes.
In parallel, generative methods like GANs (Goodfellow et al., 2014) and VAEs (Kingma & Welling, 2013) have also been explored for unsupervised learning. Although generative methods have made
∗ Equal contribution
† Work done during visiting at Berkeley
striking progress in the quality of the sampled or autoencoded data, when compared to the aforementioned discriminative methods, representations learned with these approaches demonstrate inferior performance in classification.
Toward A Unified Representation? The disparity between discriminative and generative approaches in unsupervised learning, contrasted against the fundamental goal of learning representations that are useful across many tasks, leads to a natural question that we investigate in this paper: in the unsupervised setting, is it possible to learn a unified representation that is effective for both discriminative and generative purposes? Further, do they mutually benefit each other? Concretely, we aim to learn a structured representation with the following two properties:
- The learned representation should be discriminative, such that simple classifiers applied to learned features yield high classification accuracy.
- The learned representation should be generative, with enough diversity to recover raw inputs, and structure that can be exploited for sampling and generating new images.
The fact that human visual memory serves both discriminative tasks (for example, detection and recognition) and generative or predictive tasks (for example, via replay) (Keller & Mrsic-Flogel, 2018; Josselyn & Tonegawa, 2020; Ven et al., 2020) indicates that this goal is achievable. Beyond being possible, these properties are also highly practical - successfully completing generative tasks like unsupervised conditional image generation (Hwang et al., 2021), for example, inherently requires that learned features for different classes be both structured for sampling and discriminative for conditioning. On the other hand, the generative property can serve as a natural regularization to avoid representation collapse.
Closed-Loop Transcription via a Constrained Maximin Game. The class of linear discriminative representations (LDRs) has recently been proposed for learning diverse and discriminative features for multi-class (visual) data, via optimization of the rate reduction objective (Chan et al., 2022). In the supervised setting, these representations have been shown to be be both discriminative and generative if learned in a closed-loop transcription framework via a maximin game over the rate reduction utility between an encoder and a decoder (Dai et al., 2022). Beyond the standard joint learning setting, where all classes are sampled uniformly throughout training, the closed-loop framework has also been successfully adapted to the incremental setting (Tong et al., 2022), where the optimal multi-class LDR is learned one class at a time. In the incremental (supervised) learning setting, one solves a constrained maximin problem over the rate reduction utility which keeps learned memory of old tasks intact (as constraints) while learning new tasks. It has been shown that this new framework can effectively alleviate the catastrophic forgetting suffered by most supervised learning methods.
Contributions. In this work, we show that the closed-loop transcription framework proposed for learning LDRs in the supervised setting (Chan et al., 2022) can be adapted to a purely unsupervised setting. In the unsupervised setting, we only have to view each sample and its augmentations as a 'new class' while using the rate reduction objective to ensure that learned features are both invariant to augmentation and self-consistent in generation; this leads to a constrained maximin game that is similar to the one explored for incremental learning (Tong et al., 2022). Our overall approach is illustrated in Figure 1.
As we experimentally demonstrate in Section 4, our formulation benefits from the mutual benefits of both discriminative and generative properties. It bridges the gap between two formerly distinct set of methods: by standard metrics and under comparable experimental conditions, it enables classification performance on par with and unsupervised conditional generative quality significantly higher than state-of-the-art techniques. Coupled with evidence from prior work, this suggests that the closed-loop transcription through the (constrained) maximin game between the encoder and decoder has the potential to offer a unifying framework for both discriminative and generative representation learning, across supervised, incremental, and unsupervised settings.
Related Work
Our work is mostly related to three categories of unsupervised learning methods: (1) self-supervised learning via discriminative models, (2) self-supervised learning via generative models, and (3) unsupervised conditional image generation. Table 1 compares the capabilities of models learned by various representative unsupervised learning methods.

Figure 1: Overall framework of closed-loop transcription for unsupervised learning. Two additional constraints are imposed on the Binary-CTRL method proposed in prior work (Dai et al., 2022): 1) self-consistency for sample-wise features z i and ˆ z i , say z i ≈ ˆ z i ; and 2) invariance/similarity among features of augmented samples z i and z i a , say z i ≈ z i a = f ( τ ( x i ) , θ ) , where x i a = τ ( x i ) is an augmentation of sample x i via some transformation τ ( · ) .
Table 1: Comparison of the downstream task capabilities of different unsupervised learning methods. UCIG refers to Unsupervised Conditional Image Generation (Hwang et al., 2021).
Self-Supervised Learning for Discriminative Models. On the discriminative side, works like SimCLR (Chen et al., 2020b), MoCo (He et al., 2020), and BYOL (Grill et al., 2020a) have recently shown overwhelming effectiveness in learning discriminative representations of data. MoCo (He et al., 2020) and SimCLR (Chen et al., 2020b) seek to learn features by pulling together features of augmented versions of the same sample while pushing apart features of all other samples, while BYOL (Grill et al., 2020a) trains a student network to predict the representation of a teacher network in a contrastive setting. BarlowTwins (Zbontar et al., 2021) and TCR (Li et al., 2022) learn by regularizing the covariance matrix of the embedding. However, features learned by this class of methods are typically highly compressed, and not designed to be used for generative purposes.
Self-Supervised Learning with Generative Models. On the generative side, the original GAN (Goodfellow et al., 2014) can be viewed as a natural self-supervised learning task. With an additional linear probe, works like DCGAN (Radford et al., 2015) have shown that features in the discriminator can be used for discriminative tasks. To further enhance the features, extensions like BiGAN (Donahue et al., 2016) and ALI (Dumoulin et al., 2016) introduce a third network into the GAN framework, aimed at learning an inverse mapping for the generator, which when coupled with labeled images can be used to study and supervise semantics in learned representations. Other works like SSGAN (Chen et al., 2019), SSGAN-LA (Hou et al., 2021), and ContraD (Jeong & Shin, 2021) propose to put augmentation tasks into GAN training to facilitate representation learning. Outside of GANs, variational autoencoders (VAEs) have been adapted to generate more semantically meaningful representations by trading off latent channel capacity and independence constraints with reconstruction accuracy (Higgins et al., 2016), an idea that has also been incorporated into recognition improvements using patch-level bottlenecks (Gupta et al., 2020), which encourage a V AE to focus on useful patterns in images. By incorporating data-augmentation, VAE is also shown to achieve fair discriminative performance (Falcon et al., 2021). Recently, works like MAE (He et al., 2021) and CAE (Chen et al., 2022) have learned representations by solving masked reconstruction tasks using vision transformers. Autogressive approaches like iGPT (Chen et al., 2020a) have also demonstrated decent self-supervised learning performance, which improves further with the incorporation of contrastive learning (Kim et al., 2021). However, unless supervised, features learned by those previously mentioned methods either do not have strong discriminative performance, or cannot be directly exploited to condition the generative task.
Unsupervised Conditional Image Generation (UCIG). For generative models, we often want to be able to generate images conditioned on a certain class or style, even in a completely unsupervised setting. This requires that the learned representations have structures that correspond to the desired conditioning. InfoGAN (Chen et al., 2016) proposes to learn interpretable representations by maximizing the mutual information between the observation and a subset of the latent code. ClusterGAN (Mukherjee et al., 2019) assumes a discrete Gaussian prior where discrete variables are defined as a one-hot vector and continuous variables are sampled from Gaussian distribution. Self-Conditioned GAN (Liu et al., 2020) uses clustering of discriminative features as labels to train. SLOGAN (Hwang et al., 2021) proposes a new conditional contrastive loss (U2C) to learn latent distribution of the data. Note that compared to our work, ClusterGAN and SLOGAN introduce an additional encoder that leads to increased computational complexity. On the VAE side, works like VaDE (Jiang et al., 2016) cluster based on the learned feature of a supervised ResNet. Variational Cluster (Prasad et al., 2020) simultaneously learns a prior that captures the latent distribution of the images and a posterior to help discriminate between data points in an end-to-end unsupervised setting. In this work, we will see how clusters can be estimated in a principled way in a more unified framework, by optimizing the same type of objective function that we use for learning features.
Method
Preliminaries: Rate Reduction and Closed-Loop Transcription
Assumptions on Data. Our work, as well as prior work in closed-loop transcription (Dai et al., 2022; Tong et al., 2022), considers a set of N images X = [ x 1 , x 2 , ..., x N ] ⊂ R D sampled from k classes. Borrowing notation from (Yu et al., 2020), the membership of the N samples in the k classes is denoted using k diagonal matrices: Π = { Π j ∈ R N × N } k j =1 , where the diagonal entry Π j ( i, i ) of Π j is the probability of sample i belonging to subset j . Let Ω . = { Π | ∑ Π j = I , Π j ≥ 0 . } be the set of all such matrices. WLOG, we may assume that classes are separable, with images for each belonging to a low-dimensional submanifold in the space R D .
Unsupervised Discriminative Autoencoding. The goal of transcription is to learn a unified representation, with the structure required to both classify and generate images from these k classes. Concretely, this is achieved by learning two continuous mappings: (1) an encoder parametrized by θ :
$$
$$
$$
$$
In this work, we specifically learn this mapping in an entirely unsupervised fashion, without knowing the ground-truth class labels Π at all. As stated in the introduction, a both discriminative and generative representation is difficult to achieve by standard generative methods like V AEs and GANs. This is one of the motivations for the closed-loop transcription framework (CTRL) proposed by (Dai et al., 2022), which we will generalize to the unsupervised setting.
Maximizing Rate Reduction. The CTRL framework (Dai et al., 2022) was proposed for the supervised setting, where it aims to map each class onto an independent linear subspace. As shown in (Yu et al., 2020), such a linear discriminative representation (LDR) can be achieved by maximizing a coding rate reduction objective, known as the MCR 2 principle :
$$
$$
where each Π j encodes the membership of the N samples described before. As discussed in (Chan et al., 2022), the first term R ( Z ) measures the total rate (volume) of all features whereas the second term R c measures the average rate (volume) of the k components. Our work adapts this formula to design meaningful objectives in the unsupervised setting.
Closed-Loop Transcription. To learn the autoencoding X f ( x ,θ ) - - - - → Z g ( z ,η ) - - - - → ˆ X , a fundamental question is how we measure the difference between X and the regenerated ˆ X = g ( f ( X )) . It is typically very difficult to put a proper distance measure in the image space (Wang et al., 2004). To
bypass this difficulty, the closed-loop transcription framework (Dai et al., 2022) proposes to measure the difference between X and ˆ X through the difference between their features Z and ˆ Z mapped through the same encoder:
$$
$$
The difference can be measured by the rate reduction between Z and ˆ Z , a special case of (1) with k = 2 classes:
$$
$$
Such a ∆ R is a principled distance between subspace-like Gaussian ensembles, with the property that ∆ R ( Z , ˆ Z ) = 0 iff Cov ( Z ) = Cov ( ˆ Z ) (Ma et al., 2007).
As shown in (Dai et al., 2022), applying this measure in the closed-loop CTRL formulation can already learn a decent autoencoding, even without class information. This is known as the CTRL-Binary program:
$$
$$
However, note that (4) is practically limited because it only aligns the dataset X and the regenerated ˆ X at the distribution level. There is no guarantee that for each sample x would be close to the decoded ˆ x = g ( f ( x )) . For example, (Dai et al., 2022) shows that a car sample can be decoded into a horse; the so obtained (autoencoding) representations are not sample-wise self-consistent !
Assumptions on Data.
Unsupervised Discriminative Autoencoding.
Building on U-CTRL's ability to cluster CIFAR-10 samples, we demonstrate the model's ability to perform unsupervised conditional image generation in Figure 8. In contrast to reconstruction, where images are regenerated from features corresponding to real samples, we generate images based on the feature sampling technique proposed in (Dai et al., 2022). From these results, we observe that the U-CTRL framework maintains in-cluster diversity, and that the diversity can be recovered and visualized via simple principal component analysis.
Maximizing Rate Reduction.
In the past decade, we have witnessed an explosive development in the practice of machine learning, particularly with deep learning methods. A key driver of success in practical applications has been marvelous engineering endeavors, often focused on fitting increasingly large deep networks to input data paired with task-specific sets of labels. Brute-force approaches of this nature, however, exert tremendous demands on hand-labeled data for supervision and computational resources for training and inference. As a result, an increasing amount of attention has been directed toward using selfsupervised or unsupervised techniques to learn representations that can not only learn without human annotation effort, but also be shared across downstream tasks.
Discriminative versus Generative. Tasks in unsupervised learning are typically separated into two categories. Discriminative ones frame high-dimensional observations as inputs, from which lowdimensional class or latent information can be extracted, while generative ones frame observations as generated outputs, which should often be sampled given some semantically meaningful conditioning.
Unsupervised learning approaches targeted at discriminative tasks are mainly based on a key idea: to pull different views from the same instance closer while enforcing a non-collapsed representation by either contrastive learning techniques (Chen et al., 2020b; He et al., 2020; Grill et al., 2020a), covariance regularization methods (Bardes et al., 2021; Zbontar et al., 2021), or using architecture design (Chen & He, 2020; Grill et al., 2020b). Their success is typically measured by the accuracy of a simple classifier (say a shallow network) trained on the representations that they produce, which have progressively improved over the years. Representations learned from these approaches, however, do not emphasize much about the intrinsic structure of the data distribution, and have not demonstrated success for generative purposes.
In parallel, generative methods like GANs (Goodfellow et al., 2014) and VAEs (Kingma & Welling, 2013) have also been explored for unsupervised learning. Although generative methods have made
∗ Equal contribution
† Work done during visiting at Berkeley
striking progress in the quality of the sampled or autoencoded data, when compared to the aforementioned discriminative methods, representations learned with these approaches demonstrate inferior performance in classification.
Toward A Unified Representation? The disparity between discriminative and generative approaches in unsupervised learning, contrasted against the fundamental goal of learning representations that are useful across many tasks, leads to a natural question that we investigate in this paper: in the unsupervised setting, is it possible to learn a unified representation that is effective for both discriminative and generative purposes? Further, do they mutually benefit each other? Concretely, we aim to learn a structured representation with the following two properties:
- The learned representation should be discriminative, such that simple classifiers applied to learned features yield high classification accuracy.
- The learned representation should be generative, with enough diversity to recover raw inputs, and structure that can be exploited for sampling and generating new images.
The fact that human visual memory serves both discriminative tasks (for example, detection and recognition) and generative or predictive tasks (for example, via replay) (Keller & Mrsic-Flogel, 2018; Josselyn & Tonegawa, 2020; Ven et al., 2020) indicates that this goal is achievable. Beyond being possible, these properties are also highly practical - successfully completing generative tasks like unsupervised conditional image generation (Hwang et al., 2021), for example, inherently requires that learned features for different classes be both structured for sampling and discriminative for conditioning. On the other hand, the generative property can serve as a natural regularization to avoid representation collapse.
Closed-Loop Transcription via a Constrained Maximin Game. The class of linear discriminative representations (LDRs) has recently been proposed for learning diverse and discriminative features for multi-class (visual) data, via optimization of the rate reduction objective (Chan et al., 2022). In the supervised setting, these representations have been shown to be be both discriminative and generative if learned in a closed-loop transcription framework via a maximin game over the rate reduction utility between an encoder and a decoder (Dai et al., 2022). Beyond the standard joint learning setting, where all classes are sampled uniformly throughout training, the closed-loop framework has also been successfully adapted to the incremental setting (Tong et al., 2022), where the optimal multi-class LDR is learned one class at a time. In the incremental (supervised) learning setting, one solves a constrained maximin problem over the rate reduction utility which keeps learned memory of old tasks intact (as constraints) while learning new tasks. It has been shown that this new framework can effectively alleviate the catastrophic forgetting suffered by most supervised learning methods.
Contributions. In this work, we show that the closed-loop transcription framework proposed for learning LDRs in the supervised setting (Chan et al., 2022) can be adapted to a purely unsupervised setting. In the unsupervised setting, we only have to view each sample and its augmentations as a 'new class' while using the rate reduction objective to ensure that learned features are both invariant to augmentation and self-consistent in generation; this leads to a constrained maximin game that is similar to the one explored for incremental learning (Tong et al., 2022). Our overall approach is illustrated in Figure 1.
As we experimentally demonstrate in Section 4, our formulation benefits from the mutual benefits of both discriminative and generative properties. It bridges the gap between two formerly distinct set of methods: by standard metrics and under comparable experimental conditions, it enables classification performance on par with and unsupervised conditional generative quality significantly higher than state-of-the-art techniques. Coupled with evidence from prior work, this suggests that the closed-loop transcription through the (constrained) maximin game between the encoder and decoder has the potential to offer a unifying framework for both discriminative and generative representation learning, across supervised, incremental, and unsupervised settings.
Closed-Loop Transcription.
Assumptions on Data. Our work, as well as prior work in closed-loop transcription (Dai et al., 2022; Tong et al., 2022), considers a set of N images X = [ x 1 , x 2 , ..., x N ] ⊂ R D sampled from k classes. Borrowing notation from (Yu et al., 2020), the membership of the N samples in the k classes is denoted using k diagonal matrices: Π = { Π j ∈ R N × N } k j =1 , where the diagonal entry Π j ( i, i ) of Π j is the probability of sample i belonging to subset j . Let Ω . = { Π | ∑ Π j = I , Π j ≥ 0 . } be the set of all such matrices. WLOG, we may assume that classes are separable, with images for each belonging to a low-dimensional submanifold in the space R D .
Unsupervised Discriminative Autoencoding. The goal of transcription is to learn a unified representation, with the structure required to both classify and generate images from these k classes. Concretely, this is achieved by learning two continuous mappings: (1) an encoder parametrized by θ :
$$
$$
$$
$$
In this work, we specifically learn this mapping in an entirely unsupervised fashion, without knowing the ground-truth class labels Π at all. As stated in the introduction, a both discriminative and generative representation is difficult to achieve by standard generative methods like V AEs and GANs. This is one of the motivations for the closed-loop transcription framework (CTRL) proposed by (Dai et al., 2022), which we will generalize to the unsupervised setting.
Maximizing Rate Reduction. The CTRL framework (Dai et al., 2022) was proposed for the supervised setting, where it aims to map each class onto an independent linear subspace. As shown in (Yu et al., 2020), such a linear discriminative representation (LDR) can be achieved by maximizing a coding rate reduction objective, known as the MCR 2 principle :
$$
$$
where each Π j encodes the membership of the N samples described before. As discussed in (Chan et al., 2022), the first term R ( Z ) measures the total rate (volume) of all features whereas the second term R c measures the average rate (volume) of the k components. Our work adapts this formula to design meaningful objectives in the unsupervised setting.
Closed-Loop Transcription. To learn the autoencoding X f ( x ,θ ) - - - - → Z g ( z ,η ) - - - - → ˆ X , a fundamental question is how we measure the difference between X and the regenerated ˆ X = g ( f ( X )) . It is typically very difficult to put a proper distance measure in the image space (Wang et al., 2004). To
bypass this difficulty, the closed-loop transcription framework (Dai et al., 2022) proposes to measure the difference between X and ˆ X through the difference between their features Z and ˆ Z mapped through the same encoder:
$$
$$
The difference can be measured by the rate reduction between Z and ˆ Z , a special case of (1) with k = 2 classes:
$$
$$
Such a ∆ R is a principled distance between subspace-like Gaussian ensembles, with the property that ∆ R ( Z , ˆ Z ) = 0 iff Cov ( Z ) = Cov ( ˆ Z ) (Ma et al., 2007).
As shown in (Dai et al., 2022), applying this measure in the closed-loop CTRL formulation can already learn a decent autoencoding, even without class information. This is known as the CTRL-Binary program:
$$
$$
However, note that (4) is practically limited because it only aligns the dataset X and the regenerated ˆ X at the distribution level. There is no guarantee that for each sample x would be close to the decoded ˆ x = g ( f ( x )) . For example, (Dai et al., 2022) shows that a car sample can be decoded into a horse; the so obtained (autoencoding) representations are not sample-wise self-consistent !
Sample-Wise Constraints for Unsupervised Transcription
To improve discriminative and generative properties of representations learned in the unsupervised setting, we propose two additional mechanisms for the above CTRL-Binary maximin game (4). For simplicity and uniformity, here these will be formulated as equality constraints over rate reduction measures, but in practice they can be enforced softly during optimization.
Sample-wise Self-Consistency via Closed-Loop Transcription. First, to address the issue that CTRL-Binary does not learn a sample-wise consistent autoencoding, we need to promote ˆ x to be close to x for each sample. In the CTRL framework, this can be achieved by enforcing that their corresponding features z = f ( x ) and ˆ z = f (ˆ x ) are the same or close. To promote sample-wise self-consistency, where ˆ x = g ( f ( x )) is close to x , we want the distance between z and ˆ z to be zero or small, for all N samples. This can be formulated using rate reduction; note that this again avoids measuring differences in the image space:
$$
$$
Self-Supervision via Compressing Augmented Samples. Since we do not know any class label information between samples in the unsupervised settings, the best we can do is to view every sample and its augmentations (say via translation, rotation, occlusion etc) as one 'class' - a basic idea behind almost all self-supervised learning methods. In the rate reduction framework, it is natural to compress the features of each sample and its augmentations. In this work, we adopt the standard transformations in SimCLR (Chen et al., 2020b) and denote such a transformation as τ . We denote each augmented sample x a = τ ( x ) , and its corresponding feature as z a = f ( x a , θ ) . For discriminative purposes, we hope the classifier is invariant to such transformations. Hence it is natural to enforce that the features z a of all augmentations are the same as that z of the original sample x . This is equivalent to requiring the distance between z and z a , measured in terms of rate reduction again, to be zero (or small) for all N samples:
$$
$$
Sample-wise Self-Consistency via Closed-Loop Transcription.
So far, we know the CTRL-Binary objective ∆ R ( Z , ˆ Z ) in (4) helps align the distributions while sample-wise self-consistency (5) and sample-wise augmentation (6) help align and compress features associated with each sample. Besides consistency, we also want learned representations are maximally discriminative for different samples (here viewed as different 'classes'). Notice that the rate distortion term R ( Z ) measures the coding rate (hence volume) of all features. It has been observed in (Li et al., 2022) that by maximizing this term, learned features expand and hence become more discriminative.
Unsupervised CTRL. Putting these elements together, we propose to learn a representation via the following constrained maximin program, which we refer to as unsupervised CTRL (U-CTRL):
$$
$$
In practice, the above program can be optimized by alternating maximization and minimization between the encoder f ( · , θ ) and the decoder g ( · , η ) . We adopt the following optimization strategy that works well in practice, which is used for all subsequent experiments on real image datasets:
$$
$$
$$
$$
where the constraints ∑ i ∈ N ∆ R ( z i , ˆ z i ) = 0 and ∑ i ∈ N ∆ R ( z i , z i a ) = 0 in (7) have been converted (and relaxed) to Lagrangian terms with corresponding coefficients λ 1 and λ 2 . 1
Unsupervised Conditional Image Generation via Rate Reduction. The above representation is learned without class information. In order to facilitate discriminative or generative tasks, it must be highly structured. As we will see via experiments, specific and unique structure indeed emerges naturally in the representations learned using U-CTRL: globally, features of images in the same class tend to be clustered well together and separated from other classes (Figure 2); locally, features around individual samples exhibit approximately piecewise linear low-dimensional structures (Figure 5).
The highly-structured feature distribution also suggests that the learned representation can be very useful for generative purposes. For example, we can organize the sample features into meaningful clusters, and model them with low-dimensional (Gaussian) distributions or subspaces. By sampling from these compact models, we can conditionally regenerate meaningful samples from computed clusters. This is known as unsupervised conditional image generation (Hwang et al., 2021).
To cluster features, we exploit the fact that the rate reduction framework (1) is inspired by unsupervised clustering via compression (Ma et al., 2007), which provides a principled way to find the membership Π . Concretely, we maximize the same rate reduction objective (1) over Π , but fix the learned representation Z instead. We simply view the membership Π as a nonlinear function of the features Z , say h π ( · , ξ ) : Z ↦→ Π with parameters ξ . In practice, we model this function with a simple neural network, such as an MLP head right after the output feature z . To estimate a 'pseudo' membership ˆ Π of the samples, we solve the following optimization problem over Π :
$$
$$
Experiments in Section 4.2 demonstrate that conditional image generation from clusters produced in this manner result in high-quality images that are highly similar in style.
Self-Supervision via Compressing Augmented Samples.
Unsupervised Representation Learning via Closed-Loop Transcription
So far, we know the CTRL-Binary objective ∆ R ( Z , ˆ Z ) in (4) helps align the distributions while sample-wise self-consistency (5) and sample-wise augmentation (6) help align and compress features associated with each sample. Besides consistency, we also want learned representations are maximally discriminative for different samples (here viewed as different 'classes'). Notice that the rate distortion term R ( Z ) measures the coding rate (hence volume) of all features. It has been observed in (Li et al., 2022) that by maximizing this term, learned features expand and hence become more discriminative.
Unsupervised CTRL. Putting these elements together, we propose to learn a representation via the following constrained maximin program, which we refer to as unsupervised CTRL (U-CTRL):
$$
$$
In practice, the above program can be optimized by alternating maximization and minimization between the encoder f ( · , θ ) and the decoder g ( · , η ) . We adopt the following optimization strategy that works well in practice, which is used for all subsequent experiments on real image datasets:
$$
$$
$$
$$
where the constraints ∑ i ∈ N ∆ R ( z i , ˆ z i ) = 0 and ∑ i ∈ N ∆ R ( z i , z i a ) = 0 in (7) have been converted (and relaxed) to Lagrangian terms with corresponding coefficients λ 1 and λ 2 . 1
Unsupervised Conditional Image Generation via Rate Reduction. The above representation is learned without class information. In order to facilitate discriminative or generative tasks, it must be highly structured. As we will see via experiments, specific and unique structure indeed emerges naturally in the representations learned using U-CTRL: globally, features of images in the same class tend to be clustered well together and separated from other classes (Figure 2); locally, features around individual samples exhibit approximately piecewise linear low-dimensional structures (Figure 5).
The highly-structured feature distribution also suggests that the learned representation can be very useful for generative purposes. For example, we can organize the sample features into meaningful clusters, and model them with low-dimensional (Gaussian) distributions or subspaces. By sampling from these compact models, we can conditionally regenerate meaningful samples from computed clusters. This is known as unsupervised conditional image generation (Hwang et al., 2021).
To cluster features, we exploit the fact that the rate reduction framework (1) is inspired by unsupervised clustering via compression (Ma et al., 2007), which provides a principled way to find the membership Π . Concretely, we maximize the same rate reduction objective (1) over Π , but fix the learned representation Z instead. We simply view the membership Π as a nonlinear function of the features Z , say h π ( · , ξ ) : Z ↦→ Π with parameters ξ . In practice, we model this function with a simple neural network, such as an MLP head right after the output feature z . To estimate a 'pseudo' membership ˆ Π of the samples, we solve the following optimization problem over Π :
$$
$$
Experiments in Section 4.2 demonstrate that conditional image generation from clusters produced in this manner result in high-quality images that are highly similar in style.
Unsupervised CTRL.
Building on U-CTRL's ability to cluster CIFAR-10 samples, we demonstrate the model's ability to perform unsupervised conditional image generation in Figure 8. In contrast to reconstruction, where images are regenerated from features corresponding to real samples, we generate images based on the feature sampling technique proposed in (Dai et al., 2022). From these results, we observe that the U-CTRL framework maintains in-cluster diversity, and that the diversity can be recovered and visualized via simple principal component analysis.
Unsupervised Conditional Image Generation via Rate Reduction.
Building on U-CTRL's ability to cluster CIFAR-10 samples, we demonstrate the model's ability to perform unsupervised conditional image generation in Figure 8. In contrast to reconstruction, where images are regenerated from features corresponding to real samples, we generate images based on the feature sampling technique proposed in (Dai et al., 2022). From these results, we observe that the U-CTRL framework maintains in-cluster diversity, and that the diversity can be recovered and visualized via simple principal component analysis.
Experiments
We now evaluate the performance of the proposed U-CTRL framework and compare it with representative unsupervised generative and discriminative methods. The first set of experiments (Section 4.1 show that despite being a generative method in nature, U-CTRL can learn discriminative representations competitive with state-of-the-art discriminative methods. The second set (Section 4.2) show that the learned generative representation can significantly boost the performance of unsupervised conditional image generation. Finally, the third set (Section 4.3) study how the advantages that generative represeentations have over discriminative ones.
We conduct experiments on the following datasets: CIFAR-10 (Krizhevsky et al., 2014), CIFAR-100 (Krizhevsky et al., 2009), and Tiny ImageNet (Deng et al., 2009). Standard augmentations for self-supervised learning are used across all datasets (Chen et al., 2020b).
1 Notice that computing the rate reduction terms ∆ R for all samples or a batch of samples requires computing the expensive log det of large matrices. In practice, from the geometric meaning of ∆ R for two vectors, ∆ R can be approximated with an glyph[lscript] 2 norm or the cosine distance between two vectors.
Table 2: Comparison of classification accuracy on CIFAR-10, CIFAR-100, and Tiny-ImageNet with other generative self-supervised learning methods. U-CTRL is clearly better.
We design all experiments to ensure that comparisons against U-CTRL are fair. For all methods that we compare against, we ensure that experiments are conducted with similar model sizes. If code for similar size structure can not be found, we uniformly use ResNet-18 to reproduce results for baselines, which is larger than the network used by our method. Details about network architectures and the experimental setting are given in Appendix A. All methods have runned 400 epochs or equivalent iterations (because generative models often count in iteration).
Discriminative Quality of Learned Representations
To evaluate the discriminative quality of the learned representations, we follow the standard practice of evaluating the accuracy of a simple linear classifier trained on the learned representation. Table 2 compares our method against SOTA generative self-supervised learning methods, and Table 3 compares our method against SOTA discriminative self-supervised methods. Experimental and training details are given in Appendix A.
Quantitative Comparisons of Classification Performance. From Table 2, we observe that on all chosen datasets, our method achieves substantial improvements compared to existing generative self-supervised learning methods. This includes more complex datasets like CIFAR-100 and TinyImageNet, where we surpass the current SOTA models. From Table 3, our method achieves similar performance compared to SOTA discriminative self-supervised models. These results echo our goal of seeking a more unifed generative and discriminative representations: despite resembling a generative method architecturally, our method still produces highly discriminative representations. In addition, these results lead us to ask a fundamental question: when is incorporating both discriminative and generative properties a whole greater than the sum of its parts, particularly outside of the context of computational efficiency? We provide preliminary answers in Section 4.3.
Qualitative Visualization of Learned Representations. To explain the classification performance of our method, we visualize the incoherence between features learned for the training datasets. Figure 2 shows cosine similarity heatmaps between the learned features, organized by ground-truth class labels. A block-diagonal pattern emerges automatically from U-CTRL training for all three datasets, similar to those observed in features learned in a supervised setting (Dai et al., 2022). In this case, however, these blocks emerge and correspond with classes labels despite the absence of any supervision at all.

Figure 2: Emergence of block-diagonal structures of | Z glyph[latticetop] Z | in the feature space for CIFAR-10 (left), 10 random classes from CIFAR-100 (middle), and 10 random classes from Tiny ImageNet (right).

(a) CIFAR-10 X

Figure 3: Sample-wise self-consistency: visualization of images X and reconstructed ˆ X
Improved Unsupervised Conditional Generation Quality
To evaluate the quality of unsupervised conditional image generation, we measure performance on two axes: cluster quality and image quality. We estimate clusters by optimizing (10), and show results and comparisons with both recent and classical methods in Table 4. Training details of our method for the additional MLP head can be found in the Appendix A.
Cluster Quality. We measure normalized mutual information (NMI) and clustering accuracy for cluster quality on CIFAR-10 clustered into 10 classes and CIFAR-100(20), which is clustered into 20 super-classes. From Table 4, we observe that on CIFAR-10, U-CTRL results in an NMI that is almost double that of the existing SOTA on both GAN-based and VAE-based methods, with significantly improved clustering accuracy. Unlike many baselines, we also demonstrate that our method scales to the more challenging CIFAR-100(20) dataset, where it also significantly outperforms alternatives. Our improved clustering quality suggests potential for improving unsupervised conditional image generation, which relies on first finding statistically (and hence visually) meaningful clusters.
Image Quality. We use Frechet Inception Distance (FID) (Heusel et al., 2017) and Inception Score (IS) (Salimans et al., 2016) to measure image quality. From Table 4, it is evident that U-CTRL maintains competitive image quality compared to other methods, measured both by FID and IS. We also compare original images against reconstructed ones in Figure 3, where we see that the original X is very similar to the reconstructed ˆ X ; U-CTRL indeed achieves very good sample-wise self-consistency.
Unsupervised Conditional Image Generation. In Figure 4, we visualize images generated from the ten unsupervised clusters from (10). Each block represents one cluster and each row represents one principal component for each cluster. Despite learning and training without labels, the model not only organizes samples into correct clusters, but is also able to preserve statistical diversities within each cluster/class. We can easily recover the diversity within each cluster by computing different principal components and then sample and generate accordingly! More detailed illustrations with more samples is provided in Appendix B.
Discriminative Quality of Learned Representations
As shown in the previous section, on datasets like CIFAR-10, CIFAR-100, and Tiny-ImageNet, our framework is able to achieve representation quality on par with the best discriminative self-supervised learning methods. A clear advantage of this is computational efficiency; only a single representation needs to be trained for a much broader set of tasks. This subsection aims to provide additional insights on how a unified model can be more beneficial for a broader range of tasks.
Domain Transfer. Regenerating images is demanding on the encoder, which is required to produce a more informative representation than contrastive training would. We hypothesize that the encoder

Figure 4: Unsupervised conditional image generation from each cluster of CIFAR-10, using U-CTRL. Images from different rows mean generation from different principal components of each cluster.
Table 4: Comparison of the quality of UCIG on CIFAR-10 and CIFAR-100(20). Many of the methods compared do not provide code that scales up to CIFAR-100(20), in which case we leave the corresponding table cell blank.
trained with generative task may retain more information about the image and allow the representation to generalize better. To verify this, we compare the accuracy on CIFAR-100 using models learned from CIFAR-10 in Table 5. When compared to purely discriminative self-supervised learning models, we observe that U-CTRL is 4 percent better than other methods on classification accuracy.
Visualization of Emerged Structures. The representations learned by U-CTRL are significantly different from those learned from previous either discriminative and generative methods. To illustrate this, we use t-SNE (Van der Maaten & Hinton, 2008) to visualize the learned representation in 2D. Figure 5 compares the t-SNE of representations learned for CIFAR-10 by U-CTRL and MoCoV2, respectively. It is clear that the representation learned by U-CTRL are much more structured and better organized: classes are more evident, and features within each class form clear piecewise linear structures.
Table 5: Comparing the transfer ability with purely discriminative self-supervised learning methods. All methods are trained unsupervised on CIFAR-10 and tested on CIFAR-100.

(a) U-CTRL
(b) MoCoV2
Figure 5: t-SNE visualizations of learned features of CIFAR-10 with different models.
Conclusion and Discussion
In this work, we proposed an unsupervised formulation of the closed-loop transcription framework (Dai et al., 2022). We experimentally demonstrate that it is possible to learn a unified representation for both discriminative and generative purposes, resulting in highly structured representations.
Further, we show that these two purposes mutually benefit each other in various tasks, e.g., conditional image generation and domain tranfers. Compared to the more specialized representations learned in prior works, our results suggest that such a unified representation has the potential in supporting and benefiting a wider range of new tasks. In future work, we believe the learned representations can be further improved by jointly optimizing the feature representation and feature clusters, as suggested in the original rate reduction paper (Chan et al., 2022). Features with high likelihood of belonging to the same cluster can be further linearized and compressed. Due to its unifying nature and the simplicity of the underlying concepts, this new framework may be extended beyond image data, such as sequential or dynamical observations.
Acknowledgements
Yi Ma acknowledges support from ONR grants N00014-20-1-2002 and N00014-22-1-2102, the joint Simons Foundation-NSF DMS grant #2031899, as well as partial support from Berkeley FHL Vive Center for Enhanced Reality and Berkeley Center for Augmented Cognition, Tsinghua-Berkeley Shenzhen Institute (TBSI) Research Fund, and Berkeley AI Research (BAIR).
Experiments
Training Details
Network Architectures
Table 6, 7 and Figure 6 give details on the network architecture for the decoder and the encoder networks used for experiments. The black rectangle marked with "conv, s=2" means a convlutional layer with stride 2. The orange rectangle marked with "dconv, s=2" means a deconvolutional layer with stride 2. The "x k" besides red frame means we regard these layers in red frame as a block and stack it k times. All α values in Leaky-ReLU (i.e. lReLU) of the encoder are set to 0 . 2 . We set ( nz = 128 , nc = 3 , k = 3 ) for CIFAR-10, ( nz = 256 , nc = 3 , k = 4 ) for CIFAR-100, and ( nz = 256 , nc = 3 , k = 4 ) for Tiny-ImageNet. As a comparison, ResNet-18 contains around 11 million parameters, whereas our encoder only contains between 4 and 6 million parameters depending on the choice of k.
Table 8 gives details of the network architecture for the linear classifier and Table 9 gives details of the network architecture for the additional MLP head used for unsupervised conditional image generation training.
Table 6: Network architecture of the decoder g ( · , η ) .

Optimization
For all experiments, we use Adam (Kingma & Ba, 2014) as our optimizer, with hyperparameters β 1 = 0 . 5 , β 2 = 0 . 999 . The learning rate is set to be 0.0001. We choose glyph[epsilon1] 2 = 0 . 2 . For all experiments, we adopt augmentation from SimCLR (Chen et al., 2020b).

Table 8: Network architecture of the linear classifier.
For CIFAR-10, CIFAR-100, and Tiny ImageNet, we train our framework with a batch size of 1024 over 20,000 iterations. All experiments are conducted with at most 4 RTX 3090 GPUs. Methods that are compared against in Table 3 are trained with the batch size of 256, because Chen et al. (2020b) observe that purely discriminative methods tend to perform better with smaller batch sizes. Table 2 methods have used their optimal parameters in their github code.
For training of the MLP head for unsupervised conditional image generation(10), we again use Adam (Kingma & Ba, 2014) as our optimizer with hyperparameters β 1 = 0 . 5 , β 2 = 0 . 999 . We choose the learning rate to be 0.0001 and glyph[epsilon1] 2 as 0.2, with batch size 1024 over 5000 iterations.
Additional Unsupervised Clustering and Generation Results
Cluster Reconstruction
In this subsection, we visualize the reconstruction of ten clusters that are predicted and generated by U-CTRL on the CIFAR-10 training set. Each block in Figure 7 contains both a random sample of reconstructed data in a cluster and the total number of samples within it. Note that CIFAR-10 contains 50,000 training samples, split across 10 classes. As we see in Figure 7, the number of samples in each cluster are very close to 5,000, with the largest deviator (cluster 9) containing 3,942 samples. Without any cues, one can easily identify correspond each unsupervised cluster with a CIFAR-10 class. For a class like 'bird', we observe that the model is able to group images of standing birds, flying birds, and bird heads, despite their visual differences.
Unsupervised Conditional Image Generation
Building on U-CTRL's ability to cluster CIFAR-10 samples, we demonstrate the model's ability to perform unsupervised conditional image generation in Figure 8. In contrast to reconstruction, where images are regenerated from features corresponding to real samples, we generate images based on the feature sampling technique proposed in (Dai et al., 2022). From these results, we observe that the U-CTRL framework maintains in-cluster diversity, and that the diversity can be recovered and visualized via simple principal component analysis.
Ablation Studies
The importance of each term in ours{
In this section, we study the significance of the sample-wise constraints and extra rate distortion term in the formulation 7. Table 10 presents the following objectives that we study:

Figure 7: More result on the reconstruction of clusters in CIFAR-10
Table 11 shows the result of a linear probe for representations trained using each objective on CIFAR10. From the table, it is evident that both constraints and the rate distortion term are pivotal to the success of our framework.
Table 10: Five different objective functions for U-CTRL.

(a) Cluster 1





Figure 8: Unsupervised conditional image generation on CIFAR-10. Each block represents a cluster, within which each row represents one principal component direction in the cluster, and samples along each row represent different noises applied in that principal direction.
The importance of MCR$^2$ in ours{
In this section, we verify the significance of MCR 2 term ∆ R ( Z , ˆ Z ) in our method. We do ablation study on CIFAR-10 with the same network and training condition. If we take away MCR 2 from our

(c) Cluster 3

Table 11: Ablation study on the significance of different terms in U-CTRL. formulation, it changes (11). For simplicity, we call it U-CTRL-noMCR 2
$$
$$
Table 12 shows that U-CTRL without the MCR 2 not only learns worse representation but also generalizes worse to out of distribution data. Figure 9 visualizes the reconstructed ˆ X by U-CTRLnoMCR 2 . It is clear from the image figure that without the MCR 2 , the decoder fails to reconstruct high-quality images.

(a) CIFAR-10
X

ˆ
Figure 9: Visualization of images trained by U-CTRL-noMCR 2 : X and reconstructed ˆ X on CIFAR-10 dataset. It follows our discussion in the introduction that discriminative tasks and generative tasks together learn feature that benifits each other.
Random Seed Sensitivity
This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes. While most existing unsupervised learning approaches focus on a representation for only one of these two goals, we show that a unified representation can enjoy the mutual benefits of having both. Such a representation is attainable by generalizing the recently proposed closed-loop transcription framework, known as CTRL, to the unsupervised setting. This entails solving a constrained maximin game over a rate reduction objective that expands features of all samples while compressing features of augmentations of each sample. Through this process, we see discriminative low-dimensional structures emerge in the resulting representations. Under comparable experimental conditions and network complexities, we demonstrate that these structured representations enable classification performance close to state-of-the-art unsupervised discriminative representations, and conditionally generated image quality significantly higher than that of state-of-the-art unsupervised generative models. Source code can be found at https://github.com/Delay-Xili/uCTRL.
In the past decade, we have witnessed an explosive development in the practice of machine learning, particularly with deep learning methods. A key driver of success in practical applications has been marvelous engineering endeavors, often focused on fitting increasingly large deep networks to input data paired with task-specific sets of labels. Brute-force approaches of this nature, however, exert tremendous demands on hand-labeled data for supervision and computational resources for training and inference. As a result, an increasing amount of attention has been directed toward using self-supervised or unsupervised techniques to learn representations that can not only learn without human annotation effort, but also be shared across downstream tasks.
Discriminative versus Generative. Tasks in unsupervised learning are typically separated into two categories. Discriminative ones frame high-dimensional observations as inputs, from which low-dimensional class or latent information can be extracted, while generative ones frame observations as generated outputs, which should often be sampled given some semantically meaningful conditioning.
Unsupervised learning approaches targeted at discriminative tasks are mainly based on a key idea: to pull different views from the same instance closer while enforcing a non-collapsed representation by either contrastive learning techniques (Chen et al., 2020b; He et al., 2020; Grill et al., 2020a), covariance regularization methods (Bardes et al., 2021; Zbontar et al., 2021), or using architecture design (Chen & He, 2020; Grill et al., 2020b). Their success is typically measured by the accuracy of a simple classifier (say a shallow network) trained on the representations that they produce, which have progressively improved over the years. Representations learned from these approaches, however, do not emphasize much about the intrinsic structure of the data distribution, and have not demonstrated success for generative purposes.
In parallel, generative methods like GANs (Goodfellow et al., 2014) and VAEs (Kingma & Welling, 2013) have also been explored for unsupervised learning. Although generative methods have made striking progress in the quality of the sampled or autoencoded data, when compared to the aforementioned discriminative methods, representations learned with these approaches demonstrate inferior performance in classification.
Toward A Unified Representation? The disparity between discriminative and generative approaches in unsupervised learning, contrasted against the fundamental goal of learning representations that are useful across many tasks, leads to a natural question that we investigate in this paper: in the unsupervised setting, is it possible to learn a unified representation that is effective for both discriminative and generative purposes? Further, do they mutually benefit each other? Concretely, we aim to learn a structured representation with the following two properties:
The learned representation should be discriminative, such that simple classifiers applied to learned features yield high classification accuracy.
The fact that human visual memory serves both discriminative tasks (for example, detection and recognition) and generative or predictive tasks (for example, via replay) (Keller & Mrsic-Flogel, 2018; Josselyn & Tonegawa, 2020; Ven et al., 2020) indicates that this goal is achievable. Beyond being possible, these properties are also highly practical – successfully completing generative tasks like unsupervised conditional image generation (Hwang et al., 2021), for example, inherently requires that learned features for different classes be both structured for sampling and discriminative for conditioning. On the other hand, the generative property can serve as a natural regularization to avoid representation collapse.
Closed-Loop Transcription via a Constrained Maximin Game. The class of linear discriminative representations (LDRs) has recently been proposed for learning diverse and discriminative features for multi-class (visual) data, via optimization of the rate reduction objective (Chan et al., 2022). In the supervised setting, these representations have been shown to be be both discriminative and generative if learned in a closed-loop transcription framework via a maximin game over the rate reduction utility between an encoder and a decoder (Dai et al., 2022). Beyond the standard joint learning setting, where all classes are sampled uniformly throughout training, the closed-loop framework has also been successfully adapted to the incremental setting (Tong et al., 2022), where the optimal multi-class LDR is learned one class at a time. In the incremental (supervised) learning setting, one solves a constrained maximin problem over the rate reduction utility which keeps learned memory of old tasks intact (as constraints) while learning new tasks. It has been shown that this new framework can effectively alleviate the catastrophic forgetting suffered by most supervised learning methods.
Contributions. In this work, we show that the closed-loop transcription framework proposed for learning LDRs in the supervised setting (Chan et al., 2022) can be adapted to a purely unsupervised setting. In the unsupervised setting, we only have to view each sample and its augmentations as a “new class” while using the rate reduction objective to ensure that learned features are both invariant to augmentation and self-consistent in generation; this leads to a constrained maximin game that is similar to the one explored for incremental learning (Tong et al., 2022). Our overall approach is illustrated in Figure 1.
As we experimentally demonstrate in Section 4, our formulation benefits from the mutual benefits of both discriminative and generative properties. It bridges the gap between two formerly distinct set of methods: by standard metrics and under comparable experimental conditions, it enables classification performance on par with and unsupervised conditional generative quality significantly higher than state-of-the-art techniques. Coupled with evidence from prior work, this suggests that the closed-loop transcription through the (constrained) maximin game between the encoder and decoder has the potential to offer a unifying framework for both discriminative and generative representation learning, across supervised, incremental, and unsupervised settings.
Our work is mostly related to three categories of unsupervised learning methods: (1) self-supervised learning via discriminative models, (2) self-supervised learning via generative models, and (3) unsupervised conditional image generation. Table 1 compares the capabilities of models learned by various representative unsupervised learning methods.
Self-Supervised Learning for Discriminative Models. On the discriminative side, works like SimCLR (Chen et al., 2020b), MoCo (He et al., 2020), and BYOL (Grill et al., 2020a) have recently shown overwhelming effectiveness in learning discriminative representations of data. MoCo (He et al., 2020) and SimCLR (Chen et al., 2020b) seek to learn features by pulling together features of augmented versions of the same sample while pushing apart features of all other samples, while BYOL (Grill et al., 2020a) trains a student network to predict the representation of a teacher network in a contrastive setting. BarlowTwins (Zbontar et al., 2021) and TCR (Li et al., 2022) learn by regularizing the covariance matrix of the embedding. However, features learned by this class of methods are typically highly compressed, and not designed to be used for generative purposes.
Self-Supervised Learning with Generative Models. On the generative side, the original GAN (Goodfellow et al., 2014) can be viewed as a natural self-supervised learning task. With an additional linear probe, works like DCGAN (Radford et al., 2015) have shown that features in the discriminator can be used for discriminative tasks. To further enhance the features, extensions like BiGAN (Donahue et al., 2016) and ALI (Dumoulin et al., 2016) introduce a third network into the GAN framework, aimed at learning an inverse mapping for the generator, which when coupled with labeled images can be used to study and supervise semantics in learned representations. Other works like SSGAN (Chen et al., 2019), SSGAN-LA (Hou et al., 2021), and ContraD (Jeong & Shin, 2021) propose to put augmentation tasks into GAN training to facilitate representation learning. Outside of GANs, variational autoencoders (VAEs) have been adapted to generate more semantically meaningful representations by trading off latent channel capacity and independence constraints with reconstruction accuracy (Higgins et al., 2016), an idea that has also been incorporated into recognition improvements using patch-level bottlenecks (Gupta et al., 2020), which encourage a VAE to focus on useful patterns in images. By incorporating data-augmentation, VAE is also shown to achieve fair discriminative performance (Falcon et al., 2021). Recently, works like MAE (He et al., 2021) and CAE (Chen et al., 2022) have learned representations by solving masked reconstruction tasks using vision transformers. Autogressive approaches like iGPT (Chen et al., 2020a) have also demonstrated decent self-supervised learning performance, which improves further with the incorporation of contrastive learning (Kim et al., 2021). However, unless supervised, features learned by those previously mentioned methods either do not have strong discriminative performance, or cannot be directly exploited to condition the generative task.
Unsupervised Conditional Image Generation (UCIG). For generative models, we often want to be able to generate images conditioned on a certain class or style, even in a completely unsupervised setting. This requires that the learned representations have structures that correspond to the desired conditioning. InfoGAN (Chen et al., 2016) proposes to learn interpretable representations by maximizing the mutual information between the observation and a subset of the latent code. ClusterGAN (Mukherjee et al., 2019) assumes a discrete Gaussian prior where discrete variables are defined as a one-hot vector and continuous variables are sampled from Gaussian distribution. Self-Conditioned GAN (Liu et al., 2020) uses clustering of discriminative features as labels to train. SLOGAN (Hwang et al., 2021) proposes a new conditional contrastive loss (U2C) to learn latent distribution of the data. Note that compared to our work, ClusterGAN and SLOGAN introduce an additional encoder that leads to increased computational complexity. On the VAE side, works like VaDE (Jiang et al., 2016) cluster based on the learned feature of a supervised ResNet. Variational Cluster (Prasad et al., 2020) simultaneously learns a prior that captures the latent distribution of the images and a posterior to help discriminate between data points in an end-to-end unsupervised setting. In this work, we will see how clusters can be estimated in a principled way in a more unified framework, by optimizing the same type of objective function that we use for learning features.
Our work, as well as prior work in closed-loop transcription (Dai et al., 2022; Tong et al., 2022), considers a set of N𝑁N images 𝑿=[𝒙1,𝒙2,…,𝒙N]⊂ℝD𝑿superscript𝒙1superscript𝒙2…superscript𝒙𝑁superscriptℝ𝐷\bm{X}=[\bm{x}^{1},\bm{x}^{2},...,\bm{x}^{N}]\subset\mathbb{R}^{D} sampled from k𝑘k classes. Borrowing notation from (Yu et al., 2020), the membership of the N𝑁N samples in the k𝑘k classes is denoted using k𝑘k diagonal matrices: 𝚷={𝚷j∈ℝN×N}j=1k𝚷superscriptsubscriptsubscript𝚷𝑗superscriptℝ𝑁𝑁𝑗1𝑘\bm{\Pi}={\bm{\Pi}{j}\in\mathbb{R}^{N\times N}}{j=1}^{k}, where the diagonal entry 𝚷j(i,i)subscript𝚷𝑗𝑖𝑖\bm{\Pi}{j}(i,i) of 𝚷jsubscript𝚷𝑗\bm{\Pi}{j} is the probability of sample i𝑖i belonging to subset j𝑗j. Let Ω≐{𝚷∣∑𝚷j=𝑰,𝚷j≥𝟎.}\Omega\doteq{\bm{\Pi}\mid\sum\bm{\Pi}{j}=\bm{I},\bm{\Pi}{j}\geq\bm{0}.} be the set of all such matrices. WLOG, we may assume that classes are separable, with images for each belonging to a low-dimensional submanifold in the space ℝDsuperscriptℝ𝐷\mathbb{R}^{D}.
The goal of transcription is to learn a unified representation, with the structure required to both classify and generate images from these k𝑘k classes. Concretely, this is achieved by learning two continuous mappings: (1) an encoder parametrized by θ𝜃\theta: f(⋅,θ):𝒙↦𝒛∈ℝd:𝑓⋅𝜃maps-to𝒙𝒛superscriptℝ𝑑f(\cdot,\theta):\bm{x}\mapsto\bm{z}\in\mathbb{R}^{d} with d≪Dmuch-less-than𝑑𝐷d\ll D such that all samples are mapped to their features as 𝑿→f(𝒙,θ)𝒁𝑓𝒙𝜃→𝑿𝒁\bm{X}\xrightarrow{f(\bm{x},\theta)}\bm{Z} with 𝒁=[𝒛1,𝒛2,…,𝒛N]⊂ℝd𝒁superscript𝒛1superscript𝒛2…superscript𝒛𝑁superscriptℝ𝑑\bm{Z}=[\bm{z}^{1},\bm{z}^{2},...,\bm{z}^{N}]\subset\mathbb{R}^{d}, and (2) an inverse map g(⋅,η):𝒛↦𝒙^∈ℝD:𝑔⋅𝜂maps-to𝒛^𝒙superscriptℝ𝐷g(\cdot,\eta):\bm{z}\mapsto\hat{\bm{x}}\in\mathbb{R}^{D} such that 𝒙𝒙\bm{x} and 𝒙^=g(f(𝒙))^𝒙𝑔𝑓𝒙\hat{\bm{x}}=g(f(\bm{x})) is close. In other words, 𝑿→f(𝒙,θ)𝒁→g(𝒛,η)𝑿^𝑓𝒙𝜃→𝑿𝒁𝑔𝒛𝜂→^𝑿\bm{X}\xrightarrow{f(\bm{x},\theta)}\bm{Z}\xrightarrow{g(\bm{z},\eta)}\hat{\bm{X}} forms an autoencoding.
In this work, we specifically learn this mapping in an entirely unsupervised fashion, without knowing the ground-truth class labels 𝚷𝚷\bm{\Pi} at all. As stated in the introduction, a both discriminative and generative representation is difficult to achieve by standard generative methods like VAEs and GANs. This is one of the motivations for the closed-loop transcription framework (CTRL) proposed by (Dai et al., 2022), which we will generalize to the unsupervised setting.
The CTRL framework (Dai et al., 2022) was proposed for the supervised setting, where it aims to map each class onto an independent linear subspace. As shown in (Yu et al., 2020), such a linear discriminative representation (LDR) can be achieved by maximizing a coding rate reduction objective, known as the MCR2 principle:
where each 𝚷jsubscript𝚷𝑗\bm{\Pi}_{j} encodes the membership of the N𝑁N samples described before. As discussed in (Chan et al., 2022), the first term R(𝒁)𝑅𝒁R(\bm{Z}) measures the total rate (volume) of all features whereas the second term Rcsuperscript𝑅𝑐R^{c} measures the average rate (volume) of the k𝑘k components. Our work adapts this formula to design meaningful objectives in the unsupervised setting.
To learn the autoencoding 𝑿→f(𝒙,θ)𝒁→g(𝒛,η)𝑿^𝑓𝒙𝜃→𝑿𝒁𝑔𝒛𝜂→^𝑿\bm{X}\xrightarrow{f(\bm{x},\theta)}\bm{Z}\xrightarrow{g(\bm{z},\eta)}\hat{\bm{X}}, a fundamental question is how we measure the difference between 𝑿𝑿\bm{X} and the regenerated 𝑿^=g(f(𝑿))^𝑿𝑔𝑓𝑿\hat{\bm{X}}=g(f(\bm{X})). It is typically very difficult to put a proper distance measure in the image space (Wang et al., 2004). To bypass this difficulty, the closed-loop transcription framework (Dai et al., 2022) proposes to measure the difference between 𝑿𝑿\bm{X} and 𝑿^^𝑿\hat{\bm{X}} through the difference between their features 𝒁𝒁\bm{Z} and 𝒁^^𝒁\hat{\bm{Z}} mapped through the same encoder:
The difference can be measured by the rate reduction between 𝒁𝒁\bm{Z} and 𝒁^^𝒁\hat{\bm{Z}}, a special case of (1) with k=2𝑘2k=2 classes:
Such a ΔRΔ𝑅\Delta R is a principled distance between subspace-like Gaussian ensembles, with the property that ΔR(𝒁,𝒁^)=0Δ𝑅𝒁^𝒁0\Delta R(\bm{Z},\hat{\bm{Z}})=0 iff Cov(𝒁)=Cov(𝒁^)Cov𝒁Cov^𝒁\mbox{Cov}(\bm{Z})=\mbox{Cov}(\hat{\bm{Z}}) (Ma et al., 2007).
As shown in (Dai et al., 2022), applying this measure in the closed-loop CTRL formulation can already learn a decent autoencoding, even without class information. This is known as the CTRL-Binary program:
However, note that (4) is practically limited because it only aligns the dataset 𝑿𝑿\bm{X} and the regenerated 𝑿^^𝑿\hat{\bm{X}} at the distribution level. There is no guarantee that for each sample 𝒙𝒙\bm{x} would be close to the decoded 𝒙^=g(f(𝒙))^𝒙𝑔𝑓𝒙\hat{\bm{x}}=g(f(\bm{x})). For example, (Dai et al., 2022) shows that a car sample can be decoded into a horse; the so obtained (autoencoding) representations are not sample-wise self-consistent!
To improve discriminative and generative properties of representations learned in the unsupervised setting, we propose two additional mechanisms for the above CTRL-Binary maximin game (4). For simplicity and uniformity, here these will be formulated as equality constraints over rate reduction measures, but in practice they can be enforced softly during optimization.
First, to address the issue that CTRL-Binary does not learn a sample-wise consistent autoencoding, we need to promote 𝒙^^𝒙\hat{\bm{x}} to be close to 𝒙𝒙\bm{x} for each sample. In the CTRL framework, this can be achieved by enforcing that their corresponding features 𝒛=f(𝒙)𝒛𝑓𝒙\bm{z}=f(\bm{x}) and 𝒛^=f(𝒙^)^𝒛𝑓^𝒙\hat{\bm{z}}=f(\hat{\bm{x}}) are the same or close. To promote sample-wise self-consistency, where 𝒙^=g(f(𝒙))^𝒙𝑔𝑓𝒙\hat{\bm{x}}=g(f(\bm{x})) is close to 𝒙𝒙\bm{x} , we want the distance between 𝒛𝒛\bm{z} and 𝒛^^𝒛\hat{\bm{z}} to be zero or small, for all N𝑁N samples. This can be formulated using rate reduction; note that this again avoids measuring differences in the image space:
Since we do not know any class label information between samples in the unsupervised settings, the best we can do is to view every sample and its augmentations (say via translation, rotation, occlusion etc) as one “class” — a basic idea behind almost all self-supervised learning methods. In the rate reduction framework, it is natural to compress the features of each sample and its augmentations. In this work, we adopt the standard transformations in SimCLR (Chen et al., 2020b) and denote such a transformation as τ𝜏\tau. We denote each augmented sample 𝒙a=τ(𝒙)subscript𝒙𝑎𝜏𝒙\bm{x}{a}=\tau(\bm{x}), and its corresponding feature as 𝒛a=f(𝒙a,θ)subscript𝒛𝑎𝑓subscript𝒙𝑎𝜃\bm{z}{a}=f(\bm{x}{a},\theta). For discriminative purposes, we hope the classifier is invariant to such transformations. Hence it is natural to enforce that the features 𝒛asubscript𝒛𝑎\bm{z}{a} of all augmentations are the same as that 𝒛𝒛\bm{z} of the original sample 𝒙𝒙\bm{x}. This is equivalent to requiring the distance between 𝒛𝒛\bm{z} and 𝒛asubscript𝒛𝑎\bm{z}_{a}, measured in terms of rate reduction again, to be zero (or small) for all N𝑁N samples:
So far, we know the CTRL-Binary objective ΔR(𝒁,𝒁^)Δ𝑅𝒁^𝒁\Delta R(\bm{Z},\hat{\bm{Z}}) in (4) helps align the distributions while sample-wise self-consistency (5) and sample-wise augmentation (6) help align and compress features associated with each sample. Besides consistency, we also want learned representations are maximally discriminative for different samples (here viewed as different “classes”). Notice that the rate distortion term R(𝒁)𝑅𝒁R(\bm{Z}) measures the coding rate (hence volume) of all features. It has been observed in (Li et al., 2022) that by maximizing this term, learned features expand and hence become more discriminative.
Putting these elements together, we propose to learn a representation via the following constrained maximin program, which we refer to as unsupervised CTRL (U-CTRL):
In practice, the above program can be optimized by alternating maximization and minimization between the encoder f(⋅,θ)𝑓⋅𝜃f(\cdot,\theta) and the decoder g(⋅,η)𝑔⋅𝜂g(\cdot,\eta). We adopt the following optimization strategy that works well in practice, which is used for all subsequent experiments on real image datasets:
where the constraints ∑i∈NΔR(𝒛i,𝒛^i)=0subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscript^𝒛𝑖0\sum_{i\in N}\Delta R(\bm{z}^{i},\hat{\bm{z}}^{i})=0 and ∑i∈NΔR(𝒛i,𝒛ai)=0subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscriptsubscript𝒛𝑎𝑖0\sum_{i\in N}\Delta R(\bm{z}^{i},\bm{z}{a}^{i})=0 in (7) have been converted (and relaxed) to Lagrangian terms with corresponding coefficients λ1subscript𝜆1\lambda{1} and λ2subscript𝜆2\lambda_{2}.111Notice that computing the rate reduction terms ΔRΔ𝑅\Delta R for all samples or a batch of samples requires computing the expensive logdet\log\det of large matrices. In practice, from the geometric meaning of ΔRΔ𝑅\Delta R for two vectors, ΔRΔ𝑅\Delta R can be approximated with an ℓ2superscriptℓ2\ell^{2} norm or the cosine distance between two vectors.
The above representation is learned without class information. In order to facilitate discriminative or generative tasks, it must be highly structured. As we will see via experiments, specific and unique structure indeed emerges naturally in the representations learned using U-CTRL: globally, features of images in the same class tend to be clustered well together and separated from other classes (Figure 2); locally, features around individual samples exhibit approximately piecewise linear low-dimensional structures (Figure 5).
The highly-structured feature distribution also suggests that the learned representation can be very useful for generative purposes. For example, we can organize the sample features into meaningful clusters, and model them with low-dimensional (Gaussian) distributions or subspaces. By sampling from these compact models, we can conditionally regenerate meaningful samples from computed clusters. This is known as unsupervised conditional image generation (Hwang et al., 2021).
To cluster features, we exploit the fact that the rate reduction framework (1) is inspired by unsupervised clustering via compression (Ma et al., 2007), which provides a principled way to find the membership 𝚷𝚷\bm{\Pi}. Concretely, we maximize the same rate reduction objective (1) over 𝚷𝚷\bm{\Pi}, but fix the learned representation 𝒁𝒁\bm{Z} instead. We simply view the membership 𝚷𝚷\bm{\Pi} as a nonlinear function of the features 𝒁𝒁\bm{Z}, say h𝝅(⋅,ξ):𝒁↦𝚷:subscriptℎ𝝅⋅𝜉maps-to𝒁𝚷h_{\bm{\pi}}(\cdot,\xi):\bm{Z}\mapsto\bm{\Pi} with parameters ξ𝜉\xi. In practice, we model this function with a simple neural network, such as an MLP head right after the output feature 𝒛𝒛\bm{z}. To estimate a “pseudo” membership 𝚷^^𝚷\hat{\bm{\Pi}} of the samples, we solve the following optimization problem over 𝚷𝚷\bm{\Pi}:
Experiments in Section 4.2 demonstrate that conditional image generation from clusters produced in this manner result in high-quality images that are highly similar in style.
We now evaluate the performance of the proposed U-CTRL framework and compare it with representative unsupervised generative and discriminative methods. The first set of experiments (Section 4.1 show that despite being a generative method in nature, U-CTRL can learn discriminative representations competitive with state-of-the-art discriminative methods. The second set (Section 4.2) show that the learned generative representation can significantly boost the performance of unsupervised conditional image generation. Finally, the third set (Section 4.3) study how the advantages that generative represeentations have over discriminative ones.
We conduct experiments on the following datasets: CIFAR-10 (Krizhevsky et al., 2014), CIFAR-100 (Krizhevsky et al., 2009), and Tiny ImageNet (Deng et al., 2009). Standard augmentations for self-supervised learning are used across all datasets (Chen et al., 2020b).
We design all experiments to ensure that comparisons against U-CTRL are fair. For all methods that we compare against, we ensure that experiments are conducted with similar model sizes. If code for similar size structure can not be found, we uniformly use ResNet-18 to reproduce results for baselines, which is larger than the network used by our method. Details about network architectures and the experimental setting are given in Appendix A. All methods have runned 400 epochs or equivalent iterations (because generative models often count in iteration).
To evaluate the discriminative quality of the learned representations, we follow the standard practice of evaluating the accuracy of a simple linear classifier trained on the learned representation. Table 2 compares our method against SOTA generative self-supervised learning methods, and Table 3 compares our method against SOTA discriminative self-supervised methods. Experimental and training details are given in Appendix A.
Quantitative Comparisons of Classification Performance. From Table 2, we observe that on all chosen datasets, our method achieves substantial improvements compared to existing generative self-supervised learning methods. This includes more complex datasets like CIFAR-100 and Tiny-ImageNet, where we surpass the current SOTA models. From Table 3, our method achieves similar performance compared to SOTA discriminative self-supervised models. These results echo our goal of seeking a more unifed generative and discriminative representations: despite resembling a generative method architecturally, our method still produces highly discriminative representations. In addition, these results lead us to ask a fundamental question: when is incorporating both discriminative and generative properties a whole greater than the sum of its parts, particularly outside of the context of computational efficiency? We provide preliminary answers in Section 4.3.
Qualitative Visualization of Learned Representations. To explain the classification performance of our method, we visualize the incoherence between features learned for the training datasets. Figure 2 shows cosine similarity heatmaps between the learned features, organized by ground-truth class labels. A block-diagonal pattern emerges automatically from U-CTRL training for all three datasets, similar to those observed in features learned in a supervised setting (Dai et al., 2022). In this case, however, these blocks emerge and correspond with classes labels despite the absence of any supervision at all.
To evaluate the quality of unsupervised conditional image generation, we measure performance on two axes: cluster quality and image quality. We estimate clusters by optimizing (10), and show results and comparisons with both recent and classical methods in Table 4. Training details of our method for the additional MLP head can be found in the Appendix A.
Cluster Quality. We measure normalized mutual information (NMI) and clustering accuracy for cluster quality on CIFAR-10 clustered into 10 classes and CIFAR-100(20), which is clustered into 20 super-classes. From Table 4, we observe that on CIFAR-10, U-CTRL results in an NMI that is almost double that of the existing SOTA on both GAN-based and VAE-based methods, with significantly improved clustering accuracy. Unlike many baselines, we also demonstrate that our method scales to the more challenging CIFAR-100(20) dataset, where it also significantly outperforms alternatives. Our improved clustering quality suggests potential for improving unsupervised conditional image generation, which relies on first finding statistically (and hence visually) meaningful clusters.
Image Quality. We use Frechet Inception Distance (FID) (Heusel et al., 2017) and Inception Score (IS) (Salimans et al., 2016) to measure image quality. From Table 4, it is evident that U-CTRL maintains competitive image quality compared to other methods, measured both by FID and IS. We also compare original images against reconstructed ones in Figure 3, where we see that the original 𝑿𝑿\bm{X} is very similar to the reconstructed 𝑿^^𝑿\hat{\bm{X}}; U-CTRL indeed achieves very good sample-wise self-consistency.
Unsupervised Conditional Image Generation. In Figure 4, we visualize images generated from the ten unsupervised clusters from (10). Each block represents one cluster and each row represents one principal component for each cluster. Despite learning and training without labels, the model not only organizes samples into correct clusters, but is also able to preserve statistical diversities within each cluster/class. We can easily recover the diversity within each cluster by computing different principal components and then sample and generate accordingly! More detailed illustrations with more samples is provided in Appendix B.
As shown in the previous section, on datasets like CIFAR-10, CIFAR-100, and Tiny-ImageNet, our framework is able to achieve representation quality on par with the best discriminative self-supervised learning methods. A clear advantage of this is computational efficiency; only a single representation needs to be trained for a much broader set of tasks. This subsection aims to provide additional insights on how a unified model can be more beneficial for a broader range of tasks.
Domain Transfer. Regenerating images is demanding on the encoder, which is required to produce a more informative representation than contrastive training would. We hypothesize that the encoder trained with generative task may retain more information about the image and allow the representation to generalize better. To verify this, we compare the accuracy on CIFAR-100 using models learned from CIFAR-10 in Table 5. When compared to purely discriminative self-supervised learning models, we observe that U-CTRL is 4 percent better than other methods on classification accuracy.
Visualization of Emerged Structures. The representations learned by U-CTRL are significantly different from those learned from previous either discriminative and generative methods. To illustrate this, we use t-SNE (Van der Maaten & Hinton, 2008) to visualize the learned representation in 2D. Figure 5 compares the t-SNE of representations learned for CIFAR-10 by U-CTRL and MoCoV2, respectively. It is clear that the representation learned by U-CTRL are much more structured and better organized: classes are more evident, and features within each class form clear piecewise linear structures.
In this work, we proposed an unsupervised formulation of the closed-loop transcription framework (Dai et al., 2022). We experimentally demonstrate that it is possible to learn a unified representation for both discriminative and generative purposes, resulting in highly structured representations. Further, we show that these two purposes mutually benefit each other in various tasks, e.g., conditional image generation and domain tranfers. Compared to the more specialized representations learned in prior works, our results suggest that such a unified representation has the potential in supporting and benefiting a wider range of new tasks. In future work, we believe the learned representations can be further improved by jointly optimizing the feature representation and feature clusters, as suggested in the original rate reduction paper (Chan et al., 2022). Features with high likelihood of belonging to the same cluster can be further linearized and compressed. Due to its unifying nature and the simplicity of the underlying concepts, this new framework may be extended beyond image data, such as sequential or dynamical observations.
Yi Ma acknowledges support from ONR grants N00014-20-1-2002 and N00014-22-1-2102, the joint Simons Foundation-NSF DMS grant #2031899, as well as partial support from Berkeley FHL Vive Center for Enhanced Reality and Berkeley Center for Augmented Cognition, Tsinghua-Berkeley Shenzhen Institute (TBSI) Research Fund, and Berkeley AI Research (BAIR).
Table 6, 7 and Figure 6 give details on the network architecture for the decoder and the encoder networks used for experiments. The black rectangle marked with "conv, s=2" means a convlutional layer with stride 2. The orange rectangle marked with "dconv, s=2" means a deconvolutional layer with stride 2. The "x k" besides red frame means we regard these layers in red frame as a block and stack it k times. All α𝛼\alpha values in Leaky-ReLU (i.e. lReLU) of the encoder are set to 0.20.20.2. We set (nz=128𝑛𝑧128nz=128, nc=3𝑛𝑐3nc=3, k=3𝑘3k=3) for CIFAR-10, (nz=256𝑛𝑧256nz=256, nc=3𝑛𝑐3nc=3, k=4𝑘4k=4) for CIFAR-100, and (nz=256𝑛𝑧256nz=256, nc=3𝑛𝑐3nc=3, k=4𝑘4k=4) for Tiny-ImageNet. As a comparison, ResNet-18 contains around 11 million parameters, whereas our encoder only contains between 4 and 6 million parameters depending on the choice of k.
Table 8 gives details of the network architecture for the linear classifier and Table 9 gives details of the network architecture for the additional MLP head used for unsupervised conditional image generation training.
For all experiments, we use Adam (Kingma & Ba, 2014) as our optimizer, with hyperparameters β1=0.5,β2=0.999formulae-sequencesubscript𝛽10.5subscript𝛽20.999\beta_{1}=0.5,\beta_{2}=0.999. The learning rate is set to be 0.0001. We choose ϵ2=0.2superscriptitalic-ϵ20.2\epsilon^{2}=0.2. For all experiments, we adopt augmentation from SimCLR (Chen et al., 2020b).
For CIFAR-10, CIFAR-100, and Tiny ImageNet, we train our framework with a batch size of 1024 over 20,000 iterations. All experiments are conducted with at most 4 RTX 3090 GPUs. Methods that are compared against in Table 3 are trained with the batch size of 256, because Chen et al. (2020b) observe that purely discriminative methods tend to perform better with smaller batch sizes. Table 2 methods have used their optimal parameters in their github code.
In this subsection, we visualize the reconstruction of ten clusters that are predicted and generated by U-CTRL on the CIFAR-10 training set. Each block in Figure 7 contains both a random sample of reconstructed data in a cluster and the total number of samples within it. Note that CIFAR-10 contains 50,000 training samples, split across 10 classes. As we see in Figure 7, the number of samples in each cluster are very close to 5,000, with the largest deviator (cluster 9) containing 3,942 samples. Without any cues, one can easily identify correspond each unsupervised cluster with a CIFAR-10 class. For a class like ‘bird’, we observe that the model is able to group images of standing birds, flying birds, and bird heads, despite their visual differences.
Building on U-CTRL’s ability to cluster CIFAR-10 samples, we demonstrate the model’s ability to perform unsupervised conditional image generation in Figure 8. In contrast to reconstruction, where images are regenerated from features corresponding to real samples, we generate images based on the feature sampling technique proposed in (Dai et al., 2022). From these results, we observe that the U-CTRL framework maintains in-cluster diversity, and that the diversity can be recovered and visualized via simple principal component analysis.
In this section, we study the significance of the sample-wise constraints and extra rate distortion term in the formulation 7. Table 10 presents the following objectives that we study:
Objective I is the constrained U-CTRL maximin 7.
Objective V is the U-CTRL without the augmentation compression constraint and sample-wise self-consistency constraint.
Table 11 shows the result of a linear probe for representations trained using each objective on CIFAR-10. From the table, it is evident that both constraints and the rate distortion term are pivotal to the success of our framework.
In this section, we verify the significance of MCR2 term ΔR(𝒁,𝒁^)Δ𝑅𝒁^𝒁\Delta R(\bm{Z},\hat{\bm{Z}}) in our method. We do ablation study on CIFAR-10 with the same network and training condition. If we take away MCR2 from our formulation, it changes (11). For simplicity, we call it U-CTRL-noMCR2
Table 12 shows that U-CTRL without the MCR2 not only learns worse representation but also generalizes worse to out of distribution data. Figure 9 visualizes the reconstructed 𝑿^^𝑿\hat{\bm{X}} by U-CTRL-noMCR2. It is clear from the image figure that without the MCR2, the decoder fails to reconstruct high-quality images.
It follows our discussion in the introduction that discriminative tasks and generative tasks together learn feature that benifits each other.
In this section, we verify the stability of our method against different random seeds. We report in Table 13 the accuracy of U-CTRL on CIFAR-10 with different seeds. We observe that the choice of seed has very little impact on performance.
Table: S1.T1: Comparison of the downstream task capabilities of different unsupervised learning methods. UCIG refers to Unsupervised Conditional Image Generation (Hwang et al., 2021).
| Method | Linear Probe | Image Generation | UCIG |
|---|---|---|---|
| SimCLR (Chen et al., 2020b) | ✔ | ✗ | ✗ |
| MOCO-V2 (He et al., 2020) | ✔ | ✗ | ✗ |
| ContraD (Jeong & Shin, 2021) | ✔ | ✔ | ✗ |
| PATCH-VAE (Parmar et al., 2021) | ✔ | ✔ | ✗ |
| CTRL-Binary (Dai et al., 2022) | ✔ | ✔ | ✗ |
| SLOGAN (Hwang et al., 2021) | ✗ | ✔ | ✔ |
| U-CTRL (ours) | ✔ | ✔ | ✔ |
Table: S4.T2: Comparison of classification accuracy on CIFAR-10, CIFAR-100, and Tiny-ImageNet with other generative self-supervised learning methods. U-CTRL is clearly better.
| Method | CIFAR-10 | CIFAR-100 | Tiny-ImageNet |
| Accuracy | Accuracy | Accuracy | |
| GAN based methods | |||
| SSGAN-LA(Hou et al., 2021) | 0.803 | 0.543 | 0.344 |
| DAGAN+(Antoniou et al., 2017) | 0.772 | 0.519 | 0.224 |
| ContraD(Jeong & Shin, 2021) | 0.852 | 0.514 | - |
| VAE based methods | |||
| PATCH-VAE (Parmar et al., 2021) | 0.471 | 0.325 | - |
| β𝛽\beta-VAE (Higgins et al., 2016) | 0.531 | 0.315 | - |
| CTRL based methods | |||
| CTRL-Binary(Dai et al., 2022) | 0.599 | - | - |
| U-CTRL (ours) | 0.874 | 0.552 | 0.360 |
Table: S4.T4: Comparison of the quality of UCIG on CIFAR-10 and CIFAR-100(20). Many of the methods compared do not provide code that scales up to CIFAR-100(20), in which case we leave the corresponding table cell blank.
| NMI | Accuracy | FID↓↓\downarrow | IS↑↑\uparrow | NMI | Accuracy | FID↓↓\downarrow | IS↑↑\uparrow | |
|---|---|---|---|---|---|---|---|---|
| GAN based methods | ||||||||
| Self-Conditioned GAN (Liu et al., 2020) | 0.333 | 0.117 | 18.0 | 7.7 | 0.214 | 0.092 | 24.1 | 5.2 |
| SLOGAN (Hwang et al., 2021) | 0.340 | - | 20.6 | - | - | - | - | - |
| VAE based methods | ||||||||
| GMVAE(Dilokthanakul et al., 2016) | - | 0.247 | - | - | - | - | - | - |
| Variational Clustering | - | 0.445 | - | - | - | - | - | - |
| CTRL based methods | ||||||||
| U-CTRL (ours) | 0.658 | 0.799 | 17.4 | 8.1 | 0.374 | 0.433 | 20.1 | 7.7 |
Table: A1.T6: Network architecture of the decoder g(⋅,η)𝑔⋅𝜂g(\cdot,\eta).
| 𝒛∈ℝ1×1×nz𝒛superscriptℝ11𝑛𝑧\bm{z}\in\mathbb{R}^{1\times 1\times nz} |
|---|
| ResBlockUp. 256 |
| ResBlockUp. 128 |
| ResBlockUp. 64 |
| 4 ×\times 4, stride=2, pad=1 deconv. 1 Tanh |
Table: A3.T10: Five different objective functions for U-CTRL.
| Objective I: | maxθminηR(𝒁)+ΔR(𝒁,𝒁^) s.t.∑i∈NΔR(𝒛i,𝒛^i)=0,and∑i∈NΔR(𝒛i,𝒛ai)=0formulae-sequencesubscript𝜃subscript𝜂𝑅𝒁Δ𝑅𝒁^𝒁 s.t.subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscript^𝒛𝑖0andsubscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscriptsubscript𝒛𝑎𝑖0\max_{\theta}\min_{\eta}R(\bm{Z})+\Delta R(\bm{Z},\hat{\bm{Z}})\mbox{ s.t.}\sum_{i\in N}\Delta R(\bm{z}^{i},\hat{\bm{z}}^{i})=0,\mbox{and}\sum_{i\in N}\Delta R(\bm{z}^{i},\bm{z}_{a}^{i})=0 |
| Objective II: | maxθminηR(𝒁)+ΔR(𝒁,𝒁^) s.t.∑i∈NΔR(𝒛i,𝒛^i)=0subscript𝜃subscript𝜂𝑅𝒁Δ𝑅𝒁^𝒁 s.t.subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscript^𝒛𝑖0\max_{\theta}\min_{\eta}R(\bm{Z})+\Delta R(\bm{Z},\hat{\bm{Z}})\mbox{ s.t.}\sum_{i\in N}\Delta R(\bm{z}^{i},\hat{\bm{z}}^{i})=0 |
| Objective III: | maxθminηR(𝒁)+ΔR(𝒁,𝒁^) s.t.∑i∈NΔR(𝒛i,𝒛ai)=0subscript𝜃subscript𝜂𝑅𝒁Δ𝑅𝒁^𝒁 s.t.subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscriptsubscript𝒛𝑎𝑖0\max_{\theta}\min_{\eta}R(\bm{Z})+\Delta R(\bm{Z},\hat{\bm{Z}})\mbox{ s.t.}\sum_{i\in N}\Delta R(\bm{z}^{i},\bm{z}_{a}^{i})=0 |
| Objective IV: | maxθminηΔR(𝒁,𝒁^) s.t.∑i∈NΔR(𝒛i,𝒛^i)=0,and∑i∈NΔR(𝒛i,𝒛ai)=0formulae-sequencesubscript𝜃subscript𝜂Δ𝑅𝒁^𝒁 s.t.subscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscript^𝒛𝑖0andsubscript𝑖𝑁Δ𝑅superscript𝒛𝑖superscriptsubscript𝒛𝑎𝑖0\max_{\theta}\min_{\eta}\Delta R(\bm{Z},\hat{\bm{Z}})\mbox{ s.t.}\sum_{i\in N}\Delta R(\bm{z}^{i},\hat{\bm{z}}^{i})=0,\mbox{and}\sum_{i\in N}\Delta R(\bm{z}^{i},\bm{z}_{a}^{i})=0 |
| Objective V: | maxθminηR(𝒁)+ΔR(𝒁,𝒁^)subscript𝜃subscript𝜂𝑅𝒁Δ𝑅𝒁^𝒁\max_{\theta}\min_{\eta}R(\bm{Z})+\Delta R(\bm{Z},\hat{\bm{Z}}) |
| Objective VI: | maxθminηΔR(𝒁,𝒁^)subscript𝜃subscript𝜂Δ𝑅𝒁^𝒁\max_{\theta}\min_{\eta}\Delta R(\bm{Z},\hat{\bm{Z}}) |
Table: A3.T12: Ablation study on the significance of MCR2 in U-CTRL.
| Accuracy on CIFAR-10 | Transfer Accuracy on CIFAR-100 | ||
|---|---|---|---|
| U-CTRL | 0.874 | 0.481 | |
| U-CTRL-noMCR2 | 0.836 | 0.418 |
Overall framework of closed-loop transcription for unsupervised learning. Two additional constraints are imposed on the Binary-CTRL method proposed in prior work (Dai et al., 2022): 1) self-consistency for sample-wise features 𝒛isuperscript𝒛𝑖\bm{z}^{i} and 𝒛^isuperscript^𝒛𝑖\hat{\bm{z}}^{i}, say 𝒛i≈𝒛^isuperscript𝒛𝑖superscript^𝒛𝑖\bm{z}^{i}\approx\hat{\bm{z}}^{i}; and 2) invariance/similarity among features of augmented samples 𝒛isuperscript𝒛𝑖\bm{z}^{i} and 𝒛aisuperscriptsubscript𝒛𝑎𝑖\bm{z}{a}^{i}, say 𝒛i≈𝒛ai=f(τ(𝒙i),θ)superscript𝒛𝑖superscriptsubscript𝒛𝑎𝑖𝑓𝜏superscript𝒙𝑖𝜃\bm{z}^{i}\approx\bm{z}{a}^{i}=f(\tau(\bm{x}^{i}),\theta), where 𝒙ai=τ(𝒙i)subscriptsuperscript𝒙𝑖𝑎𝜏superscript𝒙𝑖\bm{x}^{i}_{a}=\tau(\bm{x}^{i}) is an augmentation of sample 𝒙isuperscript𝒙𝑖\bm{x}^{i} via some transformation τ(⋅)𝜏⋅\tau(\cdot).
(a) CIFAR-10
(c) Tiny ImageNet
Unsupervised conditional image generation from each cluster of CIFAR-10, using U-CTRL. Images from different rows mean generation from different principal components of each cluster.
(b) MoCoV2
(a) ResBlock Up architecture
(c) Cluster 3
$$ \Delta R\big(\bm Z|\bm \Pi) \doteq \underbrace{\frac{1}{2}\log\det\left(\I + \frac{d}{N\epsilon^{2}}\Z\Z^{\top}\right)}{ \large{R(\Z)}} - \underbrace{\sum{j=1}^{k} \frac{\tr(\bm{\Pi}_j)}{2N}\log\det\left(\I + \frac{d}{\tr(\bm{\Pi}_j)\epsilon^{2}}\Z\bm{\Pi}j\Z^{\top}\right)}{ \large{R^c}}. \label{eqn:maximal-rate-reduction} $$ \tag{eqn:maximal-rate-reduction}
$$ \bm{X}\xrightarrow{\hskip 5.69054ptf(\bm{x},\theta)\hskip 5.69054pt}\bm{Z}\xrightarrow{\hskip 5.69054ptg(\bm{z},\eta)\hskip 5.69054pt}\hat{\bm{X}}\xrightarrow{\hskip 5.69054ptf(\bm{x},\theta)\hskip 5.69054pt}\ \hat{\bm{Z}}. $$ \tag{S3.E2}
$$ \Delta R\big(\Z, \hat{\Z}\big) \doteq R\big(\Z \cup \hat{\Z}\big) - \frac{1}{2} \big( R\big(\Z) + R\big(\hat \Z)\big). $$
$$ \max_\theta \min_\eta \quad \Delta R(\Z, \hat{\Z}) \label{eqn:CTRL-Binary}\vspace{-2mm} $$ \tag{eqn:CTRL-Binary}
$$ \max_\theta \min_\eta \quad & R(\Z) + \Delta R(\Z, \hat{\Z}) \label{eqn:constrained_maxmin}\ \mbox{subject to} \quad & \sum_{i\in N} \Delta R(\z^i, \hat{\z}^i) = 0, ;; \mbox{and} ;; \sum_{i\in N} \Delta R(\z^i, \z_{a}^i) = 0. \nonumber \vspace{-2mm} $$ \tag{eqn:constrained_maxmin}
Table 13: Ablation study on varying random seeds.
| Method | Linear Probe | Image Generation | UCIG |
|---|---|---|---|
| SimCLR (Chen et al., 2020b) | 4 | 7 | 7 |
| MOCO-V2 (He et al., 2020) | 4 | 7 | 7 |
| ContraD (Jeong &Shin, 2021) | 4 | 4 | 7 |
| PATCH-VAE (Parmar et al., 2021) | 4 | 4 | 7 |
| CTRL-Binary (Dai et al., 2022) | 4 | 4 | 7 |
| SLOGAN (Hwang et al., 2021) | 7 | 4 | 4 |
| U-CTRL (ours) | 4 | 4 | 4 |
| Method | CIFAR-10 Accuracy | CIFAR-100 Accuracy | Tiny-ImageNet Accuracy |
|---|---|---|---|
| GAN based methods SSGAN-LA(Hou et al., 2021) | 0.803 | 0.543 | 0.344 |
| DAGAN+(Antoniou et al., 2017) | 0.772 | 0.519 | 0.224 |
| ContraD(Jeong &Shin, 2021) | 0.852 | 0.514 | - |
| VAE based methods PATCH-VAE (Parmar et al., 2021) | 0.471 | 0.325 | - |
| β -VAE (Higgins et al., 2016) | 0.531 | 0.315 | - |
| CTRL based methods CTRL-Binary(Dai et al., 2022) | 0.599 | - | - |
| U-CTRL (ours) | 0.874 | 0.552 | 0.360 |
| Method | CIFAR-10 Accuracy | CIFAR-100 Accuracy | Tiny-ImageNet Accuracy |
|---|---|---|---|
| SIMCLR | 0.869 | 0.545 | 0.359 |
| MoCoV2 | 0.872 | 0.589 | 0.365 |
| BYOL | 0.883 | 0.581 | 0.371 |
| U-CTRL (ours) | 0.874 | 0.552 | 0.36 |
| CIFAR-10 | CIFAR-10 | CIFAR-10 | CIFAR-10 | CIFAR-100(20) | CIFAR-100(20) | CIFAR-100(20) | CIFAR-100(20) | |
|---|---|---|---|---|---|---|---|---|
| Method | NMI | Accuracy | FID ↓ | IS ↑ | NMI | Accuracy | FID ↓ | IS ↑ |
| GAN based methods Self-Conditioned GAN (Liu et al., 2020) | 0.333 | 0.117 | 18.0 | 7.7 | 0.214 | 0.092 | 24.1 | 5.2 |
| SLOGAN (Hwang et al., 2021) | 0.340 | - | 20.6 | - | - | - | - | - |
| VAE based methods GMVAE(Dilokthanakul et al., 2016) | - | 0.247 | - | - | - | - | - | - |
| Variational Clustering | - | 0.445 | - | - | - | - | - | - |
| CTRL based methods U-CTRL (ours) | 0.658 | 0.799 | 17.4 | 8.1 | 0.374 | 0.433 | 20.1 | 7.7 |
| Method | SIMCLR | MoCoV2 | BYOL | U-CTRL |
|---|---|---|---|---|
| Accuracy | 0.422 | 0.436 | 0.437 | 0.481 |
| z ∈ R 1 × 1 × nz |
|---|
| ResBlockUp. 256 |
| ResBlockUp. 128 |
| ResBlockUp. 64 |
| 4 × 4, stride=2, pad=1 deconv. 1 Tanh |
| Image x ∈ R 32 × 32 × nc |
|---|
| ResBlockDown 64 |
| ResBlockDown 128 |
| ResBlockDown 256 |
| 4 × 4, stride=1, pad=0 conv nz |
| z ∈ R 1 × 1 × nz |
|---|
| Linear(nz, number of class) |
| z ∈ R 1 × 1 × nz |
|---|
| Linear(nz, nz) ReLU |
| Linear(nz, number of clusters) |
| Method | Objective I | Objective II | Objective III | Objective IV | Objective V | Objective VI |
|---|---|---|---|---|---|---|
| Accuracy | 0.874 | 0.578 | 0.644 | 0.522 | 0.633 | 0.599 |
| Accuracy on CIFAR-10 | Transfer Accuracy on CIFAR-100 | |
|---|---|---|
| U-CTRL | 0.874 | 0.481 |
| U-CTRL-noMCR 2 | 0.836 | 0.418 |
| Random Seed | 1 | 5 | 10 | 15 | 100 |
|---|---|---|---|---|---|
| Accuracy | 0.874 | 0.876 | 0.87 | 0.874 | 0.871 |
$$ \X \xrightarrow{\hspace{2mm} f(\x, \theta)\hspace{2mm}} \Z \xrightarrow{\hspace{2mm} g(\z,\eta) \hspace{2mm}} \hat \X \xrightarrow{\hspace{2mm} f(\x, \theta)\hspace{2mm}} \ \hat \Z. $$
$$ \sum_{i\in N} \Delta R(\z^i,\hat{\z}^i) = 0. \label{eqn:sample-self-consistency}\vspace{-2mm} $$ \tag{eqn:sample-self-consistency}
| Method | Linear Probe | Image Generation | UCIG |
|---|---|---|---|
| SimCLR (Chen et al., 2020b) | 4 | 7 | 7 |
| MOCO-V2 (He et al., 2020) | 4 | 7 | 7 |
| ContraD (Jeong &Shin, 2021) | 4 | 4 | 7 |
| PATCH-VAE (Parmar et al., 2021) | 4 | 4 | 7 |
| CTRL-Binary (Dai et al., 2022) | 4 | 4 | 7 |
| SLOGAN (Hwang et al., 2021) | 7 | 4 | 4 |
| U-CTRL (ours) | 4 | 4 | 4 |
| Method | CIFAR-10 Accuracy | CIFAR-100 Accuracy | Tiny-ImageNet Accuracy |
|---|---|---|---|
| GAN based methods SSGAN-LA(Hou et al., 2021) | 0.803 | 0.543 | 0.344 |
| DAGAN+(Antoniou et al., 2017) | 0.772 | 0.519 | 0.224 |
| ContraD(Jeong &Shin, 2021) | 0.852 | 0.514 | - |
| VAE based methods PATCH-VAE (Parmar et al., 2021) | 0.471 | 0.325 | - |
| β -VAE (Higgins et al., 2016) | 0.531 | 0.315 | - |
| CTRL based methods CTRL-Binary(Dai et al., 2022) | 0.599 | - | - |
| U-CTRL (ours) | 0.874 | 0.552 | 0.360 |
| Method | CIFAR-10 Accuracy | CIFAR-100 Accuracy | Tiny-ImageNet Accuracy |
|---|---|---|---|
| SIMCLR | 0.869 | 0.545 | 0.359 |
| MoCoV2 | 0.872 | 0.589 | 0.365 |
| BYOL | 0.883 | 0.581 | 0.371 |
| U-CTRL (ours) | 0.874 | 0.552 | 0.36 |
| CIFAR-10 | CIFAR-10 | CIFAR-10 | CIFAR-10 | CIFAR-100(20) | CIFAR-100(20) | CIFAR-100(20) | CIFAR-100(20) | |
|---|---|---|---|---|---|---|---|---|
| Method | NMI | Accuracy | FID ↓ | IS ↑ | NMI | Accuracy | FID ↓ | IS ↑ |
| GAN based methods Self-Conditioned GAN (Liu et al., 2020) | 0.333 | 0.117 | 18.0 | 7.7 | 0.214 | 0.092 | 24.1 | 5.2 |
| SLOGAN (Hwang et al., 2021) | 0.340 | - | 20.6 | - | - | - | - | - |
| VAE based methods GMVAE(Dilokthanakul et al., 2016) | - | 0.247 | - | - | - | - | - | - |
| Variational Clustering | - | 0.445 | - | - | - | - | - | - |
| CTRL based methods U-CTRL (ours) | 0.658 | 0.799 | 17.4 | 8.1 | 0.374 | 0.433 | 20.1 | 7.7 |
| Method | SIMCLR | MoCoV2 | BYOL | U-CTRL |
|---|---|---|---|---|
| Accuracy | 0.422 | 0.436 | 0.437 | 0.481 |
| z ∈ R 1 × 1 × nz |
|---|
| ResBlockUp. 256 |
| ResBlockUp. 128 |
| ResBlockUp. 64 |
| 4 × 4, stride=2, pad=1 deconv. 1 Tanh |
| Image x ∈ R 32 × 32 × nc |
|---|
| ResBlockDown 64 |
| ResBlockDown 128 |
| ResBlockDown 256 |
| 4 × 4, stride=1, pad=0 conv nz |
| z ∈ R 1 × 1 × nz |
|---|
| Linear(nz, number of class) |
| z ∈ R 1 × 1 × nz |
|---|
| Linear(nz, nz) ReLU |
| Linear(nz, number of clusters) |
| Method | Objective I | Objective II | Objective III | Objective IV | Objective V | Objective VI |
|---|---|---|---|---|---|---|
| Accuracy | 0.874 | 0.578 | 0.644 | 0.522 | 0.633 | 0.599 |
| Accuracy on CIFAR-10 | Transfer Accuracy on CIFAR-100 | |
|---|---|---|
| U-CTRL | 0.874 | 0.481 |
| U-CTRL-noMCR 2 | 0.836 | 0.418 |
| Random Seed | 1 | 5 | 10 | 15 | 100 |
|---|---|---|---|---|---|
| Accuracy | 0.874 | 0.876 | 0.87 | 0.874 | 0.871 |












$$ & \max_{\theta}; R(\Z) + \Delta{R(\Z, \hat{\Z})-\lambda_{1}\sum_{i\in N} \Delta R(\z^i, \z_{a}^i)} -\lambda_{2}\sum_{i\in N} \Delta R(\z^i, \hat{\z}^i) \label{eqn:constrained_max}; \ $$ \tag{eqn:constrained_max}
References
[Bengio+chapter2007] Bengio, Yoshua, LeCun, Yann. (2007). Scaling Learning Algorithms Towards {AI. Large Scale Kernel Machines.
[Hinton06] Hinton, Geoffrey E., Osindero, Simon, Teh, Yee Whye. (2006). A Fast Learning Algorithm for Deep Belief Nets. Neural Computation.
[goodfellow2016deep] Goodfellow, Ian, Bengio, Yoshua, Courville, Aaron, Bengio, Yoshua. (2016). Deep learning.
[goodfellow2014generative] Goodfellow, Ian, Pouget-Abadie, Jean, Mirza, Mehdi, Xu, Bing, Warde-Farley, David, Ozair, Sherjil, Courville, Aaron, Bengio, Yoshua. (2014). Generative adversarial nets. Advances in neural information processing systems.
[noroozi2016unsupervised] Noroozi, Mehdi, Favaro, Paolo. (2016). Unsupervised learning of visual representations by solving jigsaw puzzles. European conference on computer vision.
[komodakis2018unsupervised] Komodakis, Nikos, Gidaris, Spyros. (2018). Unsupervised representation learning by predicting image rotations. International Conference on Learning Representations (ICLR).
[kingma2013auto] Kingma, Diederik P, Welling, Max. (2013). Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114.
[dai2022ctrl] Dai, Xili, Tong, Shengbang, Li, Mingyang, Wu, Ziyang, Psenka, Michael, Chan, Kwan Ho Ryan, Zhai, Pengyuan, Yu, Yaodong, Yuan, Xiaojun, Shum, Heung-Yeung, others. (2022). CTRL: Closed-Loop Transcription to an LDR via Minimaxing Rate Reduction. Entropy.
[lecun1998gradient] LeCun, Yann, Bottou, L{'e. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE.
[krizhevsky2014cifar] Krizhevsky, Alex, Nair, Vinod, Hinton, Geoffrey. (2014). The {CIFAR. online: http://www.cs.toronto.edu/kriz/cifar.html.
[hou2021self] Hou, Liang, Shen, Huawei, Cao, Qi, Cheng, Xueqi. (2021). Self-Supervised GANs with Label Augmentation. Advances in Neural Information Processing Systems.
[yu2020learning] Yu, Yaodong, Chan, Kwan Ho Ryan, You, Chong, Song, Chaobing, Ma, Yi. (2020). Learning diverse and discriminative representations via the principle of maximal coding rate reduction. Advances in Neural Information Processing Systems.
[krizhevsky2009learning] Krizhevsky, Alex, others. (2009). Learning multiple layers of features from tiny images. arXiv preprint arXiv:1312.6114.
[ma2007segmentation] Ma, Yi, Derksen, Harm, Hong, Wei, Wright, John. (2007). Segmentation of multivariate mixed data via lossy data coding and compression. PAMI.
[parmar2021dual] Parmar, Gaurav, Li, Dacheng, Lee, Kwonjoon, Tu, Zhuowen. (2021). Dual contradistinctive generative autoencoder. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[jeong2021training] Jeong, Jongheon, Shin, Jinwoo. (2021). Training gans with stronger augmentations via contrastive discriminator. arXiv preprint arXiv:2103.09742.
[dilokthanakul2016deep] Dilokthanakul, Nat, Mediano, Pedro AM, Garnelo, Marta, Lee, Matthew CH, Salimbeni, Hugh, Arulkumaran, Kai, Shanahan, Murray. (2016). Deep unsupervised clustering with gaussian mixture variational autoencoders. arXiv preprint arXiv:1611.02648.
[li2022neural] Li, Zengyi, Chen, Yubei, LeCun, Yann, Sommer, Friedrich T. (2022). Neural Manifold Clustering and Embedding. arXiv preprint arXiv:2201.10000.
[chen2020generative] Chen, Mark, Radford, Alec, Child, Rewon, Wu, Jeffrey, Jun, Heewoo, Luan, David, Sutskever, Ilya. (2020). Generative pretraining from pixels. International Conference on Machine Learning.
[kingma2014adam] Kingma, Diederik P, Ba, Jimmy. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
[chen2020simple] Chen, Ting, Kornblith, Simon, Norouzi, Mohammad, Hinton, Geoffrey. (2020). A simple framework for contrastive learning of visual representations. International conference on machine learning.
[li2018learning] Li, Xiaopeng, Chen, Zhourong, Poon, Leonard KM, Zhang, Nevin L. (2018). Learning latent superstructures in variational autoencoders for deep multidimensional clustering. arXiv preprint arXiv:1803.05206.
[jiang2016variational] Jiang, Zhuxi, Zheng, Yin, Tan, Huachun, Tang, Bangsheng, Zhou, Hanning. (2016). Variational deep embedding: An unsupervised and generative approach to clustering. arXiv preprint arXiv:1611.05148.
[prasad2020variational] Prasad, Vignesh, Das, Dipanjan, Bhowmick, Brojeshwar. (2020). Variational clustering: Leveraging variational autoencoders for image clustering. 2020 International Joint Conference on Neural Networks (IJCNN).
[dumoulin2016adversarially] Dumoulin, Vincent, Belghazi, Ishmael, Poole, Ben, Mastropietro, Olivier, Lamb, Alex, Arjovsky, Martin, Courville, Aaron. (2016). Adversarially learned inference. arXiv preprint arXiv:1606.00704.
[chen2019self] Chen, Ting, Zhai, Xiaohua, Ritter, Marvin, Lucic, Mario, Houlsby, Neil. (2019). Self-supervised gans via auxiliary rotation loss. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition.
[grill2020bootstrap] Grill, Jean-Bastien, Strub, Florian, Altch{'e. (2020). Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems.
[zbontar2021barlow] Zbontar, Jure, Jing, Li, Misra, Ishan, LeCun, Yann, Deny, St{'e. (2021). Barlow twins: Self-supervised learning via redundancy reduction. International Conference on Machine Learning.
[gupta2020patchvae] Gupta, Kamal, Singh, Saurabh, Shrivastava, Abhinav. (2020). Patchvae: Learning local latent codes for recognition. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[he2020momentum] He, Kaiming, Fan, Haoqi, Wu, Yuxin, Xie, Saining, Girshick, Ross. (2020). Momentum contrast for unsupervised visual representation learning. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition.
[chen2022context] Chen, Xiaokang, Ding, Mingyu, Wang, Xiaodi, Xin, Ying, Mo, Shentong, Wang, Yunhao, Han, Shumin, Luo, Ping, Zeng, Gang, Wang, Jingdong. (2022). Context autoencoder for self-supervised representation learning. arXiv preprint arXiv:2202.03026.
[higgins2016beta] Higgins, Irina, Matthey, Loic, Pal, Arka, Burgess, Christopher, Glorot, Xavier, Botvinick, Matthew, Mohamed, Shakir, Lerchner, Alexander. (2016). beta-vae: Learning basic visual concepts with a constrained variational framework. arXiv preprint arXiv:1804.03599.
[he2021masked] He, Kaiming, Chen, Xinlei, Xie, Saining, Li, Yanghao, Doll{'a. (2021). Masked autoencoders are scalable vision learners. arXiv preprint arXiv:2111.06377.
[chen2016infogan] Chen, Xi, Duan, Yan, Houthooft, Rein, Schulman, John, Sutskever, Ilya, Abbeel, Pieter. (2016). Infogan: Interpretable representation learning by information maximizing generative adversarial nets. Advances in neural information processing systems.
[mukherjee2019clustergan] Mukherjee, Sudipto, Asnani, Himanshu, Lin, Eugene, Kannan, Sreeram. (2019). Clustergan: Latent space clustering in generative adversarial networks. Proceedings of the AAAI conference on artificial intelligence.
[liu2020diverse] Liu, Steven, Wang, Tongzhou, Bau, David, Zhu, Jun-Yan, Torralba, Antonio. (2020). Diverse image generation via self-conditioned gans. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition.
[donahue2016adversarial] Donahue, Jeff, Kr{. (2016). Adversarial feature learning. arXiv preprint arXiv:1605.09782.
[hwang2021stein] Hwang, Uiwon, Kim, Heeseung, Jung, Dahuin, Jang, Hyemi, Lee, Hyungyu, Yoon, Sungroh. (2021). Stein Latent Optimization for Generative Adversarial Networks. arXiv preprint arXiv:2106.05319.
[radford2015unsupervised] Radford, Alec, Metz, Luke, Chintala, Soumith. (2015). Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434.
[tong2022incremental] Shengbang Tong, Xili Dai, Ziyang Wu, Mingyang Li, Brent Yi, Yi Ma. (2022). Incremental Learning of Structured Memory via Closed-Loop Transcription. arXiv:2202.05411.
[chan2021redunet] Kwan Ho Ryan Chan, Yaodong Yu, Chong You, Haozhi Qi, John Wright, Yi Ma. (2022). Redu{N. Journal of Machine Learning Research.
[2020Vandeven] Ven, Gido M, Siegelmann, Hava T, Tolias, Andreas S, others. (2020). Brain-inspired replay for continual learning with artificial neural networks. Nature Communications.
[Josselyn2020MemoryER] Sheena A. Josselyn, Susumu Tonegawa. (2020). Memory engrams: Recalling the past and imagining the future. Science.
[antoniou2017data] Antoniou, Antreas, Storkey, Amos, Edwards, Harrison. (2017). Data augmentation generative adversarial networks. arXiv preprint arXiv:1711.04340.
[heusel2017gans] Heusel, Martin, Ramsauer, Hubert, Unterthiner, Thomas, Nessler, Bernhard, Hochreiter, Sepp. (2017). Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems.
[salimans2016improved] Salimans, Tim, Goodfellow, Ian, Zaremba, Wojciech, Cheung, Vicki, Radford, Alec, Chen, Xi. (2016). Improved techniques for training gans. Advances in neural information processing systems.
[Keller2018-ez] Keller, Georg B, Mrsic-Flogel, Thomas D. Predictive Processing: A Canonical Cortical Computation. Neuron.
[ayub2021eec] Ayub, Ali, Wagner, Alan R. (2021). Eec: Learning to encode and regenerate images for continual learning. arXiv preprint arXiv:2101.04904.
[gatys2016image] Gatys, Leon A, Ecker, Alexander S, Bethge, Matthias. (2016). Image style transfer using convolutional neural networks. Proceedings of the IEEE conference on computer vision and pattern recognition.
[van2008visualizing] Van der Maaten, Laurens, Hinton, Geoffrey. (2008). Visualizing data using t-SNE.. Journal of machine learning research.
[Kramer1991NonlinearPC] Mark A. Kramer. (1991). Nonlinear principal component analysis using autoassociative neural networks. Aiche Journal.
[deng2009imagenet] Deng, Jia, Dong, Wei, Socher, Richard, Li, Li-Jia, Li, Kai, Fei-Fei, Li. (2009). Imagenet: A large-scale hierarchical image database. 2009 IEEE conference on computer vision and pattern recognition.
[image-similarity] Bardes, Adrien, Ponce, Jean, LeCun, Yann. (2021). Vicreg: Variance-invariance-covariance regularization for self-supervised learning. arXiv preprint arXiv:2105.04906. doi:10.1109/TIP.2003.819861.
[chen2020simsiam] Kim, Saehoon, Kim, Sungwoong, Lee, Juho. (2021). Hybrid Generative-Contrastive Representation Learning. arXiv preprint arXiv:2106.06162.
[falcon2021aavae] Falcon, William, Jha, Ananya Harsh, Koker, Teddy, Cho, Kyunghyun. (2021). AAVAE: Augmentation-Augmented Variational Autoencoders. arXiv preprint arXiv:2107.12329.
[bib1] Antoniou et al. (2017) Antreas Antoniou, Amos Storkey, and Harrison Edwards. Data augmentation generative adversarial networks. arXiv preprint arXiv:1711.04340, 2017.
[bib2] Bardes et al. (2021) Adrien Bardes, Jean Ponce, and Yann LeCun. Vicreg: Variance-invariance-covariance regularization for self-supervised learning. arXiv preprint arXiv:2105.04906, 2021.
[bib3] Chan et al. (2022) Kwan Ho Ryan Chan, Yaodong Yu, Chong You, Haozhi Qi, John Wright, and Yi Ma. ReduNet: A white-box deep network from the principle of maximizing rate reduction. Journal of Machine Learning Research, 23(114):1–103, 2022. URL http://jmlr.org/papers/v23/21-0631.html.
[bib4] Chen et al. (2020a) Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and Ilya Sutskever. Generative pretraining from pixels. In International Conference on Machine Learning, pp. 1691–1703. PMLR, 2020a.
[bib5] Chen et al. (2019) Ting Chen, Xiaohua Zhai, Marvin Ritter, Mario Lucic, and Neil Houlsby. Self-supervised gans via auxiliary rotation loss. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 12154–12163, 2019.
[bib6] Chen et al. (2020b) Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597–1607. PMLR, 2020b.
[bib7] Chen et al. (2016) Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. Advances in neural information processing systems, 29, 2016.
[bib8] Chen et al. (2022) Xiaokang Chen, Mingyu Ding, Xiaodi Wang, Ying Xin, Shentong Mo, Yunhao Wang, Shumin Han, Ping Luo, Gang Zeng, and Jingdong Wang. Context autoencoder for self-supervised representation learning. arXiv preprint arXiv:2202.03026, 2022.
[bib9] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In CVPR, 2020.
[bib10] Dai et al. (2022) Xili Dai, Shengbang Tong, Mingyang Li, Ziyang Wu, Michael Psenka, Kwan Ho Ryan Chan, Pengyuan Zhai, Yaodong Yu, Xiaojun Yuan, Heung-Yeung Shum, et al. Ctrl: Closed-loop transcription to an ldr via minimaxing rate reduction. Entropy, 24(4):456, 2022.
[bib11] Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. Ieee, 2009.
[bib12] Dilokthanakul et al. (2016) Nat Dilokthanakul, Pedro AM Mediano, Marta Garnelo, Matthew CH Lee, Hugh Salimbeni, Kai Arulkumaran, and Murray Shanahan. Deep unsupervised clustering with gaussian mixture variational autoencoders. arXiv preprint arXiv:1611.02648, 2016.
[bib13] Donahue et al. (2016) Jeff Donahue, Philipp Krähenbühl, and Trevor Darrell. Adversarial feature learning. arXiv preprint arXiv:1605.09782, 2016.
[bib14] Dumoulin et al. (2016) Vincent Dumoulin, Ishmael Belghazi, Ben Poole, Olivier Mastropietro, Alex Lamb, Martin Arjovsky, and Aaron Courville. Adversarially learned inference. arXiv preprint arXiv:1606.00704, 2016.
[bib15] Falcon et al. (2021) William Falcon, Ananya Harsh Jha, Teddy Koker, and Kyunghyun Cho. Aavae: Augmentation-augmented variational autoencoders. arXiv preprint arXiv:2107.12329, 2021.
[bib16] Goodfellow et al. (2014) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014.
[bib17] Grill et al. (2020a) Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems, 33:21271–21284, 2020a.
[bib18] Grill et al. (2020b) Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H. Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, Rémi Munos, and Michal Valko. Bootstrap your own latent: A new approach to self-supervised learning. In NeurIPS, 2020b.
[bib19] Gupta et al. (2020) Kamal Gupta, Saurabh Singh, and Abhinav Shrivastava. Patchvae: Learning local latent codes for recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4746–4755, 2020.
[bib20] He et al. (2020) Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9729–9738, 2020.
[bib21] He et al. (2021) Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. arXiv preprint arXiv:2111.06377, 2021.
[bib22] Heusel et al. (2017) Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
[bib23] Higgins et al. (2016) Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-vae: Learning basic visual concepts with a constrained variational framework. arXiv preprint arXiv:1804.03599, 2016.
[bib24] Hou et al. (2021) Liang Hou, Huawei Shen, Qi Cao, and Xueqi Cheng. Self-supervised gans with label augmentation. Advances in Neural Information Processing Systems, 34, 2021.
[bib25] Hwang et al. (2021) Uiwon Hwang, Heeseung Kim, Dahuin Jung, Hyemi Jang, Hyungyu Lee, and Sungroh Yoon. Stein latent optimization for generative adversarial networks. arXiv preprint arXiv:2106.05319, 2021.
[bib26] Jongheon Jeong and Jinwoo Shin. Training gans with stronger augmentations via contrastive discriminator. arXiv preprint arXiv:2103.09742, 2021.
[bib27] Jiang et al. (2016) Zhuxi Jiang, Yin Zheng, Huachun Tan, Bangsheng Tang, and Hanning Zhou. Variational deep embedding: An unsupervised and generative approach to clustering. arXiv preprint arXiv:1611.05148, 2016.
[bib28] Sheena A. Josselyn and Susumu Tonegawa. Memory engrams: Recalling the past and imagining the future. Science, 367, 2020.
[bib29] Georg B Keller and Thomas D Mrsic-Flogel. Predictive processing: A canonical cortical computation. Neuron, 100(2):424–435, October 2018.
[bib30] Kim et al. (2021) Saehoon Kim, Sungwoong Kim, and Juho Lee. Hybrid generative-contrastive representation learning. arXiv preprint arXiv:2106.06162, 2021.
[bib31] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
[bib32] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
[bib33] Krizhevsky et al. (2014) Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. The CIFAR-10 dataset. online: http://www.cs.toronto.edu/kriz/cifar.html, 55, 2014.
[bib34] Krizhevsky et al. (2009) Alex Krizhevsky et al. Learning multiple layers of features from tiny images. arXiv preprint arXiv:1312.6114, 2009.
[bib35] Li et al. (2022) Zengyi Li, Yubei Chen, Yann LeCun, and Friedrich T Sommer. Neural manifold clustering and embedding. arXiv preprint arXiv:2201.10000, 2022.
[bib36] Liu et al. (2020) Steven Liu, Tongzhou Wang, David Bau, Jun-Yan Zhu, and Antonio Torralba. Diverse image generation via self-conditioned gans. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14286–14295, 2020.
[bib37] Ma et al. (2007) Yi Ma, Harm Derksen, Wei Hong, and John Wright. Segmentation of multivariate mixed data via lossy data coding and compression. PAMI, 2007.
[bib38] Mukherjee et al. (2019) Sudipto Mukherjee, Himanshu Asnani, Eugene Lin, and Sreeram Kannan. Clustergan: Latent space clustering in generative adversarial networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 4610–4617, 2019.
[bib39] Parmar et al. (2021) Gaurav Parmar, Dacheng Li, Kwonjoon Lee, and Zhuowen Tu. Dual contradistinctive generative autoencoder. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 823–832, 2021.
[bib40] Prasad et al. (2020) Vignesh Prasad, Dipanjan Das, and Brojeshwar Bhowmick. Variational clustering: Leveraging variational autoencoders for image clustering. In 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–10. IEEE, 2020.
[bib41] Radford et al. (2015) Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
[bib42] Salimans et al. (2016) Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved techniques for training gans. Advances in neural information processing systems, 29, 2016.
[bib43] Tong et al. (2022) Shengbang Tong, Xili Dai, Ziyang Wu, Mingyang Li, Brent Yi, and Yi Ma. Incremental learning of structured memory via closed-loop transcription. arXiv:2202.05411, 2022.
[bib44] Van der Maaten & Hinton (2008) Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
[bib45] Ven et al. (2020) Gido M Ven, Hava T Siegelmann, Andreas S Tolias, et al. Brain-inspired replay for continual learning with artificial neural networks. Nature Communications, 11(1):1–14, 2020.
[bib46] Wang et al. (2004) Zhou Wang, A.C. Bovik, H.R. Sheikh, and E.P. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600–612, 2004. doi: 10.1109/TIP.2003.819861.
[bib47] Yu et al. (2020) Yaodong Yu, Kwan Ho Ryan Chan, Chong You, Chaobing Song, and Yi Ma. Learning diverse and discriminative representations via the principle of maximal coding rate reduction. Advances in Neural Information Processing Systems, 33:9422–9434, 2020.
[bib48] Zbontar et al. (2021) Jure Zbontar, Li Jing, Ishan Misra, Yann LeCun, and Stéphane Deny. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pp. 12310–12320. PMLR, 2021.