| Invariant Information Clustering for Unsupervised Image Classification and Segmentation (Jul 2018, arXiv 2018) |
|
88.8% |
| Xu Ji, João F. Henriques, Andrea Vedaldi We present a novel clustering objective that learns a neural network classifier from scratch, given only unlabelled data samples. The model discovers clusters that accurately match semantic classes, achieving state-of-the-art results in eight unsupervised clustering benchmarks spanning image classification and segmentation. These include STL10, an unsupervised variant of ImageNet, and CIFAR10, where we significantly beat the accuracy of our closest competitors by 6.6 and 9.5 absolute percentage points respectively. The method is not specialised to computer vision and operates on any paired dataset samples; in our experiments we use random transforms to obtain a pair from each image. The trained network directly outputs semantic labels, rather than high dimensional representations that need external processing to be usable for semantic clustering. The objective is simply to maximise mutual information between the class assignments of each pair. It is easy to implement and rigorously grounded in information theory, meaning we effortlessly avoid degenerate solutions that other clustering methods are susceptible to. In addition to the fully unsupervised mode, we also test two semi-supervised settings. The first achieves 88.8% accuracy on STL10 classification, setting a new global state-of-the-art over all existing methods (whether supervised, semi-supervised or unsupervised). The second shows robustness to 90% reductions in label coverage, of relevance to applications that wish to make use of small amounts of labels. github.com/xu-ji/IIC |
|
|
| Scaling the Scattering Transform: Deep Hybrid Networks (Mar 2017, arXiv 2017) |
76.0% (±0.6%) |
87.6% |
| Edouard Oyallon, Eugene Belilovsky, Sergey Zagoruyko We use the scattering network as a generic and fixed ini-tialization of the first layers of a supervised hybrid deep network. We show that early layers do not necessarily need to be learned, providing the best results to-date with pre-defined representations while being competitive with Deep CNNs. Using a shallow cascade of 1 x 1 convolutions, which encodes scattering coefficients that correspond to spatial windows of very small sizes, permits to obtain AlexNet accuracy on the imagenet ILSVRC2012. We demonstrate that this local encoding explicitly learns invariance w.r.t. rotations. Combining scattering networks with a modern ResNet, we achieve a single-crop top 5 error of 11.4% on imagenet ILSVRC2012, comparable to the Resnet-18 architecture, while utilizing only 10 layers. We also find that hybrid architectures can yield excellent performance in the small sample regime, exceeding their end-to-end counterparts, through their ability to incorporate geometrical priors. We demonstrate this on subsets of the CIFAR-10 dataset and on the STL-10 dataset. |
|
|
| Improved Regularization of Convolutional Neural Networks with Cutout (Aug 2017, arXiv 2017) |
|
87.26% (±0.23%) |
| Terrance DeVries, Graham W. Taylor Convolutional neural networks are capable of learning powerful representational spaces, which are necessary for tackling complex learning tasks. However, due to the model capacity required to capture such representations, they are often susceptible to overfitting and therefore require proper regularization in order to generalize well. In this paper, we show that the simple regularization technique of randomly masking out square regions of input during training, which we call cutout, can be used to improve the robustness and overall performance of convolutional neural networks. Not only is this method extremely easy to implement, but we also demonstrate that it can be used in conjunction with existing forms of data augmentation and other regularizers to further improve model performance. We evaluate this method by applying it to current state-of-the-art architectures on the CIFAR-10, CIFAR-100, and SVHN datasets, yielding new state-of-the-art results of 2.56%, 15.20%, and 1.30% test error respectively. Code is available at https://github.com/uoguelph-mlrg/Cutout |
|
|
| Deep Unsupervised Learning Through Spatial Contrasting (Oct 2016, arXiv 2016) |
|
81.3% |
| Elad Hoffer, Itay Hubara, Nir Ailon Convolutional networks have marked their place over the last few years as the best performing model for various visual tasks. They are, however, most suited for supervised learning from large amounts of labeled data. Previous attempts have been made to use unlabeled data to improve model performance by applying unsupervised techniques. These attempts require different architectures and training methods. In this work we present a novel approach for unsupervised training of Convolutional networks that is based on contrasting between spatial regions within images. This criterion can be employed within conventional neural networks and trained using standard techniques such as SGD and back-propagation, thus complementing supervised methods. |
|
|
| Stacked What-Where Auto-encoders (Jun 2015) |
74.33% |
|
| Junbo Zhao, Michael Mathieu, Ross Goroshin, Yann LeCun We present a novel architecture, the "stacked what-where auto-encoders" (SWWAE), which integrates discriminative and generative pathways and provides a unified approach to supervised, semi-supervised and unsupervised learning without relying on sampling during training. An instantiation of SWWAE uses a convolutional net (Convnet) (LeCun et al. (1998)) to encode the input, and employs a deconvolutional net (Deconvnet) (Zeiler et al. (2010)) to produce the reconstruction. The objective function includes reconstruction terms that induce the hidden states in the Deconvnet to be similar to those of the Convnet. Each pooling layer produces two sets of variables: the "what" which are fed to the next layer, and its complementary variable "where" that are fed to the corresponding layer in the generative decoder. |
|
|
| Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks (Jun 2014, arXiv 2015) |
74.2% (±0.4%) |
|
| Alexey Dosovitskiy, Philipp Fischer, Jost Tobias Springenberg, Martin Riedmiller, Thomas Brox Deep convolutional networks have proven to be very successful in learning task specific features that allow for unprecedented performance on various computer vision tasks. Training of such networks follows mostly the supervised learning paradigm, where sufficiently many input-output pairs are required for training. Acquisition of large training sets is one of the key challenges, when approaching a new task. In this paper, we aim for generic feature learning and present an approach for training a convolutional network using only unlabeled data. To this end, we train the network to discriminate between a set of surrogate classes. Each surrogate class is formed by applying a variety of transformations to a randomly sampled 'seed' image patch. In contrast to supervised network training, the resulting feature representation is not class specific. It rather provides robustness to the transformations that have been applied during training. This generic feature representation allows for classification results that outperform the state of the art for unsupervised learning on several popular datasets (STL-10, CIFAR-10, Caltech-101, Caltech-256). While such generic features cannot compete with class specific features from supervised training on a classification task, we show that they are advantageous on geometric matching problems, where they also outperform the SIFT descriptor. |
|
|
| Convolutional Clustering for Unsupervised Learning (Nov 2015) |
74.10% |
|
| Aysegul Dundar, Jonghoon Jin, Eugenio Culurciello The task of labeling data for training deep neural networks is daunting and tedious, requiring millions of labels to achieve the current state-of-the-art results. Such reliance on large amounts of labeled data can be relaxed by exploiting hierarchical features via unsupervised learning techniques. In this work, we propose to train a deep convolutional network based on an enhanced version of the k-means clustering algorithm, which reduces the number of correlated parameters in the form of similar filters, and thus increases test categorization accuracy. We call our algorithm convolutional k-means clustering. We further show that learning the connection between the layers of a deep convolutional neural network improves its ability to be trained on a smaller amount of labeled data. Our experiments show that the proposed algorithm outperforms other techniques that learn filters unsupervised. Specifically, we obtained a test accuracy of 74.1% on STL-10 and a test error of 0.5% on MNIST. |
|
|
| Deep Representation Learning with Target Coding (AAAI 2015) |
73.15% |
|
| Shuo Yang, Ping Luo, Chen Change Loy, Kenneth W. Shum, Xiaoou Tang
We consider the problem of learning deep representation when target labels are available. In this paper,
we show that there exists intrinsic relationship between target coding and feature representation learning in deep
networks. Specifically, we found that distributed binary code with error correcting capability is more capable of
encouraging discriminative features, in comparison to the 1-of-K coding that is typically used in supervised
deep learning. This new finding reveals additional benefit of using error-correcting code for deep model learning,
apart from its well-known error correcting property. Extensive experiments are conducted on popular visual
benchmark datasets. |
|
|
| Discriminative Unsupervised Feature Learning with Convolutional Neural Networks (NIPS 2014) |
72.8% (±0.4%) |
|
| Alexey Dosovitskiy, Jost Tobias Springenberg, Martin Riedmiller, Thomas Brox
Current methods for training convolutional neural networks depend on large
amounts of labeled samples for supervised training. In this paper we present an
approach for training a convolutional neural network using only unlabeled data.
We train the network to discriminate between a set of surrogate classes. Each
surrogate class is formed by applying a variety of transformations to a randomly
sampled ’seed’ image patch. We find that this simple feature learning algorithm
is surprisingly successful when applied to visual object recognition. The feature
representation learned by our algorithm achieves classification results matching
or outperforming the current state-of-the-art for unsupervised learning on several
popular datasets (STL-10, CIFAR-10, Caltech-101). |
|
|
| An Analysis of Unsupervised Pre-training in Light of Recent Advances (Dec 2014, ICLR 2015) |
70.20% (±0.7%) |
|
| Tom Le Paine, Pooya Khorrami, Wei Han, Thomas S. Huang Convolutional neural networks perform well on object recognition because of a number of recent advances: rectified linear units (ReLUs), data augmentation, dropout, and large labelled datasets. Unsupervised data has been proposed as another way to improve performance. Unfortunately, unsupervised pre-training is not used by state-of-the-art methods leading to the following question: Is unsupervised pre-training still useful given recent advances? If so, when? We answer this in three parts: we 1) develop an unsupervised method that incorporates ReLUs and recent unsupervised regularization techniques, 2) analyze the benefits of unsupervised pre-training compared to data augmentation and dropout on CIFAR-10 while varying the ratio of unsupervised to supervised samples, 3) verify our findings on STL-10. We discover unsupervised pre-training, as expected, helps when the ratio of unsupervised to supervised samples is high, and surprisingly, hurts when the ratio is low. We also use unsupervised pre-training with additional color augmentation to achieve near state-of-the-art performance on STL-10. |
|
|
| Multi-Task Bayesian Optimization (NIPS 2013) |
70.1% (±0.6%) |
|
| Kevin Swersky, Jasper Snoek
Bayesian optimization has recently been proposed as a framework for automatically
tuning the hyperparameters of machine learning models and has been shown
to yield state-of-the-art performance with impressive ease and efficiency. In this
paper, we explore whether it is possible to transfer the knowledge gained from
previous optimizations to new tasks in order to find optimal hyperparameter settings
more efficiently. Our approach is based on extending multi-task Gaussian
processes to the framework of Bayesian optimization. We show that this method
significantly speeds up the optimization process when compared to the standard
single-task approach. We further propose a straightforward extension of our algorithm
in order to jointly minimize the average error across multiple tasks and
demonstrate how this can be used to greatly speed up k-fold cross-validation.
Lastly, we propose an adaptation of a recently developed acquisition function, entropy
search, to the cost-sensitive, multi-task setting. We demonstrate the utility
of this new acquisition function by leveraging a small dataset to explore hyperparameter
settings for a large dataset. Our algorithm dynamically chooses which
dataset to query in order to yield the most information per unit cost. |
|
|
| C-SVDDNet: An Effective Single-Layer Network for Unsupervised Feature Learning (Dec 2014) |
68.23% (±0.5%) |
|
| Dong Wang, Xiaoyang Tan In this paper, we investigate the problem of learning feature representation from unlabeled data using a single-layer K-means network. A K-means network maps the input data into a feature representation by finding the nearest centroid for each input point, which has attracted researchers' great attention recently due to its simplicity, effectiveness, and scalability. However, one drawback of this feature mapping is that it tends to be unreliable when the training data contains noise. To address this issue, we propose a SVDD based feature learning algorithm that describes the density and distribution of each cluster from K-means with an SVDD ball for more robust feature representation. For this purpose, we present a new SVDD algorithm called C-SVDD that centers the SVDD ball towards the mode of local density of each cluster, and we show that the objective of C-SVDD can be solved very efficiently as a linear programming problem. Additionally, traditional unsupervised feature learning methods usually take an average or sum of local representations to obtain global representation which ignore spatial relationship among them. To use spatial information we propose a global representation with a variant of SIFT descriptor. The architecture is also extended with multiple receptive field scales and multiple pooling sizes. Extensive experiments on several popular object recognition benchmarks, such as STL-10, MINST, Holiday and Copydays shows that the proposed C-SVDDNet method yields comparable or better performance than that of the previous state of the art methods. |
|
|
| Committees of deep feedforward networks trained with few data (Jun 2014) |
68% (±0.55%) |
|
| Bogdan Miclut, Thomas Kaester, Thomas Martinetz, Erhardt Barth Deep convolutional neural networks are known to give good results on image classification tasks. In this paper we present a method to improve the classification result by combining multiple such networks in a committee. We adopt the STL-10 dataset which has very few training examples and show that our method can achieve results that are better than the state of the art. The networks are trained layer-wise and no backpropagation is used. We also explore the effects of dataset augmentation by mirroring, rotation, and scaling. |
|
|
| Stable and Efficient Representation Learning with Nonnegativity Constraints (ICML 2014) |
67.9% (±0.6%) |
|
| Tsung-Han Lin, H. T. Kung
Orthogonal matching pursuit (OMP) is an efficient approximation algorithm for computing
sparse representations. However, prior research has shown that the representations computed by
OMP may be of inferior quality, as they deliver suboptimal classification accuracy on several image
datasets. We have found that this problem is caused by OMP’s relatively weak stability under
data variations, which leads to unreliability in supervised classifier training. We show that
by imposing a simple nonnegativity constraint, this nonnegative variant of OMP (NOMP) can
mitigate OMP’s stability issue and is resistant to noise overfitting. In this work, we provide extensive
analysis and experimental results to examine and validate the stability advantage of NOMP.
In our experiments, we use a multi-layer deep architecture for representation learning, where
we use K-means for feature learning and NOMP for representation encoding. The resulting learning
framework is not only efficient and scalable to large feature dictionaries, but also is robust
against input noise. This framework achieves the state-of-the-art accuracy on the STL-10 dataset. |
|
|
| Unsupervised Feature Learning for RGB-D Based Object Recognition (ISER 2012) |
64.5% (±1%) |
|
| Liefeng Bo, Xiaofeng Ren, Dieter Fox
Recently introduced RGB-D cameras are capable of providing high
quality synchronized videos of both color and depth. With its advanced sensing
capabilities, this technology represents an opportunity to dramatically increase
the capabilities of object recognition. It also raises the problem of developing
expressive features for the color and depth channels of these sensors. In this paper
we introduce hierarchical matching pursuit (HMP) for RGB-D data. HMP uses
sparse coding to learn hierarchical feature representations from raw RGB-D data
in an unsupervised way. Extensive experiments on various datasets indicate that
the features learned with our approach enable superior object recognition results
using linear support vector machines. |
|
|
| Convolutional Kernel Networks (Jun 2014) |
62.32% |
|
| Julien Mairal, Piotr Koniusz, Zaid Harchaoui, Cordelia Schmid An important goal in visual recognition is to devise image representations that are invariant to particular transformations. In this paper, we address this goal with a new type of convolutional neural network (CNN) whose invariance is encoded by a reproducing kernel. Unlike traditional approaches where neural networks are learned either to represent data or for solving a classification task, our network learns to approximate the kernel feature map on training data. Such an approach enjoys several benefits over classical ones. First, by teaching CNNs to be invariant, we obtain simple network architectures that achieve a similar accuracy to more complex ones, while being easy to train and robust to overfitting. Second, we bridge a gap between the neural network literature and kernels, which are natural tools to model invariance. We evaluate our methodology on visual recognition tasks where CNNs have proven to perform well, e.g., digit recognition with the MNIST dataset, and the more challenging CIFAR-10 and STL-10 datasets, where our accuracy is competitive with the state of the art. |
|
|
| Discriminative Learning of Sum-Product Networks (NIPS 2012) |
62.3% (±1%) |
|
| Robert Gens, Pedro Domingos
Sum-product networks are a new deep architecture that can perform fast, exact inference
on high-treewidth models. Only generative methods for training SPNs
have been proposed to date. In this paper, we present the first discriminative
training algorithms for SPNs, combining the high accuracy of the former with
the representational power and tractability of the latter. We show that the class
of tractable discriminative SPNs is broader than the class of tractable generative
ones, and propose an efficient backpropagation-style algorithm for computing the
gradient of the conditional log likelihood. Standard gradient descent suffers from
the diffusion problem, but networks with many layers can be learned reliably using
“hard” gradient descent, where marginal inference is replaced by MPE inference
(i.e., inferring the most probable state of the non-evidence variables). The
resulting updates have a simple and intuitive form. We test discriminative SPNs
on standard image classification tasks. We obtain the best results to date on the
CIFAR-10 dataset, using fewer features than prior methods with an SPN architecture
that learns local image structure discriminatively. We also report the highest
published test accuracy on STL-10 even though we only use the labeled portion
of the dataset. |
|
|
| No more meta-parameter tuning in unsupervised sparse feature learning (Feb 2014) |
61.0% (±0.58%) |
|
| Adriana Romero, Petia Radeva, Carlo Gatta We propose a meta-parameter free, off-the-shelf, simple and fast unsupervised feature learning algorithm, which exploits a new way of optimizing for sparsity. Experiments on STL-10 show that the method presents state-of-the-art performance and provides discriminative features that generalize well. |
|
|
| Deep Learning of Invariant Features via Simulated Fixations in Video (NIPS 2012 2012) |
61% |
|
| Will Y. Zou, Shenghuo Zhu, Andrew Y. Ng, Kai Yu
We apply salient feature detection and tracking in videos to simulate fixations and
smooth pursuit in human vision. With tracked sequences as input, a hierarchical
network of modules learns invariant features using a temporal slowness constraint.
The network encodes invariance which are increasingly complex with hierarchy.
Although learned from videos, our features are spatial instead of spatial-temporal,
and well suited for extracting features from still images. We applied our features to
four datasets (COIL-100, Caltech 101, STL-10, PubFig), and observe a consistent
improvement of 4% to 5% in classification accuracy. With this approach, we
achieve state-of-the-art recognition accuracy 61% on STL-10 dataset. |
|
|
| Selecting Receptive Fields in Deep Networks (NIPS 2011) |
60.1% (±1%) |
|
| Adam Coates, Andrew Y. Ng
Recent deep learning and unsupervised feature learning systems that learn from
unlabeled data have achieved high performance in benchmarks by using extremely
large architectures with many features (hidden units) at each layer. Unfortunately,
for such large architectures the number of parameters can grow quadratically in the
width of the network, thus necessitating hand-coded “local receptive fields” that
limit the number of connections from lower level features to higher ones (e.g.,
based on spatial locality). In this paper we propose a fast method to choose these
connections that may be incorporated into a wide variety of unsupervised training
methods. Specifically, we choose local receptive fields that group together those
low-level features that are most similar to each other according to a pairwise similarity
metric. This approach allows us to harness the advantages of local receptive
fields (such as improved scalability, and reduced data requirements) when we do
not know how to specify such receptive fields by hand or where our unsupervised
training algorithm has no obvious generalization to a topographic setting. We
produce results showing how this method allows us to use even simple unsupervised
training algorithms to train successful multi-layered networks that achieve
state-of-the-art results on CIFAR and STL datasets: 82.0% and 60.1% accuracy,
respectively. |
|
|
| Learning Invariant Representations with Local Transformations (Jun 2012, ICML 2012) |
58.7% |
|
| Kihyuk Sohn, Honglak Lee Learning invariant representations is an important problem in machine learning and pattern recognition. In this paper, we present a novel framework of transformation-invariant feature learning by incorporating linear transformations into the feature learning algorithms. For example, we present the transformation-invariant restricted Boltzmann machine that compactly represents data by its weights and their transformations, which achieves invariance of the feature representation via probabilistic max pooling. In addition, we show that our transformation-invariant feature learning framework can also be extended to other unsupervised learning methods, such as autoencoders or sparse coding. We evaluate our method on several image classification benchmark datasets, such as MNIST variations, CIFAR-10, and STL-10, and show competitive or superior classification performance when compared to the state-of-the-art. Furthermore, our method achieves state-of-the-art performance on phone classification tasks with the TIMIT dataset, which demonstrates wide applicability of our proposed algorithms to other domains. |
|
|
| Pooling-Invariant Image Feature Learning (Jan 2013) |
58.28% |
|
| Yangqing Jia, Oriol Vinyals, Trevor Darrell Unsupervised dictionary learning has been a key component in state-of-the-art computer vision recognition architectures. While highly effective methods exist for patch-based dictionary learning, these methods may learn redundant features after the pooling stage in a given early vision architecture. In this paper, we offer a novel dictionary learning scheme to efficiently take into account the invariance of learned features after the spatial pooling stage. The algorithm is built on simple clustering, and thus enjoys efficiency and scalability. We discuss the underlying mechanism that justifies the use of clustering algorithms, and empirically show that the algorithm finds better dictionaries than patch-based methods with the same dictionary size. |
|
|