|
WePT1 |
Poster Session Hall |
WeP1 |
Poster Session |
|
15:00-17:10, Paper WePT1.1 | |
Selective Unsupervised Feature Learning with Convolutional Neural Network (S-CNN) |
Ghaderi, Aamir | Univ. of Texas at Arlington |
Athitsos, Vassilis | Univ. of Texas at Arlington |
Keywords: Deep learning, Classification and clustering, Artificial neural networks
Abstract: Supervised learning needs vitally huge amounts of labeled data. Labeling thousands or millions of data is very boring and needs much more time and cost. On the other hand one of the important goals in visual recognition is to create features from unlabeled data. In this paper we propose new method for training a CNN with no need to labeled instances. This algorithm for unsupervised feature learning is successful when we test for object recognition. We implement simple algorithm which can get accuracy similar to more sophisticated ones. This is easy for training and resistant to overfitting. We show the results on the popular data sets that are STL-10, CIFAR-10, and CIFAR-100 where our accuracy is competitive with the other methods. Selective Convolutional Neural Network (S-CNN) is simple and fast algorithm, that introduces new way of unsupervised feature learning and provide discriminative features which generalize well.
|
|
15:00-17:10, Paper WePT1.2 | |
Detection of Intracranial Hypertension Using Deep Learning |
Quachtran, Benjamin | Univ. of California, Los Angeles |
Hamilton, Robert | Neural Analytics |
Scalzo, Fabien | UCLA, Geffen School of Medicine |
Keywords: Deep learning, Medical image and signal analysis, Classification and clustering
Abstract: Intracranial Hypertension, a disorder characterized by elevated pressure in the brain, is typically monitored in neurointensive care and diagnosed only after elevation has occurred. This reaction-based method of treatment leaves patients at higher risk of additional complications in case of misdetec- tion. The detection of intracranial hypertension has been the subject of many recent studies in an attempt to accurately characterize the causes of hypertension, specifically examining waveform morphology. We investigate the use of Deep Learning, a hierarchical form of machine learning, to model the relationship between hypertension and waveform morphology, giving us the ability to accurately detect presence hypertension. Data from 60 patients, showing intracranial pressure levels over a half hour time span, was used to evaluate the model. We divided each patient’s recording into average normalized beats over 30 sec segments, assigning each beat a label of high (i.e. greater than 15 mmHg) or low intracranial pressure. The model was tested to predict the presence of elevated intracranial pressure. The algorithm was found to be 92.05± 2.25% accurate in detecting intracranial hypertension.
|
|
15:00-17:10, Paper WePT1.3 | |
Underwater Target Classification in Synthetic Aperture Sonar Imagery Using Deep Convolutional Neural Networks |
Williams, David | NATO Science and Tech. Organization |
Keywords: Deep learning, Pattern Recognition for Surveillance and Security, Artificial neural networks
Abstract: Deep convolutional neural networks are used to perform underwater target classification in synthetic aperture sonar (SAS) imagery. The deep networks are learned using a massive database of real, measured sonar data collected at sea during different expeditions in various geographical locations. A novel training procedure is developed specially for the data from this new sensor modality in order to augment the amount of training data available for learning and to avoid overfitting. The deep networks learned are employed for several binary classification tasks in which different classes of objects in real sonar data are to be discriminated. The proposed deep approach consistently achieves superior performance to a traditional feature-based classifier that we had relied on previously.
|
|
15:00-17:10, Paper WePT1.4 | |
Bad Teacher or Unruly Student: Can Deep Learning Say Something in Image Forensics Analysis? |
Rota, Paolo | Vienna Univ. of Tech |
Sangineto, Enver | Univ. of Trento |
Conotter, Valentina | Social IT S.r.l |
Pramerdorfer, Christopher | Vienna Univ. of Tech |
Keywords: Deep learning, Security issues, Other applications
Abstract: The pervasive availability of the Internet, coupled with the development of increasingly powerful technologies, has led digital images to be the primary source of visual information in nowadays society. However, their reliability as a true representation of reality cannot be taken for granted, due to the affordable powerful graphics editing softwares that can easily alter the original content, leaving no visual trace of any modification on the image making them potentially dangerous. This motivates developing technological solutions able to detect media manipulations without a prior knowledge or extra information regarding the given image. At the same time, the huge amount of available data has also led to tremendous advances of data-hungry learning models, which have already demonstrated in last few years to be successful in image classification. In this work we propose a deep learning approach for tampered image classification. To our best knowledge, this the first attempt to use the deep learning paradigm in an image forensic scenario. In particular, we propose a new blind deep learning approach based on Convolutional Neural Networks (CNN) able to learn invisible discriminative artifacts from manipulated images that can be exploited to automatically discriminate between forged and authentic images. The proposed approach not only detects forged images but it can be extended to localize the tampered regions within the image. This method outperforms the state-of-the-art in terms of accuracy on CASIA TIDE v2.0 dataset. The capability of automatically crafting discriminant features can lead to surprising results. For instance, detecting image compression filters used to create the dataset. This argument is also discussed within this paper.
|
|
15:00-17:10, Paper WePT1.5 | |
Semi-Supervised Tuning from Temporal Coherence |
Maltoni, Davide | Univ. of Bologna |
Lomonaco, Vincenzo | Univ. of Bologna |
Keywords: Deep learning, Semi-supervised learning and spectral methods, Artificial neural networks
Abstract: Recent works demonstrated the usefulness of temporal coherence to regularize supervised training or to learn invariant features with deep architectures. In particular, enforcing a smooth output change while presenting temporally-closed frames from video sequences, proved to be an effective strategy. In this paper we prove the efficacy of temporal coherence for semi-supervised incremental tuning. We show that a deep architecture, just mildly trained in a supervised manner, can progressively improve its classification accuracy, if exposed to video sequences of unlabeled data. The extent to which, in some cases, a semi-supervised tuning allows to improve classification accuracy (approaching the supervised one) is somewhat surprising. A number of control experiments pointed out the fundamental role of temporal coherence.
|
|
15:00-17:10, Paper WePT1.6 | |
Non-Negative Multiple Matrix Factorization with Euclidean and Kullback-Leibler Mixed Divergences |
Kohjima, Masahiro | NTT Corp |
Matsubayashi, Tatsushi | NTT Corp |
Sawada, Hiroshi | NTT |
Keywords: Machine learning and data mining
Abstract: In this paper, we tackle the problem of extracting latent structure and patterns from multiple datasets that consist of users' rating scores and activity logs (click, view, visit, ...) in order to understand the typical users' behavior. Our proposed method is based on non-negative matrix factorization, and factorizes multiple matrices simultaneously while adopting Euclidean distance and generalized KL divergence for the rating matrix and the activity matrix, respectively. We derive an optimization algorithm that offers a theoretical guarantee that it can find a locally optimal solution. Our experiments show that the proposed method outperformed existing methods when measured by mean squared error, which implies that it can extract latent structure and patterns more precisely. We also confirm that the segmentation result by the proposal helps to analyze users' behavior.
|
|
15:00-17:10, Paper WePT1.7 | |
On Combining Websensors and DTW Distance for Knn Time Series Forecasting |
Marcacini, Ricardo Marcondes | Federal Univ. of Mato Grosso Do Sul |
Carnevali, Julio César | Univ. Federal De Mato Grosso Do Sul |
Domingos, João Domingos Ferreira Mundim | UFMS |
Keywords: Machine learning and data mining
Abstract: In the pattern recognition field, different approaches have been proposed to improve time series forecasting models. In this sense, k-Nearest-Neighbour (kNN) with DTW (Dynamic Time Warping) distance is one of the most representative methods, due to its effectiveness, simplicity and intuitiveness. The great advantage of the DTW distance is the robustness to distortions in the time axis by allowing stretching and squeezing (time warping) of the time series, while traditional measures require a linear alignment between each data point. However, as well as other traditional measures, the DTW distance has the limitation of focusing only on historical time series data to predict future values, thereby not considering additional external knowledge of the problem domain. In this paper, we propose an approach called TSFW (Time Series Forecasting with Websensors) that incorporates Websensors into DTW distance to improve kNN time series forecasting. Websensors are models that represent knowledge extracted from news about the problem domain as well as the temporal evolution of this knowledge. In our proposed TSFW approach, we show that Websensors allow a more robust non-linear alignment of the time series by using similar events (extracted from news) that have occurred in the both time series. Thus, distortions in the time axis among the time series can be corrected more accurately compared to the traditional technique that uses only the original values of the time series.
|
|
15:00-17:10, Paper WePT1.8 | |
Exploiting Social and Mobility Patterns for Friendship Prediction in Location-Based Social Networks |
Valverde-Rebaza, Jorge Carlos | Univ. of São Paulo |
Roche, Mathieu | Cirad, TETIS & LIRMM |
Poncelet, Pascal | LIRMM |
de Andrade Lopes, Alneu | Univ. of São Paulo |
Keywords: Machine learning and data mining, Classification and clustering
Abstract: Link prediction is a ``hot topic'' in network analysis and has been largely used for friendship recommendation in social networks. With the increased use of location-based services, it is possible to improve the accuracy of link prediction methods by using the mobility of users. The majority of the link prediction methods focus on the importance of location for their visitors, disregarding the strength of relationships existing between these visitors. We, therefore, propose three new methods for friendship prediction by combining, efficiently, social and mobility patterns of users in location-based social networks (LBSNs). Experiments conducted on real-world datasets demonstrate that our proposals achieve a competitive performance with methods from the literature and, in most of the cases, outperform them. Moreover, our proposals use less computational resources by reducing considerably the number of irrelevant predictions, making the link prediction task more efficient and applicable for real world applications.
|
|
15:00-17:10, Paper WePT1.9 | |
Detecting Contextual Collective Anomalies at a Glance |
Prado-Romero, Mario Alfonso | CENATAV |
Gago Alonso, Andrés | Advanced Tech. Application Centre (CENATAV) |
Keywords: Machine learning and data mining, Classification and clustering
Abstract: Many phenomena in our world can be modeled as networks, from neurons in the human brain, computer networks and bank transactions to social interactions. Anomaly detection is an important data mining task consisting in detecting rare objects that deviate from the majority of the data. Contextual collective anomaly detection techniques can be applied to intrusion detection in computer networks, bank fraud detection, or finding people with strange behavior in social networks. In this work, a fast and intuitive algorithm to detect collective contextual anomalies is presented. Furthermore, the importance of selecting algorithms which find meaningful outliers for the application domain specialists is analyzed.
|
|
15:00-17:10, Paper WePT1.10 | |
Dual Approximated Nuclear Norm Based Matrix Regression Via Adaptive Line Search Scheme |
Luo, Lei | Nanjing Univ. of Science and Tech |
Yang, Jian | Nanjing Univ. of Science and Tech |
Tu, Qinghua | School of Computer Science and Engineering, Nanjing Univ. Of |
Zhang, Yigong | Nanjing Univ. of Science and Tech |
Keywords: Machine learning and data mining, Classification and clustering
Abstract: Face recognition with partial occlusion is one of the urgent and challenging problems in the pattern recognition research. Using the Alternating Direction Method of Multipliers (ADMM), the recently proposed nuclear norm based matrix regression model (NMR) has been shown a great potential in dealing with the structural noise. And yet, ADMM needs to bring into an auxiliary variable and only exploits the convexity of NMR. Compared with ADMM, the gradient based methods are simpler. To make use of these methods, this paper considers the Approximated NMR (ANMR) model. Utilizing the singular value shrinkage operator and strong convexity of ANMR, the dual problem of ANMR (DANMR) is derived and a crucial result is obtained: the primal optimal solution of ANMR can be converted as the matrix function associated with the dual optimal solution. Due to the differentiability of DANMR, an adaptive line search scheme is developed to solve it. This approach combines the advantages of the accelerated gradient technique and adaptive parameters updating strategy. Therefore, a convergence rate of O(1/N2) can be guaranteed. Experimental results show the superiority of the proposed algorithm over some existing methods.
|
|
15:00-17:10, Paper WePT1.11 | |
Aggregation Procedure of Gaussian Mixture Models for Additive Features |
Ridi, Antonio | Univ. of Applied Sciences Western Switzerland |
Gisler, Christophe | Univ. of Fribourg, Switzerland |
Hennebert, Jean | Univ. of Applied Sciences Western Switzerland |
Keywords: Machine learning and data mining, Classification and clustering
Abstract: In this work we provide details on a new and effective approach able to generate Gaussian Mixture Models (GMMs) for the classification of aggregated time series. More specifically, our procedure can be applied to time series that are aggregated together by adding their features. The procedure takes advantage of the additive property of the Gaussians that complies with the additive property of the features. Our goal is to classify aggregated time series, i.e. we aim to identify the classes of the single time series contributing to the total. The standard approach consists in training the models using the combination of several time series coming from different classes. However, this has the drawback of being a very slow operation given the amount of data. The proposed approach, called GMMs aggregation procedure, addresses this problem. It consists of three steps: (i) modeling the independent classes, (ii) generation of the models for the class combinations and (iii) simplification of the generated models. We show the effectiveness of our approach by using time series in the context of electrical appliance consumption, where the time series are aggregated by adding the active and reactive power. Finally, we compare the proposed approach with the standard procedure.
|
|
15:00-17:10, Paper WePT1.12 | |
Learning Multi-View Strategies with Boosting for Classification |
Peng, Jing | Montclair State Univ. |
Aved, Alex | AFRL |
|
15:00-17:10, Paper WePT1.13 | |
Multiview Clustering Based on Robust and Regularized Matrix Approximation |
Pu, Jiameng | Computer School, Wuhan Univ |
Zhang, Qian | Beijing Samsung Telecom R&D Center |
Zhang, Lefei | Department of Computing, the Hong Kong Pol. Univ |
Du, Bo | School of Computer, Wuhan Univ. Wuhan 430079, China |
You, Jane | The Hong Kong Pol. Univ |
Keywords: Machine learning and data mining, Classification and clustering, Dimensionality reduction and manifold learning
Abstract: Pattern recognition tasks such as the data classification and clustering usually can be represented by the perspective of multiple views or feature spaces. Obviously, the performance of the classification and clustering should be greatly improved if we carefully consider the discriminabilities from multiple views and explore the complementary information among them. However, multiple features also bring new challenges to handle them. In the literature, many existed multiview feature learning methods dealt with different views equally, thus they couldn't optimally utilize the complementary property of them. On the other hand, matrix factorization based clustering algorithms usually adopt the conventional ell_{2}-norm based squared residue minimization to measure the loss, which is easily influenced by the outliers and noises from the multiple sources of input. In this paper, we propose a novel multiview data clustering algorithm based on the matrix factorization to relieve the above issues. The basic idea for the proposed Robust and Regularized Matrix Approximation (RRMA) is that the observed data matrix could be low-rank approximated by a cluster centroid matrix and a cluster indicator matrix, respectively, and the major contributions of our work lie in the introduction of the robust ell_{2,1}-norm and ensemble manifold regularization to regularize the matrix factorization and make the model more discriminative for multiview data clustering. We properly adjust the importance of different views by assigning a set of trainable weights on the views. Moreover, we propose an efficient solution featured with impactful updating rules to seek the local optimal parameters. Encouraging experimental results on numerous public multiview datasets demonstrate the superiority of our model compared to some state-of-the-art methods.
|
|
15:00-17:10, Paper WePT1.14 | |
Detecting Low-Quality Reference Time Series in Stream Recognition |
Dupont, Marc | IRISA |
Marteau, Pierre-Francois | Univ. De Bretagne Sud |
Ghouaiel, Nehla | IRISA Labaratory, Univ. of South Brittany |
Keywords: Machine learning and data mining, Classification and clustering, Gesture and Behavior Analysis
Abstract: On-line supervised spotting and classification of subsequences can be performed by comparing some distance between the stream and previously learnt time series. However, learning a few incorrect time series can trigger disproportionately many false alarms. In this paper, we propose a fast technique to prune bad instances away and automatically select appropriate distance thresholds. Our main contribution is to turn the ill-defined spotting problem into a collection of single well-defined binary classification problems, by segmenting the stream and by ranking subsets of instances on those segments very quickly. We further demonstrate our technique's effectiveness on a gesture recognition application.
|
|
15:00-17:10, Paper WePT1.15 | |
Instance Selection Using Non-Linear Sparse Modeling |
Dornaika, Fadi | Univ. of the Basque Country |
Kamal Aldine, Ihab | Univ. OF THE BASQUE COUNTRY |
Ruichek, Yassine | Univ. De Tech. De Belfort-Montbeliard |
Keywords: Machine learning and data mining, Classification and clustering, Statistical, syntactic and structural pattern recognition
Abstract: Sparse Modeling Representative Selection (SMRS) has been recently introduced for selecting the most relevant examples in datasets. SMRS exploits data self-representativeness coding in order to infer a coding matrix with block sparsity constraint. The relevance scores of samples are then derived from the estimated matrix of coefficients. Since SMRS is based on a linear model for data self-representation, it cannot always provide good relevant samples. Besides, most of its selected samples can be found in dense areas in input space. In this paper, we propose to overcome the SMRS method's shortcomings that are related to the coding matrix estimation. We introduce two non-linear data self-representativeness coding schemes that are based on Hilbert space and column generation. Experimental evaluation is carried out on summarizing a video movie and on summarizing training image datasets used for classification tasks. These experiments demonstrated that the proposed non-linear methods can outperform state-of-the art selection methods including the SMRS method.
|
|
15:00-17:10, Paper WePT1.16 | |
Constrained Dominant Sets for Retrieval |
Mequanint, Eyasu Zemene | Ca'Foscari Uiversity of Venice |
Alemu, Leulseged Tesfaye | Ca'Foscari Univ. of Venice |
Pelillo, Marcello | Ca' Foscari Univ |
Keywords: Machine learning and data mining, Content based image retrieval and data mining
Abstract: Learning new global relations based on an initial affinity of the database objects has shown significant improvements in similarity retrievals. Locally constrained diffusion process is one of the recent effective tools in learning the intrinsic manifold structure of a given data. Existing methods, which constrain the diffusion process locally, have problems - manual choice of optimal local neighborhood size, do not allow for intrinsic relation among the neighbors, fix initialization vector to extract dense neighbor - which negatively affect the affinity propagation. We propose a new approach, which alleviate these issues, based on some properties of a family of quadratic optimization problems related to dominant sets, a well-known graph theoretic notion of a cluster which generalizes the concept of a maximal clique to edge-weighted graphs. In particular, we show that by properly controlling a regularization parameter which determines the structure and the scale of the underlying problem, we are in a position to extract dominant set cluster which is constrained to contain user-provided query. Experimental results on standard benchmark datasets show the effectiveness of the proposed approach.
|
|
15:00-17:10, Paper WePT1.17 | |
Hyperparameter Tuning for Big Data Using Bayesian Optimisation |
Theckel Joy, Tinu | Deakin Univ |
Rana, Santu | Deakin Univ |
Gupta, Sunil Kumar | Deakin Univ |
Venkatesh, Svetha | Deakin Univ |
Keywords: Machine learning and data mining, Model selection, Deep learning
Abstract: Hyperparameters play a crucial role in the model selection of machine learning algorithms. Tuning these hyperparameters can be exhaustive when the data is large. Bayesian optimisation has emerged as an efficient tool for hyperparameter tuning of machine learning algorithms. In this paper, we propose a novel framework for tuning the hyperparameters for big data using Bayesian optimisation. We divide the big data into chunks and generate hyperparameter configurations for the chunks using the standard Bayesian optimisation. We utilise the information from the chunks for the hyperparameter tuning for the big data using a transfer learning setting. We evaluate the performance of the proposed method on the task of tuning hyperparameters of two machine learning algorithms. We show that our method achieves the best available hyperparameter configuration within less computational time compared to the state-of-art hyperparameter tuning methods.
|
|
15:00-17:10, Paper WePT1.18 | |
 Bayesian Regression Selecting Valuable Subset from Mixed Bag Training Data |
Katsuki, Takayuki | IBM Res. - Tokyo |
Inoue, Masato | Waseda Univ |
Attachments: Supplementary material
Keywords: Machine learning and data mining, Other applications
Abstract: This paper addresses a problem in which we learn a regression model from sets of training data. Each of the sets has an only single label, and only one of the training data in the set reflects the label. This is particularly the case when the label is attached to a group of data, such as time-series data. The label is not attached to the point of the sequence but rather attached to particular time window of the sequence. As such, a small part of the time window likely reflects the label, whereas the other larger part of the time window likely does not reflect it. We design an algorithm for estimating which of the training data in each of the sets corresponds to the label, as well as for training the regression model on the basis of Bayesian modeling and posterior inference with variational Bayes. Our experimental results show that our approach perform better than baseline methods on an artificial dataset and on a real-world dataset.
|
|
15:00-17:10, Paper WePT1.19 | |
Quantile Regression of Interval-Valued Data |
Fagundes, Roberta | Univ. De Pernambuco |
Souza, Renata | Univ. Federal De Pernambuco |
Soares, Yanne | Univ. De Pernambuco |
Keywords: Machine learning and data mining, Performance Evaluation, Model selection
Abstract: Linear regression is a standard statistical method widely used for prediction. It focuses on modeling the mean the target variable without accounting for all the distributional properties of this variable. In contrast, the quantile regression model facilitates the analysis of the full distributional properties, it allows to model different quantities of the target variable. This paper proposes a quantile regression model for interval data. In this model, each interval variable of the input data is represented by its range and center and a smooth function between two vectors composed by interval variables are defined. In order to test the usefulness of the proposed model, a simulation study is undertaken and an application using a scientific production interval data set of institutions from Brazil are performed. The quality of the interval prediction obtained by the proposed model is assessed by mean magnitude of relative error calculated from test data.
|
|
15:00-17:10, Paper WePT1.20 | |
Reinforcement Learning Via Recurrent Convolutional Neural Networks |
Shankar, Tanmay | Indian Inst. of Tech. Guwahati |
Dwivedy, Santosha Kumar | Indian Inst. of Tech. Guwahati |
Guha, Prithwijit | Department of EEE, IIT Guwahati |
Keywords: Reinforcement learning and temporal models, Deep learning, Artificial neural networks
Abstract: Deep Reinforcement Learning has enabled the learning of policies for complex tasks in partially observable environments, without explicitly learning the underlying model of the tasks. While such model-free methods do achieve considerable performance, they often ignore the structure of task. We present a more natural representation of the solutions to Reinforcement Learning (RL) problems, within 3 Recurrent Convolutional Neural Network (RCNN) architectures to better exploit this inherent structure. The forward passes of each RCNN execute an efficient Value Iteration, propagate beliefs of state in partially observable environments, and choose optimal actions respectively. Applying back-propagation to these RCNNs allows the system to explicitly learn the Transition Model and Reward Function associated with the underlying MDP, serving as an elegant alternative to classical model-based RL. We evaluate the proposed algorithms in simulation, considering a robot planning problem. We demonstrate the capability of our framework to reduce the cost of re-planning, learn accurate MDP models, and finally replan with learned models to achieve near-optimal policies.
|
|
15:00-17:10, Paper WePT1.21 | |
A Distance-Based Shape Descriptor Invariant to Similitude and Its Application to Shape Classification |
Presles, Benoit | Univ. Coll. London |
Debayle, Johan | Ec. Nationale Supérieure Des Mines De Saint-Etienne |
Keywords: Representation and analysis in pixel/voxel images, Classification and clustering
Abstract: Pattern recognition usually requires to describe or represent shapes with some features, called shape escriptors. A shape descriptor generally needs to be invariant to some geometrical transformations (translation, rotation, scaling...). In addition, it has to be robust against slight deformations or noise damaging of the shape. In this paper, a novel shape descriptor based on distances and invariant to similitude transformation is proposed. A metric associated to the proposed descriptor is then introduced to measure the dissimilarity between shapes. Performance tests are evaluated on the Kimia and MPEG7 image databases to evaluate the quality of the proposed descriptor. More specifically, the proposed method shows a better performance for shape classification in comparison to some methods from the literature.
|
|
15:00-17:10, Paper WePT1.22 | |
A Soft-Labeled Self-Training Approach |
Mey, Alexander | TU Delft |
Loog, Marco | Delft Univ. of Tech. / Univ. of Copenhagen |
Keywords: Semi-supervised learning and spectral methods
Abstract: Semi-supervised classification methods try to improve a supervised learned classifier with the help of unlabeled data. In many cases one assumes a certain structure on the data, as for example the manifold assumption, the smoothness assumption or the cluster assumption. Self-training is a method that does not need any assumptions on the data itself. The idea is to use the supervised trained classifier to label the unlabeled points and to enlarge this way the training data. This paper aims to show that a self-training approach with soft-labeling is preferable in many cases in terms of expected loss (risk) minimization. The main idea is to use a soft-labeling to minimize the risk on labeled and unlabeled data together, in which the hard-labeled self-training is an extreme case.
|
|
15:00-17:10, Paper WePT1.23 | |
Deep Sparse-Coded Network (DSN) |
Gwon, Youngjune | Harvard Univ |
Cha, Miriam | Harvard Univ |
Kung, H. T. | Harvard Univ |
Keywords: Semi-supervised learning and spectral methods, Classification and clustering, Deep learning
Abstract: We present Deep Sparse-coded Network (DSN), a deep architecture based on multilayer sparse coding. It has been considered difficult to learn a useful feature hierarchy by stacking sparse coding layers in a straightforward manner. The primary reason is the modeling assumption for sparse coding that takes in a dense input and yields a sparse output vector. Applying a sparse coding layer on the output of another tends to violate the modeling assumption. We overcome this shortcoming by interlacing nonlinear pooling units. Average- or max-pooled sparse codes are aggregated to form dense input vectors for the next sparse coding layer. Pooling achieves nonlinear activation analogous to neural networks while not introducing diminished gradient flows during the training. We introduce a novel backpropagation algorithm to finetune the proposed DSN beyond the pretraining via greedy layerwise sparse coding and dictionary learning. We build an experimental 4-layer DSN with the L1-regularized LARS and the greedy-L0 OMP, and demonstrate superior performance over a similarly-configured stacked autoencoder (SAE) on CIFAR-10.
|
|
15:00-17:10, Paper WePT1.24 | |
Dynamic Adaptive Graph Construction: Application to Graph-Based Multi-Observation Classification |
Dornaika, Fadi | Univ. of the Basque Country |
Dahbi, Radouan | Univ. of Tech. of Belfort-Montbéliard |
Bosaghzadeh, Alireza | Univ. of Basque Country |
Ruichek, Yassine | Univ. De Tech. De Belfort-Montbeliard |
Keywords: Semi-supervised learning and spectral methods, Classification and clustering, Machine learning and data mining
Abstract: Most of graph construction techniques assume a transductive setting in which the whole data collection is available at construction time. Addressing graph construction for inductive setting, in which data are coming sequentially, has received much less attention. Constructing the graph from scratch can be very time consuming. In this paper, we propose an efficient dynamic graph construction method that adds new samples (labeled or unlabeled) to a previously constructed graph. We use a Two Phase Weighted Regularized Least Square (TPWRLS) coding scheme to represent new sample(s) with respect to an existing data set. The representative coefficients are then used to update the graph affinity matrix. The proposed method not only appends the new samples to the graph but also updates the whole graph structure by discovering which nodes are affected by the introduction of new samples and by updating their edge weights. The proposed construction framework is applied to the problem of graph-based label propagation using multiple observations in a semi-supervised scenario. Experiments on three public image databases show that, without any significant loss in the accuracy of the final classification, the proposed dynamic graph construction is more efficient than the batch graph construction.
|
|
15:00-17:10, Paper WePT1.25 | |
Model-Based Classification and Novelty Detection for Point Pattern Data |
Vo, Ba-Ngu | Curtin Univ |
Tran, Nhat-Quang | Curtin Univ |
Phung, Dinh | Deakin Univ |
Vo, Ba-Tuong | Curtin Univ |
Keywords: Semi-supervised learning and spectral methods, Classification and clustering, Machine learning and data mining
Abstract: Point patterns are sets or multi-sets of unordered elements that can be found in numerous data sources. However, in data analysis tasks such as classification and novelty detection, appropriate statistical models for point pattern data have not received much attention. This paper proposes the modelling of point pattern data via random finite sets (RFS). In particular, we propose appropriate likelihood functions, and a maximum likelihood estimator for learning a tractable family of RFS models. In novelty detection, we propose novel ranking functions based on RFS models, which substantially improve performance.
|
|
15:00-17:10, Paper WePT1.26 | |
A PAC Bound for Joint Matrix Completion Via Partially Collective Matrix Factorization |
Lan, Chao | Univ. of Kansas |
Li, Xiaoli | Univ. of Kansas |
Deng, Yujie | Univ. of Kansas |
Amand, Joseph St. | Univ. of Kansas |
Huan, Jun | Univ. of Kansas |
Keywords: Semi-supervised learning and spectral methods, Transfer learning, Machine learning and data mining
Abstract: Collective Matrix Factorization (CMF) is a popular model for the joint matrix completion task, but limited by its strong assumption that all matrices share the same low-rank structure. Recently, an alternative model was proposed with a relaxed assumption that matrix low-rank structures are partly shared. We refer this model as Partially Collective Matrix Factorization (P-CMF). This paper presents a first PAC generalization error bound for joint matrix completion based on the P-CMF model. Our technical contributions are tri-facet. First, we derive a new PAC bound for single matrix completion, which fundamentally improves the existing PAC bound in multiple aspects. Then, based on it we derive the first PAC bound for joint matrix completion based on the P-CMF model. This not only justifies the theoretical soundness of P-CMF, but also reveals its several insights. Finally, we present a model construction criterion for P-CMF based methods, which specifies the degrees of sharing between matrix low-rank structures. We demonstrate the effectiveness of this criterion in simulation.
|
|
15:00-17:10, Paper WePT1.27 | |
Shape Classification with a Vertex Clustering Graph Kernel |
Bai, Lu | Central Univ. of Finance and Ec |
Cui, Lixin | School of Information, Central Univ. of Finance and Ec |
Wang, Yue | Central Univ. of Finance and Ec |
Bai, Xiao | Beihang Univ |
Hancock, Edwin | Univ. of York |
Jin, Xin | Central Univ. of Finance and Ec. Beijing, China |
Keywords: 2D/3D object detection and recognition, Machine learning and data mining
Abstract: Graph kernels are powerful tools for structural analysis in computer vision. Unfortunately, most existing state-of-the-art graph kernels ignore the locational or structural correspondence information between graphs, based on the visual background. This drawback influences the performance of existing kernels for computer vision based classification problems, e.g., classification of shapes, point clouds and digital images. The aim of this paper is to address the problem with existing kernels, by developing a novel vertex clustering graph kernel. We show that this kernel not only overcomes the shortcoming of ignoring correspondence information between isomorphic substructures that arises in most existing graph kernels, but also guarantees the transitivity between the correspondence information. Our kernel can easily outperform state-of-the-art graph kernels in terms of classification accuracy on standard shape based graph datasets.
|
|
15:00-17:10, Paper WePT1.28 | |
Ensemble-Based Local Learning for High-Dimensional Data Regression |
Raytchev, Bisser | Hiroshima Univ |
Katamoto, Yoshinari | Hiroshima Univ |
Koujiba, Miku | Hiroshima Univ |
Tamaki, Toru | Hiroshima Univ |
Kaneda, Kazufumi | Hiroshima Univ |
Keywords: Active and ensemble learning, Classification and clustering
Abstract: In this paper we propose a new local learning based regression method which utilizes ensemble-learning as a form of regularization to reduce the variance of local estimators. This makes it possible to use local learning methods even with very high-dimensional datasets. The efficacy of the proposed method is illustrated on two publicly available high-dimensional sets in comparison with several global learning methods, and it is shown that the proposed ensemble-based local learning method significantly outperforms the global ones.
|
|
15:00-17:10, Paper WePT1.29 | |
Active Learning Using Uncertainty Information |
Yang, Yazhou | Delft Univ. of Tech |
Loog, Marco | Delft Univ. of Tech. / Univ. of Copenhagen |
Keywords: Active and ensemble learning, Machine learning and data mining, Model selection
Abstract: Many active learning methods belong to the retraining-based approaches, which select one unlabeled instance, add it to the training set with its possible labels, retrain the classification model, and evaluate the criteria that we base our selection on. However, since the true label of the selected instance is unknown, these methods resort to calculating the average-case or worse-case performance with respect to the unknown label. In this paper, we propose a different method to solve this problem. In particular, our method aims to make use of the uncertainty information to enhance the performance of retraining-based models. We apply our method to two state-of-the-art algorithms and carry out extensive experiments on a wide variety of real-world datasets. The results clearly demonstrate the effectiveness of the proposed method and indicate it can reduce human labeling efforts in many real-life applications.
|
|
15:00-17:10, Paper WePT1.30 | |
D-LSM: Deep Liquid State Machine with Unsupervised Recurrent Reservoir Tuning |
Wang, Qian | Texas A&M Univ |
Li, Peng | Texas A&M Univ |
Keywords: Artificial neural networks, Deep learning
Abstract: The Liquid State Machine (LSM) is a biologically plausible model of computation for recurrent spiking neural networks, which offers promising solutions to real-world applications in both software and hardware based systems. At the same time, deep feedforward rate-based neural networks such as convolutional neural networks (CNNs) have achieved great success in many computer vision related applications. However, a systematic exploration of deep recurrent spiking neural networks is lacking. We propose a new model of Deep Liquid State Machine (D-LSM), which simultaneously explores the powers of recurrent spiking networks and deep architectures. D-LSM consists of multiple basic LSM processing and pooling stages. Recurrent reservoir networks across different LSM stages act as nonlinear filters capable of extracting spatio-temporal features of increasingly higher levels from the input. We propose to train the D-LSM practically by adopting unsupervised training (e.g. through STDP) for recurrent reservoirs and spike-based supervised rules for the final readout stage. The perspective of realizing D-LSM based hardware processors is also presented.
|
|
15:00-17:10, Paper WePT1.31 | |
Generating Commentaries for Tennis Videos |
Yan, Fei | Univ. of Surrey |
Mikolajczyk, Krystian | Univ. of Surrey |
Kittler, Josef | Univ. of Surrey |
Keywords: Artificial neural networks, Image and video analysis and understanding, Deep learning
Abstract: We present an approach to automatically generating verbal commentaries for tennis games. We introduce a novel application that requires a combination of techniques from computer vision, natural language processing and machine learning. A video sequence is first analysed using state-of-the-art computer vision methods to track the ball, fit the detected edges to the court model, track the players, and recognise their strokes. Based on the recognised visual attributes we formulate the tennis commentary generation problem in the framework of long short-term memory recurrent neural networks as well as structured SVM. In particular, we investigate pre-embedding of descriptive terms and loss function for LSTM. We introduce a new dataset of 633 annotated pairs of tennis videos and corresponding commentary. We perform an automatic as well as human based evaluation, and demonstrate that the proposed pre-embedding and loss function lead to substantially improved accuracy of the generated commentary.
|
|
15:00-17:10, Paper WePT1.32 | |
Linear Model Optimizer vs Neural Networks: A Comparison for Improving the Quality and Saving of LED-Lighting Control Systems |
Lobato-Rios, Victor | Inst. Nacional De Astrofisica, Optica Y Electronica |
Hernandez-Castañon, Viviana del Rocio | Inst. Nacional De Astrofisica, Optica Y Electronica |
Carrasco-Ochoa, Jesus Ariel | National Inst. of Astrophysics, Optics and Electronics |
Martinez-Trinidad, Francisco | National Inst. of Astrophysics, Optics and Electronics |
Keywords: Artificial neural networks, Machine learning and data mining
Abstract: Lighting systems represents about 38% of the total energy consumption in office buildings, however, a great amount of this energy is wasted because luminaires keep working at its maximum power even when just a single person is present. In order to improve the performance of LED-Lighting control systems, we propose a linear model that considers the luminaire’s influence on its neighborhood and takes into account visual comfort and energy consumption. The proposed linear model was contrasted against two Neural Network configurations that were trained to find the best dimming levels. Our experiments demonstrate that a linear optimizer applied to our proposed linear model have a better performance than any of the two tested neural networks.
|
|
15:00-17:10, Paper WePT1.33 | |
 Point Cloud Labeling Using 3D Convolutional Neural Network |
Huang, Jing | Univ. of Southern California |
You, Suya | Univ. of Southern California |
Attachments: Supplementary material
Keywords: Deep learning, Representation and analysis in pixel/voxel images, Scene understanding
Abstract: In this paper, we tackle the labeling problem for 3D point clouds. We introduce a 3D point cloud labeling scheme based on 3D Convolutional Neural Network. Our approach minimizes the prior knowledge of the labeling problem and does not require a segmentation step or hand-crafted features as most previous approaches did. Particularly, we present solutions for large data handling during the training and testing process. Experiments performed on the urban point cloud dataset containing 7 categories of objects show the robustness of our approach.
|
|
15:00-17:10, Paper WePT1.35 | |
Evolutionary Data Purification for Social Media Classification |
James, Stuart | Univ. of Surrey |
Collomosse, John Philip | Univ. of Surrey |
Keywords: Image and video analysis and understanding, Classification and clustering, Model selection
Abstract: We present a novel algorithm for the semantic labeling of photographs shared via social media. Such imagery is diverse, exhibiting high intra-class variation that demands large training data volumes to learn representative classifiers. Unfortunately image annotation at scale is noisy resulting in errors in the training corpus that confound classifier accuracy. We show how evolutionary algorithms may be applied to select a 'purified' subset of the training corpus to optimize classifier performance. We demonstrate our approach over a variety of image descriptors (including deeply learned features) and support vector machines.
|
|
15:00-17:10, Paper WePT1.36 | |
Estimates of Classification Complexity for Myoelectric Pattern Recognition |
Nilsson, Niclas | Chalmers Univ. of Tech |
Ortiz-Catalan, Max | Chalmers Univ. of Tech |
Keywords: Machine learning and data mining, Classification and clustering, Dimensionality reduction and manifold learning
Abstract: Myoelectric pattern recognition (MPR) can be used for intuitive control of virtual and robotic effectors in clinical applications such as prosthetic limbs and the treatment of phantom limb pain. The conventional approach is to feed classifiers with descriptive electromyographic (EMG) features that represent the aimed movements. The complexity and consequently classification accuracy of MPR is highly affected by the separability of such features. In this study, classification complexity estimating algorithms were investigated as a potential tool to estimate MPR performance. An early prediction of MPR accuracy could inform the user of faulty data acquisition, as well as suggest the repetition or elimination of detrimental movements in the repository of classes. Two such algorithms, Nearest Neighbor Separability (NNS) and Separability Index (SI), were found to be highly correlated with classification accuracy in three commonly used classifiers for MPR (Linear Discriminant Analysis, Multi-Layer Perceptron, and Support Vector Machine). These Classification Complexity Estimating Algorithms (CCEAs) were implemented in the open source software “BioPatRec” and are available freely online. This work deepens the understanding of the complexity of MPR for the prediction of motor volition.
|
|
15:00-17:10, Paper WePT1.37 | |
Multiple Instance Dictionary Learning Using Functions of Multiple Instances |
Jiao, Changzhe | Univ. of Missouri |
Zare, Alina | Univ. of Missouri |
Keywords: Machine learning and data mining, Classification and clustering, Image and video analysis and understanding
Abstract: Dictionary Learning Functions of Multiple Instances (DL-FUMI) is proposed to address target detection problems with inaccurate training labels. DL-FUMI is a multiple instance dictionary learning method that estimates target atoms that describe distinctive and representative features of the target class and background atoms that account for the shared features found across both target and non-target data points. Experimental results show that the target atoms estimated by DL-FUMI are more discriminative and representative of the target class than comparison methods. DL-FUMI is shown to have improved performance on several detection problems as compared to other multiple instance dictionary learning algorithms.
|
|
15:00-17:10, Paper WePT1.38 | |
One-Shot Learning of Temporal Sequences Using a Distance Dependent Chinese Restaurant Process |
Orrite, Carlos | Univ. of Zaragoza |
Rodríguez, Mario | Univ. of Zaragoza |
Medrano, Carlos | Univ. of Zaragoza |
Keywords: Machine learning and data mining, Classification and clustering, Model selection
Abstract: Activity recognition in videos is a challenging task, mainly if a scarce number of samples is available for modelling the problem. The task becomes even harder when using generative models such as mixture models or Hidden Markov Models (HMMs), as they demand a lot of samples to determinate their parameters. Additionally, these models rely on the appropriate selection of some parameters, for instance the number of hidden states. Therefore, we propose in this paper the creation of a Universal Background Model (UBM) of features, using videos from public datasets, applied to the activity encoding and an unsupervised modelling of the activities with a distance dependent Chinese Restaurant Process (ddCRP), where the number of states (tables in the Chinese Restaurant descriptions) is automatically determined by the process. In order to classify an incoming video-sequence we propose to model it as a ddCRP distribution and to apply a nearest neighbour algorithm based on a kernel between distributions. To carry out this process we use a Probability Product Kernel (PPK) algorithm by previously mapping the ddCRP into a HMM with discrete observations. Preliminary experiments in two public data sets, as Weizmann and KTH, show that this proposal achieves state-of-the-art results.
|
|
15:00-17:10, Paper WePT1.39 | |
 Regression-Based Metric Learning |
Moutafis, Panagiotis | Univ. of Houston |
Leng, Mengjun | Univ. of Houston |
Kakadiaris, Ioannis | Univ. of Houston |
Attachments: Supplementary material
Keywords: Machine learning and data mining, Classification and clustering, Other applications
Abstract: Existing distance metric learning methods define an objective function and seek a distance metric (or equivalently a projection) that minimizes it. In this paper, we propose a different approach that illustrates how to formulate distance metric learning as a regression problem. First, the objective function is minimized to learn target representations. Then, a regression method is employed to learn a projection that maps the input to the target representations. This global projection function is the single output of the proposed algorithm. Our contribution is a different perspective on how to train a distance metric learning algorithm. The advantages are: (i) this approach has the potential to simplify the optimization process; and (ii) it allows researchers to leverage the power of existing regression methods and those to be invented. Experimental results on several publicly available datasets illustrate that the proposed framework can learn a distance metric with discriminative properties.
|