Entropic one-class classifiers

Entropic one-class classifiers
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

The one-class classification problem is a well-known research endeavor in pattern recognition. The problem is also known under different names, such as outlier and novelty/anomaly detection. The core of the problem consists in modeling and recognizing patterns belonging only to a so-called target class. All other patterns are termed non-target, and therefore they should be recognized as such. In this paper, we propose a novel one-class classification system that is based on an interplay of different techniques. Primarily, we follow a dissimilarity representation based approach; we embed the input data into the dissimilarity space by means of an appropriate parametric dissimilarity measure. This step allows us to process virtually any type of data. The dissimilarity vectors are then represented through a weighted Euclidean graphs, which we use to (i) determine the entropy of the data distribution in the dissimilarity space, and at the same time (ii) derive effective decision regions that are modeled as clusters of vertices. Since the dissimilarity measure for the input data is parametric, we optimize its parameters by means of a global optimization scheme, which considers both mesoscopic and structural characteristics of the data represented through the graphs. The proposed one-class classifier is designed to provide both hard (Boolean) and soft decisions about the recognition of test patterns, allowing an accurate description of the classification process. We evaluate the performance of the system on different benchmarking datasets, containing either feature-based or structured patterns. Experimental results demonstrate the effectiveness of the proposed technique.


💡 Research Summary

The paper introduces a novel one‑class classification (OCC) framework that unifies several advanced concepts to handle arbitrary data types, including feature vectors and structured objects such as labeled graphs. The core pipeline begins by embedding the original data into a dissimilarity space (DS) using a parametric dissimilarity measure d_I(x, r; p) defined over a representation set R. By varying the parameters p, the view of the data in DS can be adapted to highlight relevant structures. The rows of the resulting dissimilarity matrix serve as the embedded vectors, allowing any subsequent algebraic operation.

Once embedded, the data points become vertices of a complete weighted Euclidean graph G(p). Edge weights are the squared Euclidean distances in DS, which are normalized to lie in


Comments & Academic Discussion

Loading comments...

Leave a Comment