Javascript must be enabled to continue!
Small Stochastic Data Compactification Concept Justified in the Entropy Basis
View through CrossRef
Measurement is a typical way of gathering information about an investigated object, generalized by a finite set of characteristic parameters. The result of each iteration of the measurement is an instance of the class of the investigated object in the form of a set of values of characteristic parameters. An ordered set of instances forms a collection whose dimensionality for a real object is a factor that cannot be ignored. Managing the dimensionality of data collections, as well as classification, regression, and clustering, are fundamental problems for machine learning. Compactification is the approximation of the original data collection by an equivalent collection (with a reduced dimension of characteristic parameters) with the control of accompanying information capacity losses. Related to compactification is the data completeness verifying procedure, which is characteristic of the data reliability assessment. If there are stochastic parameters among the initial data collection characteristic parameters, the compactification procedure becomes more complicated. To take this into account, this study proposes a model of a structured collection of stochastic data defined in terms of relative entropy. The compactification of such a data model is formalized by an iterative procedure aimed at maximizing the relative entropy of sequential implementation of direct and reverse projections of data collections, taking into account the estimates of the probability distribution densities of their attributes. The procedure for approximating the relative entropy function of compactification to reduce the computational complexity of the latter is proposed. To qualitatively assess compactification this study undertakes a formal analysis that uses data collection information capacity and the absolute and relative share of information losses due to compaction as its metrics. Taking into account the semantic connection of compactification and completeness, the proposed metric is also relevant for the task of assessing data reliability. Testing the proposed compactification procedure proved both its stability and efficiency in comparison with previously used analogues, such as the principal component analysis method and the random projection method.
Title: Small Stochastic Data Compactification Concept Justified in the Entropy Basis
Description:
Measurement is a typical way of gathering information about an investigated object, generalized by a finite set of characteristic parameters.
The result of each iteration of the measurement is an instance of the class of the investigated object in the form of a set of values of characteristic parameters.
An ordered set of instances forms a collection whose dimensionality for a real object is a factor that cannot be ignored.
Managing the dimensionality of data collections, as well as classification, regression, and clustering, are fundamental problems for machine learning.
Compactification is the approximation of the original data collection by an equivalent collection (with a reduced dimension of characteristic parameters) with the control of accompanying information capacity losses.
Related to compactification is the data completeness verifying procedure, which is characteristic of the data reliability assessment.
If there are stochastic parameters among the initial data collection characteristic parameters, the compactification procedure becomes more complicated.
To take this into account, this study proposes a model of a structured collection of stochastic data defined in terms of relative entropy.
The compactification of such a data model is formalized by an iterative procedure aimed at maximizing the relative entropy of sequential implementation of direct and reverse projections of data collections, taking into account the estimates of the probability distribution densities of their attributes.
The procedure for approximating the relative entropy function of compactification to reduce the computational complexity of the latter is proposed.
To qualitatively assess compactification this study undertakes a formal analysis that uses data collection information capacity and the absolute and relative share of information losses due to compaction as its metrics.
Taking into account the semantic connection of compactification and completeness, the proposed metric is also relevant for the task of assessing data reliability.
Testing the proposed compactification procedure proved both its stability and efficiency in comparison with previously used analogues, such as the principal component analysis method and the random projection method.
Related Results
Entropy and Wealth
Entropy and Wealth
While entropy was introduced in the second half of the 19th century in the international vocabulary as a scientific term, in the 20th century it became common in colloquial use. Po...
Influence of ideals in compactifications
Influence of ideals in compactifications
Abstract
One point compactification is studied in the light of ideal of subsets of ℕ. ????-proper map is introduced and showed that a continuous map can be extended ...
Discussion on the Full Entropy Assumption of the SP 800-90 Series
Discussion on the Full Entropy Assumption of the SP 800-90 Series
NIST SP 800-90 series support the generation of high-quality random bits for cryptographic and non-cryptographic use. The security of a random number generator depends on the unpre...
Cross-Subject Emotion Recognition Using Fused Entropy Features of EEG
Cross-Subject Emotion Recognition Using Fused Entropy Features of EEG
Emotion recognition based on electroencephalography (EEG) has attracted high interest in fields such as health care, user experience evaluation, and human–computer interaction (HCI...
Justification, epistemic
Justification, epistemic
The term ‘justification’ belongs to a cluster of normative terms that also includes ‘rational’, ‘reasonable’ and ‘warranted’. All these are commonly used in epistemology, but there...
Entropy-guided sevoflurane administration during cardiopulmonary bypass surgery in the paediatric population
Entropy-guided sevoflurane administration during cardiopulmonary bypass surgery in the paediatric population
Background
Maintaining optimal anesthetic depth during cardiopulmonary bypass (CPB) in pediatric patients is challenging due to altered physiology and unreliable conven...
Quantum wave entropy
Quantum wave entropy
In quantum mechanics, particles have a new type of probabilistic property, which is quantum wave probability. Corresponding to this new probability, the particle has the property o...
Thermodynamics of High Temperature Plasmas
Thermodynamics of High Temperature Plasmas
In this work we discuss how and to what extent the thermodynamic concepts and the thermodynamic formalism can be extended to the description of high temperature states of the plasm...

