 Research
 Open Access
 Published:
A data value metric for quantifying information content and utility
Journal of Big Data volume 8, Article number: 82 (2021)
Abstract
Datadriven innovation is propelled by recent scientific advances, rapid technological progress, substantial reductions of manufacturing costs, and significant demands for effective decision support systems. This has led to efforts to collect massive amounts of heterogeneous and multisource data, however, not all data is of equal quality or equally informative. Previous methods to capture and quantify the utility of data include value of information (VoI), quality of information (QoI), and mutual information (MI). This manuscript introduces a new measure to quantify whether larger volumes of increasingly more complex data enhance, degrade, or alter their information content and utility with respect to specific tasks. We present a new informationtheoretic measure, called Data Value Metric (DVM), that quantifies the useful information content (energy) of large and heterogeneous datasets. The DVM formulation is based on a regularized model balancing data analytical value (utility) and model complexity. DVM can be used to determine if appending, expanding, or augmenting a dataset may be beneficial in specific application domains. Subject to the choices of data analytic, inferential, or forecasting techniques employed to interrogate the data, DVM quantifies the information boost, or degradation, associated with increasing the data size or expanding the richness of its features. DVM is defined as a mixture of a fidelity and a regularization terms. The fidelity captures the usefulness of the sample data specifically in the context of the inferential task. The regularization term represents the computational complexity of the corresponding inferential method. Inspired by the concept of information bottleneck in deep learning, the fidelity term depends on the performance of the corresponding supervised or unsupervised model. We tested the DVM method for several alternative supervised and unsupervised regression, classification, clustering, and dimensionality reduction tasks. Both real and simulated datasets with weak and strong signal information are used in the experimental validation. Our findings suggest that DVM captures effectively the balance between analyticalvalue and algorithmiccomplexity. Changes in the DVM expose the tradeoffs between algorithmic complexity and data analytical value in terms of the samplesize and the featurerichness of a dataset. DVM values may be used to determine the size and characteristics of the data to optimize the relative utility of various supervised or unsupervised algorithms.
Introduction
Background
Big data sets are becoming ubiquitous, emphasizing the importance of solving the challenge of balancing information utility, data value, resource costs, computational efficiency, and inferential reliability [1]. This manuscript tackles this problem by developing a new measure, called the Data Value Metric (DVM), that quantifies the energy, or information content, of large and complex datasets, which can be used as a yardstick to determine if appending, expanding, or otherwise augmenting the data size or complexity may be beneficial in specific application domains. In practice, DVM provides a mechanism to balance, or tradeoff, a pair of competing priorities (1) costs or tradeoffs associated with increasing or decreasing the size of heterogeneous datasets (sample size) and controlling the sampling error rate, and (2) expected gains (e.g., decisionmaking improvement) or losses (e.g., decrease of precision or variability increase) associated with the corresponding scientific inference. The computational complexity of the DVM method is directly proportional to that of calculating mutual information, which is linear in terms of the data size. Thus, the DVM complexity is determined directly by the inferential method or technique used to obtain the classification, regression, or clustering results, which may itself be nonlinear. Hence, DVM calculations do not add significant overhead to the standard analytical protocol.
Although several performance measures exist for supervised and unsupervised inference tasks, it is difficult to use established methods to infer the sufficiency of the data for each specific inferential task. For example, one could use accuracy measures for a classification task. Assume that the accuracy of \(70\%\) is achieved for a nonrandom, nonstationary, or nonhomogeneous dataset. Then, the question is whether we can expect an increase of the accuracy by adding more samples or more features, or maybe use alternative models to increase the value of the resulting inference. In general, such questions are difficult to answer solely by considering a particular measure of performance on a given dataset. Several of the previous approaches measuring the quality of data are summarized below.
Related work
Several previous studies have proposed metrics for assessing the information gain of a given dataset. For example, value of information (VoI) analysis, originally proposed in [2] with overviews in [3,4,5], is a decisiontheoretic statistical framework representing the expected increased inference accuracy or reduction in loss based on additional prospective information [6]. The basic three types of VoI methods include (1) inferential and modeling cases for linear objective functions under simplified parameter distribution restrictions, which limits their broad practical applicability [3, 7]; (2) methods for estimating the expected value of partial perfect information (EVPPI) involving partitioning of the parameter space into smaller subsets and assuming constant and optimal inference over the local neighborhoods, within subsets [8, 9]; and (3) Gaussian process regression methods approximating the expected inference [10,11,12]. More specifically, for a particular parameter \(\phi\), the EVPPI is the expected inferential gain, or reduction in loss, when \(\phi\) is perfectly estimated. As the perfect \(\phi\) is unknown in advance, this reduction of loss expectation is taken over the entire parameter space \(\phi \in \Phi\):
where d is decision, inference or action, \(d^*_{\phi }\) is the optimal inference obtained when \(\phi\) is known, \(\theta\) is the the model parameter vector, E is the expectation, and \(L(d,\theta )\) is the likelihood function [6]. Note that VoI techniques are mainly suitable for specific types of problems, such as evidence synthesis in the context of decision theory. Further, their computational complexity tends to be high and require nested Monte Carlo procedures.
Another relevant study [13] utilizes a unique decomposition of the differences (errors) between theoretical (population) parameters and their sampledriven estimates (statistics) into three independent components. If \(\theta\) and \({\hat{\theta }}\) represent a theoretical characteristic of interest (e.g., population mean) and its samplebased parameter estimate (e.g., sample arithmetic average), respectively, then, the error can be canonically decomposed as:
Suppose J is a (uniform) random subset indexing a sample from the entire (finite, N) population. For a sample \(\{X_i: j\in I_n \}\), \(R_j\) is a randomsample indicator function (with values 0 or 1) capturing whether \(j\in I_n\). Of course, \(\sum _{j=1}^N {R_j}=n\), X is a multidimensional design matrix capturing the attributes of the data (features), \(g:X \longrightarrow \mathfrak {R}\) is a linking map that allows us to compute on samples (e.g., polynomial functions for moment calculations or indicator functions for distribution functions), \(g_j=g(X_j)\) is a mapping of the jth feature, \(A=A(g,R)\) is a measure of association between \(R_J\) and \(G_J\), the sampling rate \(f=E_J(R_J)=\frac{n}{N}\) (ratio of sampletopopulation size), \(B=\sqrt{\frac{1f}{f}}\), and C is a measure encoding the difficulty of estimating the samplebased parameters (\({\hat{\theta }}\)).
Bayes error rate is another metric that quantifies the intrinsic classification limits. In classification problems, the Bayes error rate represents the minimal classification error achieved by any classifier [14, 15]. The Bayes error rate only depends on the distributions of the classes and characterizes the minimum achievable error of any classifier. Several previous studies proposed effective estimation methods for the Bayes error rate [14,15,16,17]. In particular, [18] obtains a rateoptimal nonparametric estimator of the Bayes error rate. The Bayes error rate may not be attainable with a practical classifier.
The proposed data value metric addresses the problem of measuring and tracking data information content relative to the intrinsic limits within the context of a specific analytical inferential model.
Data value metric
For a given dataset, the informationtheoretic definition of DVM employs mutual information (MI) [19, 20] to quantify the inferential gain corresponding to increasing the data size or the richness of its features. In general, mutual information evaluates the degree of relatedness between a pair of data sets. In particular, MI may be used to assess the information gain between an initial data set and its augmented counterpart representing an enhanced version of the former. When both random variables X and Y are either discrete or continuous, the mutual information can be defined by:
where p(x) and p(y) are the marginal probability distribution functions and p(x, y) is the joint probability function of X and Y. The nonnegative and symmetric MI measure expresses the intrinsic dependence in the joint distribution of X and Y, relative to the assumption of X and Y independence. Thus, MI captures the X and Y dependence in the sense that \(I(X; Y)=0\) if and only if X and Y are independent random variables, and for dependent X and Y, \(I(X; Y)>0\). Further, the conditional mutual information is defined as follows:
DVM relies on a lowdimensional representation of the data and tracks the quality of the extracted features. Either the extracted features or the predicted values from a model can be used in a lowdimensional representation in the DVM formulation. For each dataset, the DVM quantifies the performance of a specified supervised or unsupervised inference method. The DVM formulation is inspired by the concept of information bottleneck in deep neural networks (DNNs) [21, 22]. Information bottleneck represents the tradeoff between two mutual information measures: I(X; T) and I(T; Y), where X and Y are respectively the input and output of the deep learning model and T is an intermediate feature layer.
Instead of simply computing sampledriven parameter estimates, the DVM approach examines the informationtheoretic properties of datasets relative to their samplesizes, featurerichness, and the algorithmic complexity of the corresponding scientific inference. There are both similarities and differences between DMV and other VoI metrics. The main difference is that for modelbased inference, some VoI metrics may have known, exact, or asymptotic expectations based on exact, or Markov chain Monte Carlo (MCMC), posterior estimates [23,24,25]. Whereas, under modelfree inference, estimating the DVM theoretical or ergodic properties is difficult, in general. This challenge prevents the derivation of an exact linear decomposition of the error between population characteristics and their sampledriven counterparts.
This manuscript is organized as follows. In "Methods" section , we define the data value metric (DVM) as an informationtheoretic function of the (training and testing) data and the specific inferential technique. This section also includes the computational details about an effective mutual information (MI) estimator and ensemble dependency graph estimator (EDGE) [22] as well as the implementation details of a DVM Python package we built, validated, and openly shared. A feature selection application of DVM is also discussed in this section. The estimation of the mutual information using the ensemble dependency graph estimator (EDGE) is discussed in "Mutual information estimation" section. "Results" section includes experimental results illustrating the behavior of the proposed DVM metric on a wide range of real and simulated data, low and highenergy signals, featurepoor and featurerich datasets. "Conclusion and discussion" section summarizes the conclusions and provides a discussion about applications, possible improvements, limitations, and future work. In the Appendix, we provide DVM implementation details, source code references, additional results, and references to interactive 3D plots of DVM performance on real and simulated data.
Methods
There are a wide range of artificial intelligence, machine learning, and statistical inference methods for classification, regression and clustering [1, 26,27,28]. The DVM metric is applicable to unsupervised and supervised, modelbased and modelfree approaches. We employed the following supervised classification methods to identify, predict, or label predefined classes, linear models [29, 30], random forest [31], adaptive [32] and gradient [33] boosting, and knearest neighbors [34]. In addition, we tested several unsupervised clustering approaches for categorizing and grouping objects into subsets without explicit a priori labels, Kmeans [35], Affinity Propagation [36], and Agglomerative clustering [37].
The data value metric (DVM) technique utilizes MI to quantify the energy of datasets relative to the corresponding inferential technique applied to interrogate the data. Our approach is based on transforming the triple (T, S, g), representing the training (model estimation) dataset, the testing (validation) dataset, and the specific inferential method, respectively, into random variables \(X=g(X_T,X_S)\) and \(Y=Y_S\) whose MI captures the datamethod information content in the triple.
Depending upon the type of the intended inference on the data, we will define the DVM separately for supervised modeling and for unsupervised clustering. While the two definitions are congruent, this dichotomy is necessary to provide explicitly constructive definitions that can be used for a wide range of domain applications. Expanding the general regularization problem formulation, given a dataset, D, the DVM is defined as a mixture blending a fidelity term, F(D), and a regularization term, R(D):
The DVM fidelity term captures the usefulness of the sample data for the specified inferential task (supervised or unsupervised). The second, regularization, term penalizes the DVM based on the computational complexity of the corresponding inferential method. Thus, broadly speaking, the DVM depends on the data (including both training and testing sets) as well as the dataanalytic technique used to obtain the desired inference.
Let’s first explain the rationale behind mixing fidelity and regularization in the DVM definition. Consider a casestudy where a highenergy (lownoise) dataset provides sufficient information to derive either good prediction accuracy, for supervised modeling, or obtain stable clustering results, for unsupervised inference. Expanding heterogeneous data by either appending the number of samples or expanding the set of features may not always increase the DVM and may add substantial costs associated of collecting, managing, quality control, and processing the larger datasets. The penalty term in the DVM accounts for some of these potential detrimental effects due to inflating the data. The effect of the regularization term is mediated by the size of the penalty coefficient \(\lambda\), which controls the DVM balance between quality of the inference and the algorithmic complexity. There are many possible alternative forms of the regularizer term, R(D), such as runtime, computational complexity, or computing costs. In our experiments, we use the BigO computational complexity of training the predictor to quantify the regularization penalty term \(R(D)= f(n)\). Table 1 shows the computational complexities of several commonly used classification (C) and regression (R) classifiers. The table uses the following notation: n represents the size of the training sample, p is the number of features, \(k_{trees}\) is the number of trees (for treebased classifiers), \(m_{sv}\) is the number of support vectors (for SVM), and \(o_{l_i}\) is the number of neurons at layer i in a deep neural network classifier.
Next, we will focus solely on the more complex DVM fidelity term, which will be defined separately for the two alternative approachessupervised prediction and unsupervised clustering.
Representation of the fidelity term in lowdimensions
First we will define the DVM fidelity term based on lowdimensional representations of the data. The motivation behind this definition of the fidelity is driven by the neural networks (NNs) process of optimizing an objective function and identifying feature contributions. Let X, T and Y respectively denote the NN input layer, an intermediate feature layer, and the output layer.
In [21, 22], the mutual information measures I(X; T) and I(T; Y) are used to demonstrate the evolution of training in deep neural networks. I(T; Y) represents how the trained feature layer T is informative about the label. In the training process of a deep neurals network (DNN), I(T; Y) keeps increasing [21, 22]. On the other hand, I(X; T) shows the complexity of the representation T. In DNN, I(X; T) increases in the first training phase and it decreases in the compression phase [21, 22]. Thus, T is a good representation of X if its information about Y is maximized for a constrained complexity. This is equivalent to maximizing the following information bottleneck (IB) loss function [38]:
where \(\beta\) is a Lagrange multiplier with the condition \(\beta >0\).
The DVM formulation is inspired by the NN definition of information bottleneck loss function in equation (5). Intuitively, a feature vector T has high quality if it is informative about the label and its representation complexity is small. Thus, IB might be used as a measure of feature quality.
However, there are also problems with considering IB as a feature quality measure. First, in general, IB has no fixed range and it’s not a priori clear what values of IB represent high and low salient features. Second, the penalty term in the IB function, I(X; T), represents the information of the feature T about X, which captures both necessary and unnecessary information in order to predict Y. It may be better to only consider the information that is independent of Y as a penalty term. In terms of information theoretic measures, one could formulate this as conditional information I(X; TY). Note that this penalty term is minimized when the representation T yields the information of Y without extra information about X. An example of this case is when Y is an invertible function of T.
Thus, the proposed fidelity term for the Data Value Metric (DVM) is defined in terms of the mutual information and conditional mutual information measures introduced in (1) and (2) as follows:
The following remarks include some of the properties of the proposed fidelity measure.
Remark 1.a
The following inequality holds
and the fidelity term of the DVM always has the following upper bound:
Remark 1.b
\(F(T)=1\) if and only if the following equations are true:
The proof for the Remarks 1.a and 1.b is given in Appendix 1.
Remark 2
The fidelity term of the DVM can be simplified to the form of the standard information bottleneck [38]:
As a simple demonstration of the behavior of \(DVM=F\lambda R\), we fit a 5layer DNN to predict the 10 class labels of the MNIST dataset [39], and used the DVM to track the feature quality across epochs and layers. The results of the DVM performance on the digit recognition is given Fig. 1. Since the network is trained as a whole with all layers, the regularizer term R is considered fixed for all layers. At a fixed training epoch, the DVM values in different network layers represent the tradeoff between the information about the labels and the information about the input. These complementary information components are the first and second terms in the numerator of the DVM fidelity (11). During the iterative network training process, the information about the labels and the fidelity term increase, which suggests improvement of the quality of the feature layers.
Supervised modeling
The DVM fidelity term definition, equation (6), relies on lowdimensional representations. Using the supervisedmodel predicted values, we can obtain lowdimensional representations that can be used to measure data quality in both supervised and unsupervised problems. Unsupervised inference problems will be considered later.
In supervised inference, we assume that we have a set of independent and identically distributed (i.i.d.) samples \(X_i, 1\le i \le n\) with a joint distribution f(x) and associated known labels \(Y_i\). We define encoderdecoder pair \(({\mathcal {E}}, {\mathcal {D}})\), where \({\mathcal {E}}\) maps the highdimensional input X into a lower dimensional representation T, and \({\mathcal {D}}\) maps the representation T to the predicted labels. In practice, we can think of \({\mathcal {E}}\) as a dimensionalityreduction method, or the intermediate representations of a deep neural network. In addition, \({\mathcal {D}}\) performs the classification task based on the lower dimensional representations. Note that if T is simply the predicted labels, the fidelity would depend on the specific classifier. However, if T is some lowdimensional representation of the data, such as extracted features or any intermediate layer of a deep neural network, the fidelity would be independent from the classifier and would only depend on the encoder (feature extraction) method.
The definition of the fidelity measure is based on a crossvalidation type average of the definition (6) using the estimated mutual information measures. Given any random variables X, Y and Z, with corresponding sets of i.i.d. samples \(\mathbf{X },\mathbf{Y }\) and \(\mathbf{Z }\), \(I(\mathbf{X };\mathbf{Y })\) denotes the estimated mutual information using the sample sets \(\mathbf{X },\mathbf{Y }\).
We randomly split the feature set \(\mathbf{X }\) into two subsets \((\mathbf{X }',\widetilde{\mathbf{X }})\). The first subset \((\mathbf{X }')\) is used for training, whereas the second one \((\widetilde{\mathbf{X }})\) is used for independent testing and validation. Also let \(\widetilde{\mathbf{T }}\) denote the set of intermediate representation (or predicted labels), and \(\widetilde{\mathbf{Y }}\) represent the true labels associated with the test dataset \(\widetilde{\mathbf{X }}\). Then, we can define the DVM fidelity term by:
Using a weight coefficient, \(\beta\), this fidelity term formulation, equation (12), mixes two components, \(I(\widetilde{\mathbf{T }}_i;\widetilde{\mathbf{Y }}_i)\) and \(I(\widetilde{\mathbf{X }}_i;\widetilde{\mathbf{T }}_i\widetilde{\mathbf{Y }}_i)\), via normalization by \(I(\mathbf{X };\mathbf{Y })\).
The first term, \(I(\widetilde{\mathbf{T }}_i;\widetilde{\mathbf{Y }}_i)\), accounts for the fidelity of the low dimensional representation of the output labels, \(\widetilde{\mathbf{Y }}_i\), whereas the second (penalty) term, \(I(\widetilde{\mathbf{X }}_i;\widetilde{\mathbf{T }}_i\widetilde{\mathbf{Y }}_i)\), accounts for the compression of the lowerdimensional representation.
The pseudo code below (Algorithm 1) outlines the computational implementation strategy we employ in the DVM package for evaluating the DVM. The metric captures the relative analytical value of the dataset relative to the computational complexity of the supervised prediction, classification, or regression problem. In practice, the regularization term, R(g), is estimated according to the known algorithmic complexity, see Table 1.

Input : Data sets \(\mathbf{X }\), \(\mathbf{Y }\), model g, parameters \(\beta \), \(\lambda \)

for a random split \((\mathbf{X }'_i,\widetilde{\mathbf{X }}_i)\) of \(\mathbf{X }\) do

train g based on \((\mathbf{X }'_i,\mathbf{Y }'_i)\)

\(\widetilde{\mathbf{T }}_i\leftarrow g(\widetilde{\mathbf{X }}_i)\)

\(F_i\leftarrow \frac{I(\widetilde{\mathbf{T }}_i;\widetilde{\mathbf{Y }}_i) \beta I(\widetilde{\mathbf{X }}_i;\widetilde{\mathbf{T }}_i\widetilde{\mathbf{Y }}_i)}{I(\mathbf{X };\mathbf{Y })} \)


\(\widehat{D} \leftarrow \frac{1}{M} \sum _{i=1}^M F_i  \lambda R(g)\)

Output : \(\widehat{D}\)
Algorithm 1: DVM calculation for supervised problems.
Feature selection
Since DVM can be used to measure the quality of a feature set T, it can also serve as a feature selection method. In this section, we demonstrate a heuristic algorithm for sequential feature selection based on DVM values.
For a classification problem, the feature selection is defined as follows. Based on an initial feature set, choose a smaller set of features that yields a minimum prediction error. Let \(X=\{X^1,...,X^d\}\) denote the d initial features. The objective is to select a smaller set of r features with maximum DVM score. One specific approach is based on a forward selection involving r iterative steps. At each step, we select a feature from the initial feature set, \(\{X^1,...,X^d\}\), which increases DVM score the most. For a given (initial or intermediate) feature set \({{\mathcal {F}}}\), \(DVM\{{{\mathcal {F}}}\}\) represents the DVM score corresponding to that specific feature set \({{\mathcal {F}}}\). The pseudocode implementing this strategy for DVMbased feature selection is given in Algorithm 2.
Unsupervised inference
We can extend the definition of DVM for supervised problems to unsupervised clustering models. In the unsupervised problems, we don’t have explicit outcomes to evaluate the model performance.

Input: Input dataset, \(\mathbf{X }=\{X_1,...,X_N\}\)

Labels, \(\mathbf{Y }=\{Y_1,...,Y_N\}\)

Desired number of output features, r


\(\mathcal {F}:=\phi \), \(\mathcal {R}:=\{1,...,r\}\)

for each \(i\in \mathcal {R}\) do

\(f \leftarrow _{j\in \mathcal {R}\mathcal {F}} ( DVM\{\mathcal {F}\}  DVM\{\mathcal {F}\cup X^j\} )\)

Add f into \(\mathcal {F}\)


Output: \(\mathcal {F}\)
Algorithm 2: DVMbased feature selection.
Intuitively, the definition of fidelity for an unsupervised clustering method reflects the stability of the derived clusters, regardless of the clustering labels.
Our strategy for estimating the DVM fidelity for unsupervised clustering methods is based on randomly splitting the dataset \(\mathbf{X }\) into three subsets \((\mathbf{X }',\mathbf{X }'',\widetilde{\mathbf{X }})\).
The first two of these sets, \((\mathbf{X }',\mathbf{X }'')\), are used for crossvalidation training, whereas the remaining one, \(\widetilde{\mathbf{X }}\), is used for independent testing and validation. By training the classifier on the first subset \((\mathbf{X }')\), we obtain derived computed labels. These predicted labels, \({\widehat{\mathbf{Y }}}\), may be used as baseline for computing the fidelity based on the information bottleneck in equation (12). Let \(\widetilde{\mathbf{T }}\) be the representation layer (or predicted indices associated with the test dataset \(\widetilde{\mathbf{X }}\)). The DVM fidelity term for unsupervised learners may then be defined as follows:
where the index i in the above definition denote the the variables associated with the ith randomized splitting of \(\mathbf{X }\). Just as we did for the supervised problems, we can explicate the DVM algorithmic implementation via the pseudo code used in the DVM package.
The algorithm below (Algorithm 3) shows the DVM calculation for unsupervised clustering and classification problems. Again, the regularization term is derived using the approximate estimate of the computational complexity associated with the classifier (R(g)), see Table 1.

Input: Data sets \(\mathbf{X }\), \(\mathbf{Y }\), model g, parameters \(\beta \) and \(\lambda \)

for a random split \((\mathbf{X }'_i,\mathbf{X }'', \widetilde{\mathbf{X }}_i)\) of \(\mathbf{X}\) do

Apply unsupervised model g based on \(\widetilde{\mathbf{X }}_i\)

\(\hat{\mathbf{Y }}_i\leftarrow g(\mathbf{X }'_i)\)

\(\widetilde{\mathbf{T }}_i\leftarrow g(\mathbf{X }''_i)\)

\(F_i\leftarrow \frac{I(\widetilde{\mathbf{T }}_i;\hat{\mathbf{Y }}_i) \beta I(\widetilde{\mathbf{X }}_i;\widetilde{\mathbf{T }}_i\hat{\mathbf{Y }}_i)}{I(\mathbf{X };\hat{\mathbf{Y }})} \)


\(\widehat{D} \leftarrow \frac{1}{T} \sum _{i=1}^T F_i  \lambda R(g)\)

Output: \(\widehat{D}\)
Algorithm 3: DVM calculation for unsupervised problems.
Mutual information estimation
In many areas, including data science and machine learning, the density of the data is unknown. In these cases, one needs to estimate the mutual information from the data points. Examples of MI estimation strategies include KSG [40], KDE [41], Parzen window density estimation [42], and adaptive partitioning [43].
The computational complexity and convergence rate are two important performance metrics of various MI estimators. The process of MI estimation is computationally intensive for large data sets, e.g., the computational complexity of the KDE method is \(O(n^2)\), while the KSG method takes \(O(k\ n \log (n))\) time to compute MI (k is a parameter of the KSG estimator). More computationally efficient estimators such as [44] provide improvements with estimated MI estimation time of \(O(n \log (n))\).
Thus, estimation of mutual information for large and complex data sets requires some approximation. For instance, we can use one of the standard estimators that exist for the nonparametric distributions. Nonparametric estimators are a family of estimators, for which we consider minimal assumptions on the density functions. There are several previous approaches, e.g., [45,46,47,48], that guarantee optimal convergence rates. Among these estimators, the hashbased estimator proposed in [48] has linear computational complexity. As we deal with large and complex data sets, here we employ a hashbased mutual information estimator, called the ensemble dependency graph estimator (EDGE) [22]. EDGE has an optimal mean square error (MSE) convergence rate and low computational complexity that make it suitable for our task of detecting the information gain associated with augmenting a data set.
Results
We conducted a number of experiments to illustrate the use of the proposed DVM on a wide range of real and simulated datasets. Each dataset was labeled as low, medium, or high energy, indicating the strength of the signal information content in the data. The results of different machine learning and statistical modeling methods, their quality, accuracy, and reproducibility heavily depend on the intrinsic signal energy. We contrast the proposed DVM against classifieraccuracy and Bayes optimal classifier accuracy, which is a measure of classification task difficulty. In this paper, we define the Bayes classifier accuracy as the additive complement of the classical Bayes error rate (risk), i.e., Bayesian Accuracy = 1 Bayesian Error.
Datasets
MNIST Handwritten Digits Data: The Modified National Institute of Standards and Technology (MNIST) dataset consists of a large number of fixedsize, grayscale images of handwritten digits. It includes a set of 60,000 training images, and a set of 10,000 test images. Each image has a dimension \(28\times 28\), and each pixel intensity takes a value between 0 and 255. The training data are also paired with a label (0, 1, 2, ...,9) indicating the correct number represented in the corresponding image [39].
ALS dataset: Amyotrophic lateral sclerosis (ALS) is a complex progressive neurodegenerative disorder with an estimated prevalence of about 5 per 100,000 people in the United States. The disease severity is enormous with many the patients surviving only a few years after ALS diagnosis, and few living with ALS for decades [49]. We used the ProACT openaccess database [50], which collects and aggregates clinical data of 16 ALS clinical trials and one observational study completed in the recent twenty years [51].
This dataset contains the information of 2,424 patients with 249 clinical features, tracked over 12 months. The ALS disease progression, which is measured by the change of Amyotrophic Lateral Sclerosis Functional Rating Scale (ALSFRS) score over time, is used as the target variable. ALSFSR is a realvalued number in the range [0, 5].
Simulated dataset: Synthetic data were generated using make_blobs function in scikitlearn (https://scikitlearn.org). There were five centers for the dataset. Each dataset had 2,000 samples and 800 features. The standard deviation for strong signal data was 20, while it was 40 for weak signal data.
The continuous data was generated using the following formula:
where X was generated by sampling 800 random observations from a multivariate Gaussian distribution. The mean vector of this multivariate Gaussian distribution was generated from a Gaussian distribution with mean zero and variance 25. The eigenvalues of the diagonal variancecovariance matrix of the multivariate Gaussian distribution were generated from a Uniform(2; 12) distribution. The noise term follows a standard Gaussian distribution and its magnitude term, K, was chosen to be 10 for the strong signal or 50 for the weak signal simulated datasets.
Validation experimental design
Our experimental design included supervised and unsupervised machine learning methods using real and simulated datasets with different signal profiles – weak and strong signals. Figure 2 shows the specific supervised and unsupervised methods, and the type of data used in the DVM validation protocol. The labels strong and weak associated with different datasets qualify the relative size of the information content in the data, i.e., the relative signal to noise ratio. For the observed datasets, this information content reflects the power of the covariate features to predict an outcome (for supervised problems) or the consistency of the derived labels (for unsupervised problems). For the simulated data, the information energy is directly related to signaltonoise ratio (\(SNR<0.2\) vs. \(SNR>2.0\)). For each of the cells in the validation design, we computed the DVM as a parametric surface defined over the 2D grid parameterized over data samplesize and numberoffeatures. The reported results include 2D plots of crosssections of the DVM surface for a fixed samplesize or a fixed numberoffeatures. We also plotted the complete 3D DVM surfaces rendered as triangulated 2manifolds. These interactive 3D plots are available in supplementary materials and are accessible on our webserver.
Strong signal datasets: Fig. 3 compares the DVM value to the classification accuracy and Bayes accuracy rates on the MNIST dataset using the Random Forest classifier. As the sample size and the number of features increase, the classification accuracy, Bayes accuracy, and the DVM increase. The \(95\%\) confidence interval is represented by the shaded area around the DVM curve.
Using the MNIST data, the results in Fig. 3a imply that both the classification accuracy and DVM drastically increase with increase of the sample size between 500 and 4,500. The accuracy converges to around 0.85 when the sample size approaches 4,500. In the same range, the DVM also converges to around 0.8. Similar results relative to the increase of the number of features are show in Fig. 3b. As the number of features approaches 800, the accuracy converges to around 0.86 and the DVM approaches 0.8.
Using strongsignal simulated data, the results in Fig. 4a show that classification accuracy, Bayes accuracy, and DVM increase as the sample size grows from 200 to 2000. The accuracy converges to around 0.95 and DVM approaches 0.92 for large samplesizes. The result in Fig. 4b also shows the growth of the classification accuracy and DVM as the number of features increases from 100 to 800, but plateaus around 300 features.
Figure 5 displays a 3D surface plot for the classification accuracy and DVM parameterized by the sample size and the number of features. This graph provides more information compared to the crosssectional linear plots shown in Figs. 3, 4. Interactive 3D surface plots for all experiments are available online (see Appendix 1, 2).
These results illustrate that for some strong signals, there may be little gain of increasing the samplesize or the number of features.
Weaksignal datasets: Fig. 6 shows the results of the accuracy and DVM for the real (ALS) weaksignal dataset. As expected, the DVM pattern is less stable, but still suggests that adding additional cases or enhancing the features of the data adds little value to improve the unsupervised clustering of the data (Kmeans clustering).
Figure 7 depicts the DVM trends using the weak simulated data. Again, the overall low DVM values suggest that increasing the size of augmenting the complexity of weaksignal data may not significantly improve the subsequent unsupervised clustering.
Interactive 2D and 3D DVM surface plots illustrating the results of each experiment are available online at https://socr.umich.edu/docs/uploads/2020/DVM/. These graphs show the behavior of the DVM spanning the domain of possible number of cases and number of features for the real and simulated datasets.
In the appendix, we show examples of cases (pairs of datasets and classifiers) where the DVM may actually decrease with an increase of the number of samples or the number of features.
Feature selection
We demonstrate the feature selection algorithm introduced in algorithm 2 on a simulated dataset. The simulated dataset consists of 1000 samples randomly drawn from a 4cluster 2DGaussian distribution. The clusters are on a square with edge size 1, where the label for each sample determines the distribution cluster. The dimension of the samples is 20 and the problem is to select up to 15 features. Figure 8 represents the steps of the feature selection algorithm. At each step, the best of all features is selected using DVM and added to the chosenfeatures set. Note that due to the dimensionality and runtime complexity terms in the DVM definition, we do not expect a monotonic graph, however, the local maximums suggest an appropriate stopping criterion for the feature selection process. Figure 8 shows the performance of the DVMbased feature selection yielding a 6element feature set, \(\{F18,F4,F13,F9,F5,F12\}\), corresponding to a high DVM value, \(DVM=0.84\).
Conclusion and discussion
This manuscript presents the mathematical formulation, algorithmic implementation, and computational validation of a data value metric (DVM) for quantifying the analyticalvalue and informationcontent (energy) of a dataset. DVM depends on the intended data processing, modeling, forecasting or classification strategies used to interrogate the data. The significance of the work is the introduction of a new measure of intrinsic data value, the DVM, that complements other traditional measures of analytic performance, e.g., accuracy, sensitivity, logodds ratio, Bayesian risk, positive predictive power, and area under the receiver operating characteristic curve. Through the experiments presented herein, authors discovered that the DVM captures the important trends of traditional measures applied to different types of datasets. The DVM tuning parameter (alpha) provides flexibility for balancing between algorithmic performance and computational complexity, which facilitates a dataspecific quantization of the relative information content in a dataset.
As the DVM is applicable for a wide range of datasets and a broad gamut of supervised and unsupervised analytical methods, it can be used as a single unified measure to guide the process of data augmentation, data reduction, and feature selection. It would be interesting to compare the DVMdriven feature selection to other variable selection methods [1], e.g., filtering methods such as information gain and Markov blanket filtering, wrapper techniques such as recursive feature elimination and simulated annealing, and embedded strategies such as random forests and weightedSVM.
The DVM evaluates the analytical value of a dataset relative to a predefined analytical technique for the data interrogation. The two primary benefits of using an informationtheoretic measure, such as the regularized DVM, as a datautility metric include (1) the estimate of the DVM is easy to compute for each triple of a dataset, analytical strategy, and performance measures, and (2) the DVM magnitude (high or low value) serves as a proxy translating specific datamining challenges and observable data into a continuous pseudodistance metric of informationcontent relative to computationalcomplexity.
The normalization of the DVM fidelity term ensures that the informationvalue of the data is standardized in a uniform range, [0,1]. Relative to an a priori analytical strategy, extreme fidelity values close to 0 or 1 correspond respectively to lowquality and highinformationcontent datasets. The real data and simulationbased results show that there is a connection between error rate and the DVM values. However, theoretical bounds on the discrepancy between the prediction error rate and the informationbased DVM are not yet understood. Future studies are needed to explore this theoretical relation for various types of analytical methods and data characteristics.
As firm supporters of openscience, we have shared all code, data, and results on the DVM GitHub page (https://github.com/SOCR/DVM/) and the SOCR DVM documentation site (https://socr.umich.edu/docs/uploads/2020/DVM/).
Availability of data and materials
All data generated, simulated, or analysed during this study are included in this published article [and its supplementary information files].
References
 1.
Dinov ID. Data science and predictive analytics biomedical and health applications using R. Berlin: Springer; 2018.
 2.
Raiffa H, Schlaifer R. Applied statistical decision theory 1961.
 3.
Baio G. Statistical modeling for health economic evaluations. Ann Revi Statist Appl. 2018;5(1):289–309. https://doi.org/10.1146/annurevstatistics031017100404.
 4.
Baio G, Heath A. When simple becomes complicated: why Excel should lose its place at the top table. London: SAGE Publications Sage UK; 2017.
 5.
Parmigiani G, Inoue L. Decision Theory: Principles and Approaches, vol. 812. Hoboken: Wiley; 2009.
 6.
Jackson C, Presanis A, Conti S, Angelis DD. Value of information: sensitivity analysis and research design in bayesian evidence synthesis. J Am Statist Associat. 2019;114(528):1436–49. https://doi.org/10.1080/01621459.2018.1562932.
 7.
Madan J, Ades AE, Price M, Maitland K, Jemutai J, Revill P, Welton NJ. Strategies for efficient computation of the expected value of partial perfect information. Med Decis Making. 2014;34(3):327–42.
 8.
Strong M, Oakley JE. An efficient method for computing singleparameter partial expected value of perfect information. Med Decis Making. 2013;33(6):755–66.
 9.
Sadatsafavi M, Bansback N, Zafari Z, Najafzadeh M, Marra C. Need for speed: an efficient algorithm for calculation of singleparameter expected value of partial perfect information. Value Health. 2013;16(2):438–48.
 10.
Strong M, Oakley JE, Brennan A. Estimating multiparameter partial expected value of perfect information from a probabilistic sensitivity analysis sample: a nonparametric regression approach. Med Decis Making. 2014;34(3):311–26.
 11.
Strong M, Oakley JE, Brennan A, Breeze P. Estimating the expected value of sample information using the probabilistic sensitivity analysis sample: a fast, nonparametric regressionbased method. Med Decis Making. 2015;35(5):570–83.
 12.
Heath A, Manolopoulou I, Baio G. Estimating the expected value of partial perfect information in health economic evaluations using integrated nested laplace approximation. Statist Med. 2016;35(23):4264–80.
 13.
Meng XL. Statistical paradises and paradoxes in big data (i): law of large populations, big data paradox, and the 2016 us presidential election. Ann Appl Stat. 2018;12(2):685–726. https://doi.org/10.1214/18AOAS1161SF.
 14.
Wang Q, Kulkarni SR, Verdú S. Divergence estimation of continuous distributions based on datadependent partitions. IEEE Transact Informat Theory. 2005;51(9):3064–74.
 15.
Póczos B, Xiong L, Schneider J. Nonparametric divergence estimation with applications to machine learning on distributions. In: UAI (also arXiv Preprint arXiv:1202.3758 2012) 2011.
 16.
Berisha V, Wisler A, Hero AO, Spanias A. Empirically estimable classification bounds based on a nonparametric divergence measure. IEEE Transact Signal Process. 2016;64(3):580–91.
 17.
Noshad M, Hero A. Scalable hashbased estimation of divergence measures. In: International Conference on Artificial Intelligence and Statistics, 2018;pp. 1877–1885.
 18.
Noshad M, Xu L, Hero A. Learning to benchmark: Determining best achievable misclassification error from training data. arXiv preprint arXiv:1909.07192 2019.
 19.
Ho SW, Verdú S. Convexity/concavity of renyi entropy and \(\alpha\)mutual information. In: Information Theory (ISIT), 2015 IEEE International Symposium On, 2015;pp. 745–749. IEEE
 20.
Cover TM, Thomas JA. Elements of information theory. Hoboken: Wiley; 2012.
 21.
ShwartzZiv R, Tishby N. Opening the black box of deep neural networks via information. arXiv preprint arXiv:1703.00810 2017.
 22.
Noshad M, Zeng Y, Hero AO. Scalable mutual information estimation using dependence graphs. In: ICASSP 20192019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2019;pp. 2962–2966. IEEE
 23.
Ades A, Sutton A. Multiparameter evidence synthesis in epidemiology and medical decisionmaking: current approaches. J Royal Stat Soci Series A. 2006;169(1):5–35.
 24.
Oakley JE, O’Hagan A. Probabilistic sensitivity analysis of complex models: a bayesian approach. J Royal Statist Soc Series B. 2004;66(3):751–69.
 25.
Saltelli A, Tarantola S, Campolongo F, Ratto M. Sensitivity analysis in practice: a guide to assessing scientific models. Chichester. 2004.
 26.
Pan SJ, Yang Q. A survey on transfer learning. IEEE Transact Knowl Data Eng. 2010;22(10):1345–59. https://doi.org/10.1109/TKDE.2009.191.
 27.
Denison DD, Hansen MH, Holmes CC, Mallick B, Yu B. Nonlinear Estimation and Classification. Lecture Notes in Statistics. Springer. 2013. https://books.google.com/books?id=0IDuBwAAQBAJ
 28.
Ghahramani Z. Probabilistic machine learning and artificial intelligence. Nature. 2015;521(7553):452–9.
 29.
Faraway JJ. Extending the Linear Model with R: Generalized Linear, Mixed Effects and Nonparametric Regression Models. Chapman and Hall/CRC, ??? 2016.
 30.
Tibshirani R. The lasso method for variable selection in the cox model. Statist Med. 1997;16(4):385–95.
 31.
Liaw A, Wiener M, et al. Classification and regression by randomforest. R News. 2002;2(3):18–22.
 32.
Margineantu DD, Dietterich TG. Pruning adaptive boosting. In: ICML, 1997;vol. 97, pp. 211–218. Citeseer
 33.
Chen T, Guestrin C. Xgboost: A scalable tree boosting system. In: Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD ’16, pp. 785–794. ACM, New York 2016. https://doi.org/10.1145/2939672.2939785.
 34.
Dudani SA. The distanceweighted knearestneighbor rule. IEEE Transact Syst Man Cybernet. 1976;4:325–7.
 35.
Hartigan JA, Wong MA. Algorithm as 136: a kmeans clustering algorithm. J Royal Statist Soc Series C. 1979;28(1):100–8.
 36.
Bodenhofer U, Kothmeier A, Hochreiter S. Apcluster: an r package for affinity propagation clustering. Bioinformatics. 2011;27(17):2463–4.
 37.
Murtagh F, Legendre P. Ward’s hierarchical agglomerative clustering method: which algorithms implement ward’s criterion? J Classificat. 2014;31(3):274–95.
 38.
Alemi AA, Fischer I, Dillon JV, Murphy K. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410 2016.
 39.
Deng L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process Magaz. 2012;29(6):141–2. https://doi.org/10.1109/MSP.2012.2211477.
 40.
Kraskov A, Stögbauer H, Grassberger P. Estimating mutual information. Phys Rev E. 2004;69(6):066138.
 41.
Moon Y, Rajagopalan B, Lall U. Estimation of mutual information using kernel density estimators. Phys Rev E. 1995;52(3):2318.
 42.
Kwak N, Choi CH. Input feature selection by mutual information based on parzen window. IEEE Transact Pattern Analy Mach Intell. 2002;24(12):1667–71.
 43.
Stowell D, Plumbley MD. Fast multidimensional entropy estimation by \(k\)d partitioning. IEEE Signal Process Lett. 2009;16(6):537–40. https://doi.org/10.1109/LSP.2009.2017346.
 44.
Evans D. A computationally efficient estimator for mutual information. In: Proceedings of the Royal Society of London A: Mathematical, Physical and Engineering Sciences, 2008;vol. 464, pp. 1203–1215. The Royal Society
 45.
WaltersWilliams J, Li Y. Estimation of mutual information: A survey. In: International Conference on Rough Sets and Knowledge Technology, 2009;pp. 389–396. Springer
 46.
Singh S, Póczos B. Generalized exponential concentration inequality for rényi divergence estimation. In: International Conference on Machine Learning, 2014;pp. 333–341.
 47.
Noshad M, Moon KR, Sekeh SY, Hero AO. Direct estimation of information divergence using nearest neighbor ratios. In: 2017 IEEE International Symposium on Information Theory (ISIT), 2017;pp. 903–907. IEEE
 48.
Noshad M, Hero AO. Scalable hashbased estimation of divergence measures. In: 2018 Information Theory and Applications Workshop (ITA), 2018; pp. 1–10. IEEE
 49.
Tang M, Gao C, Goutman SA, Kalinin A, Mukherjee B, Guan Y, Dinov ID. Modelbased and modelfree techniques for amyotrophic lateral sclerosis diagnostic prediction and patient clustering. Neuroinformatics. 2019;17(3):407–21. https://doi.org/10.1007/s1202101894069.
 50.
Rahme R, Yeatts SD, Abruzzo TA, Jimenez L, Fan L, Tomsick TA, Ringer AJ, Furlan AJ, Broderick JP, Khatri P. Early reperfusion and clinical outcomes in patients with m2 occlusion: pooled analysis of the proact ii, ims, and ims ii studies. J Neurosurgery JNS. 2014;121(6):1354–8.
 51.
Glass JD, Hertzberg VS, Boulis NM, Riley J, Federici T, Polak M, Bordeau J, Fournier C, Johe K, Hazel T, Cudkowicz M, Atassi N, Borges LF, Rutkove SB, Duell J, Patil PG, Goutman SA, Feldman EL. Transplantation of spinal cord–derived neural stem cells for als. Neurology. 2016;87(4):392–400. https://doi.org/10.1212/WNL.0000000000002889. https://n.neurology.org/content/87/4/392.full.pdf
Acknowledgements
This work was supported in part by NSF grants 1916425, 1734853, 1636840, 1416953, 0716055 and 1023115, NIH grants P20 NR015331, U54 EB020406, P50 NS091856, P30 DK089503, UL1TR002240, R01CA233487, R01MH121079, and K23 ES027221, and ARO grant W911NF1510479. The funders played no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. Colleagues at the University of Michigan Statistics Online Computational Resource (SOCR) and the Michigan Institute for Data Science (MIDAS) contributed ideas, infrastructure, and support for the project.
Author information
Affiliations
Contributions
ID and MN contributed to the study conception and design. Material preparation, data collection, coding, and analysis were performed by MN, JC, YS, and ID. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1
Proof of Remark 1
First note that the following equation holds [22]:
Using the chain rule for the mutual information we have
On the other hand we also have the following equation:
From (16) and (18) we obtain the following inequality:
and the equality holds if and only if \(I(X;YT)=0\). Therefore, \(F(1)=1\) if and only if the second term in (6) is equal to zero and we have \(I(X;YT)=0\). An example of a case with conditions in (9) is when Y is an invertible function of T. \(\square\)
Appendix 2
Implementation
Below, we briefly describe the DVM Python package organization and invocation. We have implemented a DVM python package for our data value metric framework and made it available on GitHub (https://github.com/SOCR/DVM).
The DVM package can be used on any dataset and any userdefined supervised or supervised tasks in order to evaluate the quality of the data. The package consists of three main python files, DVM.py, methods.py, and DVM_plot.py. Please note that DVM.py uses the mutual information estimator file, EDGE.py, as its dependency.
DVM.py gets the input datasets X and in the case of a supervised task, a set of corresponding labels denoted by Y. Further, the user needs to specify the input parameters, \(\beta\), problem_type and method. \(\beta\) is the coefficient of the regularizer term of DVM. problem_type specifies whether the task is supervised or unsupervised, and method is the learning method that is used by user. For a given method, we can also input the corresponding required arguments. For example, if we are using KNN_classifier from methods.py as our method, it requires the parameters n_neighbors (number of neighbors) and weights (type of weights) as input:
There are two DVM output values: DVM_value and confidence_band. DVM_value gives the average value computed according to the DVM formula in equation (6) and confidence_band gives the \(95\%\) confidence limits of the DVM values computed by different subsets in equation (13).
The methods.py file consists of various supervised and supervised methods. Each supervised method takes the following arguments: X_train, Y_train, X_test, **kwargs. X_train, Y_train, X_test respectively are the train data set and labels, and the test data set for which we would like to predict labels. **kwargs specifies all of the arguments that the given method requires. An example of the format is as follows:
The output of the method is a numpy array of predicted labels. Note that in addition to the methods listed in methods.py file, any other user defined method that satisfies the above format can be used for DVM.
The DVM_plot.py gets the input datasets X, and in the case of a supervised task, a set of corresponding labels denoted by Y. Further the user need to specify the input parameters continuous, \(\beta\), problem_type, method, plot_type. Continuous indicates whether the response variable is continuous variable or discrete variable. plot_type specifies the plots that the user wants to generate, where plot_type = ’3D’ generates 3D plots of DVM, plot_type = ’3D’ generates 2D plots, and plot_type = ’Both’ generates both 2D and 3D plots.\(\beta\), problem_type, method have the same meaning as in the DVM.py. For a given method, we can also input the corresponding required arguments like DVM.py. The same example as DVM.py is used here to illustrate the syntax:
In addition to the 2D and 3D DVM plots, DVM_plot also outputs a dictionary containing Accuracy, MI(mutual information), Complexity, DVM,Sample Number (a sequence of different number of samples) and Feature Number (a sequence of different of features.)
As calculating the DVM measure actually involves another parameter (\(\lambda\)) that represents the weightaveraging of the DVM fidelity and regularization terms, the actual DVM manifold is intrinsically a surface embedded in 4D. We have designed a kimesurface visualization that allows us to explore the properties of the DVM manifold by including a \(\lambda\) slider that reduces the DVM into an animated 3D space.
Supplementary experiments
The appendix below includes four additional tables of results that illustrate some of the DVM performance in different situations. Readers are encouraged to view the corresponding DVM interactive 2D plots and 3D surface graphs on the website,
https://socr.umich.edu/docs/uploads/2020/DVM/. See Tables 2, 3, 4, 5
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Noshad, M., Choi, J., Sun, Y. et al. A data value metric for quantifying information content and utility. J Big Data 8, 82 (2021). https://doi.org/10.1186/s40537021004466
Received:
Accepted:
Published:
Keywords
 Data energy
 Artificial intelligence
 Machine learning
 Data utility
 Information content